| { |
| "paper_id": "2020", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T06:34:36.681501Z" |
| }, |
| "title": "Sentence Boundary Detection on Line Breaks in Japanese", |
| "authors": [ |
| { |
| "first": "Yuta", |
| "middle": [], |
| "last": "Hayashibe", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Megagon Labs", |
| "location": { |
| "addrLine": "Ltd. 7-3-5 Ginza Chuo-ku", |
| "postCode": "104-8227", |
| "settlement": "Tokyo, Tokyo", |
| "region": "Recruit Co", |
| "country": "Japan, Japan" |
| } |
| }, |
| "email": "hayashibe@megagon.ai" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "For NLP, sentence boundary detection (SBD) is an essential task to decompose a text into sentences. Most of the previous studies have used a simple rule that uses only typical characters as sentence boundaries. However, some characters may or may not be sentence boundaries depending on the context. We focused on line breaks in them. We newly constructed annotated corpora, implemented sentence boundary detectors, and analyzed performance of SBD in several settings.", |
| "pdf_parse": { |
| "paper_id": "2020", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "For NLP, sentence boundary detection (SBD) is an essential task to decompose a text into sentences. Most of the previous studies have used a simple rule that uses only typical characters as sentence boundaries. However, some characters may or may not be sentence boundaries depending on the context. We focused on line breaks in them. We newly constructed annotated corpora, implemented sentence boundary detectors, and analyzed performance of SBD in several settings.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Many NLP tasks treat a sentence as a unit of processing. The task for decomposing a text into sentences is called sentence boundary detection (SBD). In Japanese, periods (e.g. \" \", \".\"), exclamation marks, and question marks are delimiters to segment sentences in most cases. For this reason, the SBD in most studies takes only the positions of these typical delimiters as sentence boundaries. For example, in the construction of the \"Web Japanese N-gram database 1 \" provided by Google, Inc., they extracted sentences by segmenting on their positions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "However, line breaks can also indicate sentence boundaries without periods as the following text 2 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "(Ms. Okubo of \"Oasiz\" has been on my mind lately </s>What TV shows does she often appear on?)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "</s>", |
| "sec_num": null |
| }, |
| { |
| "text": "(1) * Both authors contributed equally. 1 https://www.gsk.or.jp/files/catalog/ GSK2007-C/GSK2007C_README.utf8.txt", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "</s>", |
| "sec_num": null |
| }, |
| { |
| "text": "2 In this paper, we use \" \" to show line breaks and \"</s>\" to show sentence boundaries.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "</s>", |
| "sec_num": null |
| }, |
| { |
| "text": "Many line breaks do not follow the typical delimiters. For example, 33.4% of line breaks in the balanced corpus of contemporary written Japanese (BCCWJ) (Maekawa, 2008) were not followed by them. On the other hand, line breaks may be placed in the middle of a sentence. Therefore, we can not simply treat the positions of line breaks as sentence boundaries.", |
| "cite_spans": [ |
| { |
| "start": 153, |
| "end": 168, |
| "text": "(Maekawa, 2008)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "</s>", |
| "sec_num": null |
| }, |
| { |
| "text": "(Among recent movies are there any with Gary Oldman?) 2This type of line break is used to make long sentences easy to read. Shinmori et al. (2003) performed a structural analysis of Japanese patent documents. They reported that 48.5% of the first claim in the 59,968 patent documents contain line breaks in the sentence. They explain it is common that claims written in Japanese are described in one sentence and the use of line breaks is intended to improve readability.", |
| "cite_spans": [ |
| { |
| "start": 124, |
| "end": 146, |
| "text": "Shinmori et al. (2003)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "</s>", |
| "sec_num": null |
| }, |
| { |
| "text": "There can be more sentence boundaries than these. Nishimura (2003) showed that there are more than six variations of Japanese sentence boundaries in an online forum: Description of actions (e.g. \"( )\": embarrassment, \"( )\": tears), \"Smiley\" Icons (e.g. \"( *\u02c6 \u02c6*)\", \" \u02c6 \u02c6 \"), and so on. Sakai (2013) conducted a linguistic analysis of Japanese emails written by young people on their mobile phones and found that about 63% of the emails used emoticons instead of punctuation marks for sentence boundaries.", |
| "cite_spans": [ |
| { |
| "start": 50, |
| "end": 66, |
| "text": "Nishimura (2003)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "</s>", |
| "sec_num": null |
| }, |
| { |
| "text": "In this paper, we focus on SBD on line breaks in Japanese. We newly construct annotated corpora to answer the following three research questions: 2 Related Work Zhu et al. (2007) removed noise in English email text by removing extra lines and spaces and restoring wrong cases of characters. They show that 49.5% of the noise in about 5,000 texts is due to line breaks. Three labels for line breaks were trained and predicted by the Conditional Random Fields (CRF) algorithm (Lafferty et al., 2001 ): PRV (Preserve line break), RPA (Replace line break by space), DEL (Delete line break). The accuracy is reported as F-measure 93.75. Huang and Chen (2011) insist that the concept of \"sentences\" is fuzzier and less-defined in Chinese, and Native Chinese writers seldom follow the usage guidelines of punctuation marks. They listed the symbols used as sentence boundaries, such as whitespaces, commas, periods, line breaks. They reported F1 of manual SBD is 81.18 and one of CRF is 77.48.", |
| "cite_spans": [ |
| { |
| "start": 161, |
| "end": 178, |
| "text": "Zhu et al. (2007)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 474, |
| "end": 496, |
| "text": "(Lafferty et al., 2001", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "</s>", |
| "sec_num": null |
| }, |
| { |
| "text": "Stanza 3 (Qi et al., 2020 ) is a language-agnostic fully neural pipeline for text analysis, including tokenization, multiword token expansion, lemmatization, part-of-speech and morphological feature tagging, dependency parsing, and named entity recognition. Unlike most existing toolkits, it does tokenization and SBD at the same time by using a bidirectional long short-term memory network (Graves and Schmidhuber, 2005 ) (Bi-LSTM) for characters in texts. It provides models for 66 languages including Japanese. The Japanese model is trained with with UD Japanese GSD 4 . Its architecture enables SBD on any characters, including line breaks. However the training corpus does not contain line breaks. Therefore the model can not perform SBD on line breaks. tence boundaries (Konishi et al., 2015) and morphological information. It covers a wide range of genres such as books, magazines, newspapers, business reports, blogs, internet forums, and textbooks. Some of them contain line breaks. Table 1 shows the statistics of the corpus 5,6 It consists 44,770 sentences in 2,918 documents. They contain 1,721 line breaks that do not segment sentences out of 26,056 line breaks. Such line breaks are contained only limited domains: QA texts and blog texts on the web.", |
| "cite_spans": [ |
| { |
| "start": 9, |
| "end": 25, |
| "text": "(Qi et al., 2020", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 391, |
| "end": 420, |
| "text": "(Graves and Schmidhuber, 2005", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 776, |
| "end": 798, |
| "text": "(Konishi et al., 2015)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "</s>", |
| "sec_num": null |
| }, |
| { |
| "text": "We create two kinds of Japanese corpora with sentence boundary annotation: Jalan-F and Jalan-A, in order to perform experiments in another domain and another writing style. Both of them are composed of a part of hotel reviews posted on Jalan 7 , which is a popular travel information web site. Table 1 shows the statistics of the corpora. All annotations are performed by one worker and confirmed by another worker. Jalan-F 8 comprises 500 reviews. We fully anno-tated sentence boundaries for all texts. As a result, we found 3,290 sentences. It contains 1,484 line breaks. Out of them, 170 line breaks do not segment sentences. Jalan-A 9 comprises 298 reviews in an atypical writing style. They do not contain typical Japanese periods (\" \"). This is an example. </s> </s> (We stayed in a private room </s> The room was clean and spacious, so we'll be back again </s> The staffs were great) 3Some line breaks segment sentences and some do not. We only annotated sentence boundaries on line breaks. While the number of boundaries is 1,374, there may be more sentences. It contains 1,983 line breaks. Out of them, 153 line breaks do not segment sentences.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 294, |
| "end": 301, |
| "text": "Table 1", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Jalan Corpora", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "To answer the third research question, we created two pseudo annotation corpora: P-BCCWJ and P-Jalan. First, we removed all line breaks from BCCWJ and 10,000 reviews additionally extracted from Jalan. Then, we replaced typical Japanese sentence boundaries \" \" into line breaks and regard all of them as sentence boundaries. Finally, we replaced ideographic commas \" \" into line breaks with 50% probability. This is an example.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Pseudo Annotation Corpora", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "(It is to look into the distance from a good view.) Pseudo annotation:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Original:", |
| "sec_num": null |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "</s>", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Original:", |
| "sec_num": null |
| }, |
| { |
| "text": "4 Experiments", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Original:", |
| "sec_num": null |
| }, |
| { |
| "text": "We create sentence boundary detectors by fineturning the BERT (Devlin et al., 2019 ) model 10 pre-trained on Japanese Wikipedia by Tohoku 9 The \"A\" is an abbreviation for \"atypical.\" 10 https://huggingface.co/cl-tohoku/ bert-base-japanese-whole-word-masking University. Texts are first tokenized with MeCab 11 morphological parser and then spitted into subwords by WordPiece. Its vocabulary size is 32,000. We exploit implementations of sequence labeling in \"transformers\" 12 by Hugging Face with three labels 13 : \"Sentence boundary\" (SB) and \"Not sentence boundary\" (NSB) for line breaks, and \"Others\" (O) for tokens that are not line breaks. We only use predictions for line breaks. Table 2 shows an example of input, output, and evaluation for detectors. In training, all tokens are labeled \"O\" except for line breaks. Whatever predictions are output for them, we do not consider them in the evaluation. Line breaks are labeled \"SB\" or \"NSB\" for training. We recognize sentence boundaries only on the tokens whose predictions are \"SB.\"", |
| "cite_spans": [ |
| { |
| "start": 62, |
| "end": 82, |
| "text": "(Devlin et al., 2019", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 686, |
| "end": 693, |
| "text": "Table 2", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experiment Settings", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "We set the maximum sequence length 320, the training batch size 32, and the number of epochs five. If the maximum number of input tokens is exceeded, we divide them into multiple inputs. We perform the Unicode NFKC normalization for all inputs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiment Settings", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "For training and evaluation, we exclude 663 documents from BCCWJ and 164 documents from Jalan-F that do not contain line breaks. Each corpus of BCCWJ, Jalan-F, and Jalan-A is divided into 8:2 for learning and training. We built four models by using the three training sources and the data from the combination of Jalan-F and Jalan-A.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiment Settings", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "First, we investigate the impact of domains. As shown in Table 3 Table 5 : SBD Performance on line breaks by models trained with pseudo corpora bad compared to one of the model BCCWJ (98.2). This shows that we can make reasonably accurate models using training data even from different domains. On the other hand, F 1 scores for Jalan-F test data are close to 100 for all models. Therefore, we consider Jalan-F only contains simple cases.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 57, |
| "end": 64, |
| "text": "Table 3", |
| "ref_id": "TABREF6" |
| }, |
| { |
| "start": 65, |
| "end": 72, |
| "text": "Table 5", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experiments 1: Impact of Domains", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Second, we investigate the impact of writing styles. As shown in Table 4 , the F 1 score of the model BCCWJ is the best (97.2) among the four models. This shows that models trained on a large amount of data are more accurate, even if the writing styles are different.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 65, |
| "end": 72, |
| "text": "Table 4", |
| "ref_id": "TABREF7" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experiments 2: Impact of Writing Styles", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Third, we investigate the effect of pseudo corpora. Table 5 shows the result. The F 1 scores of the model P-BCCWJ for BCCWJ is 78.8. It is much worse than one of the model BCCWJ (98.2). This is an example of a false negative (FN) by the model P-BCCWJ.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 52, |
| "end": 59, |
| "text": "Table 5", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experiments 3: Effect of Pseudo Corpora", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "</s> . . . (It is necessary to build disaster prevention measures. </s>In the fire and disaster management agency, . . .) 5They were often wrong even in the almost obvious cases where periods \" \" were just before line breaks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments 3: Effect of Pseudo Corpora", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "The F 1 scores of the models P-BCCWJ and P-Jalan are respectively 94.8 and 92.8. Though they are better than one of the model Jalan-F (90.2), worse than one of the model Jalan-F+A (95.1).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments 3: Effect of Pseudo Corpora", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "These results suggest that although a sentence boundary detector with pseudo-corpus could achieve moderate performance, we can obtain better detectors by training with annotated corpora.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experiments 3: Effect of Pseudo Corpora", |
| "sec_num": "4.4" |
| }, |
| { |
| "text": "We implemented sentence boundary detectors by using BERT and revealed the following facts:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 It is possible to train a sentence boundary detector on line breaks with annotated corpora.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 Training with much annotation data is effective even for texts in another writing style.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 Although it is possible to train a sentence boundary detector even with pseudo-corpus to some extent, more performance will be gained by training with annotated corpora.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "There are two main issues that we need to address in the future. The first issue is to do is to use active learning to increase the number of learning examples and improve accuracy. The second issue is to perform SBD for other atypical sentence boundary expressions other than line breaks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "5" |
| }, |
| { |
| "text": "https://stanfordnlp.github.io/ stanza/ 4 https://universaldependencies.org/ treebanks/ja_gsd/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "In this paper, we removed all line breaks at the end of documents because they are obvious sentence boundaries. Additionally, if there is a series of line breaks or a space before or after a line break, we replaced it with a single line break.6 We only used the \"core\" in BCCWJ. Its annotation is manual while no manual correction is performed for \"noncore.\"7 https://www.jalan.net 8 The \"F\" is an abbreviation for \"full annotation.\"", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "https://github.com/taku910/mecab 12 https://github.com/huggingface/ transformers13 We did a preliminary experiment with binary labels \"Sentence boundary\" (SB) and \"Not sentence boundary\" (NSB), but it was low performance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "We recognize Dr. Yuki Arase at Osaka University for the many discussions and insightful comments. Furthermore, we thank the anonymous reviewers for their careful reading and valuable comments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding", |
| "authors": [ |
| { |
| "first": "Jacob", |
| "middle": [], |
| "last": "Devlin", |
| "suffix": "" |
| }, |
| { |
| "first": "Ming-Wei", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| } |
| ], |
| "year": 2019, |
| "venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| "volume": "1", |
| "issue": "", |
| "pages": "4171--4186", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jacob Devlin, Ming-Wei Chang, et al. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Hu- man Language Technologies, Volume 1, pages 4171- 4186.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures", |
| "authors": [ |
| { |
| "first": "Alex", |
| "middle": [], |
| "last": "Graves", |
| "suffix": "" |
| }, |
| { |
| "first": "J\u00fcrgen", |
| "middle": [], |
| "last": "Schmidhuber", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Neural Networks", |
| "volume": "18", |
| "issue": "5", |
| "pages": "602--610", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Alex Graves and J\u00fcrgen Schmidhuber. 2005. Frame- wise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures. Neural Networks, 18(5):602-610.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Pause and Stop Labeling for Chinese Sentence Boundary Detection", |
| "authors": [ |
| { |
| "first": "", |
| "middle": [], |
| "last": "Hen-Hsen", |
| "suffix": "" |
| }, |
| { |
| "first": "Hsin-Hsi", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Proceedings of the International Conference Recent Advances in Natural Language Processing 2011", |
| "volume": "", |
| "issue": "", |
| "pages": "146--153", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hen-Hsen Huang and Hsin-Hsi Chen. 2011. Pause and Stop Labeling for Chinese Sentence Boundary De- tection. In Proceedings of the International Con- ference Recent Advances in Natural Language Pro- cessing 2011, pages 146-153.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Correction of Sentence Boundaries in the Balanced Corpus of Contemporary Written Japanese DVD Version 1.0. NINJAL Research Papers", |
| "authors": [ |
| { |
| "first": "Hikari", |
| "middle": [], |
| "last": "Konishi", |
| "suffix": "" |
| }, |
| { |
| "first": "Takenori", |
| "middle": [], |
| "last": "Nakamura", |
| "suffix": "" |
| } |
| ], |
| "year": 2015, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "81--100", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hikari Konishi, Takenori Nakamura, et al. 2015. Cor- rection of Sentence Boundaries in the Balanced Cor- pus of Contemporary Written Japanese DVD Ver- sion 1.0. NINJAL Research Papers, (9):81-100.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data", |
| "authors": [ |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Lafferty", |
| "suffix": "" |
| }, |
| { |
| "first": "Andrew", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the 18th International Conference on Machine Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "282--289", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "John Lafferty, Andrew McCallum, et al. 2001. Condi- tional Random Fields: Probabilistic Models for Seg- menting and Labeling Sequence Data. In Proceed- ings of the 18th International Conference on Ma- chine Learning, pages 282-289.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Balanced Corpus of Contemporary Written Japanese", |
| "authors": [ |
| { |
| "first": "Kikuo", |
| "middle": [], |
| "last": "Maekawa", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Proceedings of the 6th Workshop on Asian Language Resources", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Kikuo Maekawa. 2008. Balanced Corpus of Contem- porary Written Japanese. In Proceedings of the 6th Workshop on Asian Language Resources.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Linguistic Innovations and Interactional Features of Casual Online Communication in Japanese", |
| "authors": [ |
| { |
| "first": "Yukiko", |
| "middle": [], |
| "last": "Nishimura", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Journal of Computer-Mediated Communication", |
| "volume": "9", |
| "issue": "1", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yukiko Nishimura. 2003. Linguistic Innovations and Interactional Features of Casual Online Communi- cation in Japanese. Journal of Computer-Mediated Communication, 9(1).", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Stanza: A Python Natural Language Processing Toolkit for Many Human Languages", |
| "authors": [ |
| { |
| "first": "Peng", |
| "middle": [], |
| "last": "Qi", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuhao", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| } |
| ], |
| "year": 2020, |
| "venue": "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations", |
| "volume": "", |
| "issue": "", |
| "pages": "101--108", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peng Qi, Yuhao Zhang, et al. 2020. Stanza: A Python Natural Language Processing Toolkit for Many Hu- man Languages. In Proceedings of the 58th Annual Meeting of the Association for Computational Lin- guistics: System Demonstrations, pages 101-108.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "The role of sentence closing as an emotional marker: A case of Japanese mobile phone e-mail. Discourse, Context & Media", |
| "authors": [ |
| { |
| "first": "Noboru", |
| "middle": [], |
| "last": "Sakai", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "", |
| "volume": "2", |
| "issue": "", |
| "pages": "149--155", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Noboru Sakai. 2013. The role of sentence closing as an emotional marker: A case of Japanese mobile phone e-mail. Discourse, Context & Media, 2(3):149-155.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Patent Claim Processing for Readability: Structure Analysis and Term Explanation", |
| "authors": [ |
| { |
| "first": "Akihiro", |
| "middle": [], |
| "last": "Shinmori", |
| "suffix": "" |
| }, |
| { |
| "first": "Manabu", |
| "middle": [], |
| "last": "Okumura", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of the ACL-2003 Workshop on Patent Corpus Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "56--65", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Akihiro Shinmori, Manabu Okumura, et al. 2003. Patent Claim Processing for Readability: Structure Analysis and Term Explanation. In Proceedings of the ACL-2003 Workshop on Patent Corpus Process- ing, pages 56-65.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "A Unified Tagging Approach to Text Normalization", |
| "authors": [ |
| { |
| "first": "Conghui", |
| "middle": [], |
| "last": "Zhu", |
| "suffix": "" |
| }, |
| { |
| "first": "Jie", |
| "middle": [], |
| "last": "Tang", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "688--695", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Conghui Zhu, Jie Tang, et al. 2007. A Unified Tagging Approach to Text Normalization. In Proceedings of the 45th Annual Meeting of the Association of Com- putational Linguistics, pages 688-695.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "TABREF2": { |
| "type_str": "table", |
| "html": null, |
| "num": null, |
| "content": "<table/>", |
| "text": "Statistics of corpus. LB means \"line break,\" and SB means \"sentence boundary.\" Each two letters for BCCWJ represents newspaper articles (PN), books (PB), magazines (PM), white papers (OW), QA texts in the Internet (OC) and blog texts (OY)." |
| }, |
| "TABREF4": { |
| "type_str": "table", |
| "html": null, |
| "num": null, |
| "content": "<table/>", |
| "text": "An example of input, output, and evaluation for sentence boundary detectors" |
| }, |
| "TABREF6": { |
| "type_str": "table", |
| "html": null, |
| "num": null, |
| "content": "<table><tr><td colspan=\"3\">: SBD Performance on line breaks by models</td></tr><tr><td colspan=\"3\">trained with annotated corpora</td></tr><tr><td>Test</td><td>Train</td><td>TP TN FP FN F 1</td></tr><tr><td colspan=\"3\">Jalan-A BCCWJ 210 46 1 11 97.2</td></tr><tr><td/><td>Jalan-F</td><td>188 39 8 33 90.2</td></tr><tr><td/><td>Jalan-A</td><td>204 27 20 17 91.7</td></tr><tr><td/><td colspan=\"2\">Jalan-F+A 202 45 2 19 95.1</td></tr></table>", |
| "text": "" |
| }, |
| "TABREF7": { |
| "type_str": "table", |
| "html": null, |
| "num": null, |
| "content": "<table><tr><td>Test</td><td>Train</td><td colspan=\"2\">TP TN FP FN F 1</td></tr><tr><td colspan=\"4\">BCCWJ P-BCCWJ 2,715 570 48 1,410 78.8</td></tr><tr><td/><td>P-Jalan</td><td colspan=\"2\">1,868 575 43 2,257 61.9</td></tr><tr><td colspan=\"3\">Jalan-A P-BCCWJ 200 46 1</td><td>21 94.8</td></tr><tr><td/><td>P-Jalan</td><td>192 46 1</td><td>29 92.8</td></tr></table>", |
| "text": "SBD Performance on line breaks in atypical writing style by models trained with annotated corpora" |
| } |
| } |
| } |
| } |