| { |
| "paper_id": "W06-0139", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T04:03:29.189231Z" |
| }, |
| "title": "Description of the NCU Chinese Word Segmentation and Named Entity Recognition System for SIGHAN Bakeoff 2006", |
| "authors": [ |
| { |
| "first": "Yu-Chieh", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "National Central University National Central University National Central University Taoyuan", |
| "location": { |
| "country": "Taiwan Taoyuan, Taiwan Taoyuan, Taiwan" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Jie-Chi", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "National Central University National Central University National Central University Taoyuan", |
| "location": { |
| "country": "Taiwan Taoyuan, Taiwan Taoyuan, Taiwan" |
| } |
| }, |
| "email": "yang@cl.ncu.edu.tw" |
| }, |
| { |
| "first": "Qian-Xiang", |
| "middle": [], |
| "last": "Lin", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "National Central University National Central University National Central University Taoyuan", |
| "location": { |
| "country": "Taiwan Taoyuan, Taiwan Taoyuan, Taiwan" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "Asian languages are far from most western-style in their non-separate word sequence especially Chinese. The preliminary step of Asian-like language processing is to find the word boundaries between words. In this paper, we present a general purpose model for both Chinese word segmentation and named entity recognition. This model was built on the word sequence classification with probability model, i.e., conditional random fields (CRF). We used a simple feature set for CRF which achieves satisfactory classification result on the two tasks. Our model achieved 91.00 in F rate in UPUC-Treebank data, and 78.71 for NER task.", |
| "pdf_parse": { |
| "paper_id": "W06-0139", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "Asian languages are far from most western-style in their non-separate word sequence especially Chinese. The preliminary step of Asian-like language processing is to find the word boundaries between words. In this paper, we present a general purpose model for both Chinese word segmentation and named entity recognition. This model was built on the word sequence classification with probability model, i.e., conditional random fields (CRF). We used a simple feature set for CRF which achieves satisfactory classification result on the two tasks. Our model achieved 91.00 in F rate in UPUC-Treebank data, and 78.71 for NER task.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "With the rapid expansion of text media sources such as news articles, technical reports, there is an increasing demand for text mining and processing. Among different cultures and countries, the Asian languages are far from the other languages, there is not an explicit boundary between words, for example Chinese. Similar to English, the preliminary step of most natural language processing is to \"tokenize\" each word. In Chinese, the word tokeniza-tion is also known as word segmentation or Chinese word tokenization.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "To support the above targets, it is necessary to detect the boundaries between words in a given sentence. In tradition, the Chinese word segmentation technologies can be categorized into three types, (heuristic) rule-based, machine learning, and hybrid. Among them, the machine learning-based techniques showed excellent performance in many research studies (Peng et al., 2004; Zhou et al., 2005; Gao et al., 2004) . This method treats the word segmentation problem as a sequence of word classification. The classifier online assigns either \"boundary\" or \"non-boundary\" label to each word by learning from the large annotated corpora. Machine learning-based word segmentation method is quite similar to the word sequence inference techniques, such as part-of-speech (POS) tagging, phrase chunking (Wu et al., 2006a) and named entity recognition (Wu et al., 2006b) .", |
| "cite_spans": [ |
| { |
| "start": 358, |
| "end": 377, |
| "text": "(Peng et al., 2004;", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 378, |
| "end": 396, |
| "text": "Zhou et al., 2005;", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 397, |
| "end": 414, |
| "text": "Gao et al., 2004)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 797, |
| "end": 815, |
| "text": "(Wu et al., 2006a)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 845, |
| "end": 863, |
| "text": "(Wu et al., 2006b)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we present a prototype for Chinese word segmentation and named entity recognition based on the word sequence inference model. Unlike previous researches (Zhou et al., 2005; Shi, 2005) , we argue that without using the word segmentation information, Chinese named entity recognition task can also be viewed as a variant word segmentation technique. Therefore, the two tasks can be accomplished without adapting the word sequence inference model. The preliminary experimental result show that in the word segmentation task, our method can achieve 91.00 in F rate for the UPUC Chinese Treebank data, while it at-tends 78.76 F rate for the Microsoft Chinese named entity recognition task.", |
| "cite_spans": [ |
| { |
| "start": 168, |
| "end": 187, |
| "text": "(Zhou et al., 2005;", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 188, |
| "end": 198, |
| "text": "Shi, 2005)", |
| "ref_id": "BIBREF5" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The rest of this paper is organized as follows. Section 2 describes the word sequence inference model and the used learner. Experimental result and evaluations are reported in section 3. Finally, in section 4, we draw conclusion and future remarks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this section, we firstly describe the overall system architecture for the word segmentation and named entity recognition tasks. In section 2.2, the employed classification model-conditional random fields (CRF) is then presented.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Description", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Similar to English text chunking (Ramshaw and Marcus, 1995; Wu et al., 2006a) , the word sequence classification model aims to classify each word via encoding its context features. An example can be shown in Figure 1 . In Figure1, the model is classifying the Chinese character \"\u570b\" (country). The second row in Figure 1 means the corresponding category of each in the word-segmentation (WS) task, while the third row indicates the class in the named entity recognition (NER) task. For the WS task, there are only two word types, B-CP (Begin of Chinese phrase) and I-CP (Interior of Chinese phrase). In contrast, the word types in the NER task depend on the pre-defined named class. For example, both in MSR and CityU datasets, person, location, and organization should be identified. In this paper, we used the similar IOB2 representation style (Wu et al., 2006a) to express the Chinese word structures. By encoding with IOB style, both WS and NER problems can be viewed as a sequence of word classification. During testing, we seek to find the optimal word type for each Chinese character. These types strongly reflect the actual word boundaries for Chinese words or named entity phrases.", |
| "cite_spans": [ |
| { |
| "start": 33, |
| "end": 59, |
| "text": "(Ramshaw and Marcus, 1995;", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 60, |
| "end": 77, |
| "text": "Wu et al., 2006a)", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 845, |
| "end": 863, |
| "text": "(Wu et al., 2006a)", |
| "ref_id": "BIBREF6" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 208, |
| "end": 216, |
| "text": "Figure 1", |
| "ref_id": "FIGREF1" |
| }, |
| { |
| "start": 311, |
| "end": 319, |
| "text": "Figure 1", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Word Sequence Classification", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "To effect classify each character, in this paper, we employ 13 feature templates to capture the context information of it. Table 1 lists the adopted feature templates. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 123, |
| "end": 130, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Word Sequence Classification", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "W -2 \uf9b4 W 0 + W +1 \u570b+\u611b W -1 \u4e2d W +1 + W +2 \u611b+\uf914 W 0 \u570b W +1 + W +2 \u611b+\uf914 W +1 \u611b W -2 +W -1 +W 0 \uf9b4+\u4e2d+\u570b W +2 \uf914 W -1 +W 0 +W +1 \u4e2d+\u570b+\u611b W -2 + W -1 \uf9b4+\u4e2d W 0 +W +1 +W +2 \u570b+\u611b+\uf914 W -1 + W 0 \u4e2d+\u570b", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Word Sequence Classification", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Conditional random field (CRF) was an extension of both Maximum Entropy Model (MEMs) and Hidden Markov Models (HMMs) that was firstly introduced by (Lafferty et al., 2001) . CRF defined conditional probability distribution P(Y|X) of given sequence given input sentence where Y is the \"class label\" sequence and X denotes as the observation word sequence. A CRF on (X,Y) is specified by a feature vector F of local context and the corresponding feature weight \u03bb. The F can be treated as the combination of state transition and observation value in conventional HMM. To determine the optimal label sequence, the CRF uses the following equation to estimate the most probability. The most probable label sequence y can be efficiently extracted via the Viterbi algorithm. However, training a CRF is equivalent to estimate the parameter set\u03bbfor the feature set. In this paper, we directly use the quasi-Newton L-BFGS 1 method (Nocedal and Wright, 1999) to iterative update the parameters.", |
| "cite_spans": [ |
| { |
| "start": 148, |
| "end": 171, |
| "text": "(Lafferty et al., 2001)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 920, |
| "end": 946, |
| "text": "(Nocedal and Wright, 1999)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conditional Random Fields", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "We evaluated our model in the close track on UPUC Chinese Treebank for Chinese word segmentation task, and CityU corpus for Chinese NER task. Both settings are the same for the two tasks. The evaluations of the two tasks were mainly measured by the three metrics, namely recall, precision, and f1-measurement. However, the evaluation style for the NER and WS is quite different. In WS, participant should reformulate the testing data into sentence level whereas the NER was evaluated in the token-level. Table 2 lists the results of the two tasks with our preliminary model. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 504, |
| "end": 511, |
| "text": "Table 2", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dataset and Evaluations", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "To explore the effectiveness of our method, we go on extend our model to the other three tasks for the WS track, namely CityU, MSR. Table3 shows the experimental results of our model in the all close WS track except for CKIP corpus. These results do not officially provided by the SIGHAN due to the time limitation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Result on Word Segmentation Task", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "1 http://www-unix.mcs.anl.gov/tao/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Result on Word Segmentation Task", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In the second experiment, we focus on directly adapting our method for the NER track. Table 4 lists the experimental result of our method in the CityU and MSR datasets. It is worth to note that due to the different evaluation style in NER tracks, our tokenization rules did not consistent with the SIGHAN provided testing tokens. Our preliminary tokenization rules produced 371814 characters for the testing data, while there are 364356 tokens in the official provided testing set. Such a big trouble deeply earns the actual performance of our model. To propose a reliable and actual result, we directly evaluate our method in the official provided testing set again. As shown in Table 4 , the our method achieved 0.787 in F rate with non-correct version. In contrast, after correcting the Chinese tokenization rules as well as SIGHAN official provided tokens, our method significantly improved from 0.787 to 0.868. Similarly, our method performed very on the MSR track which reached 0.818 in F rate. ", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 86, |
| "end": 93, |
| "text": "Table 4", |
| "ref_id": "TABREF3" |
| }, |
| { |
| "start": 680, |
| "end": 687, |
| "text": "Table 4", |
| "ref_id": "TABREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Experimental Result on Named Entity Recognition Task", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "Chinese word segmentation is the most important foundations for many Chinese linguistic technologies such as text categorization and information retrieval. In this paper, we present simple Chinese word segmentation and named entity recognition models based on the conventional sequence classification technique. The main focus of our work is to provide a light-weight and simple model that could be easily ported to different domains and languages. Without any prior knowledge and rules, such a simple technique shows satisfactory results on both word segmentation and named entity recognition tasks. To reach state-of-the-art this model still needs to employed more detail feature engines and analysis. In the future, one of the main directions is to extend this model toward full unsuper-vised learning from large un-annotated text. Mining from large unlabeled data have been showed benefits to improve the original accuracy. Thus, not only the more stochastic feature analysis, but also adjust the learner from unlabeled data are important future remarks.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions and Future Work", |
| "sec_num": "4" |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Conditional Random Field: Probabilistic models for segmenting and labeling sequence data", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Lafferty", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Pereira", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the International Conference on Machine Learning", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lafferty, J., McCallum, A., and Pereira, F. 2001. Conditional Random Field: Probabilistic models for segmenting and labeling sequence data. In Pro- ceedings of the International Conference on Ma- chine Learning.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Adaptive Chinese word segmentation", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Gao", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [ |
| "N" |
| ], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "X", |
| "middle": [], |
| "last": "Xia", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Qin", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings the 41st Annual Meeting of the Association for Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "21--26", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Gao, J., Wu, A., Li, M., Huang, C. N., Li, H., Xia, X., and Qin, H. 2004. Adaptive Chinese word segmen- tation. In Proceedings the 41st Annual Meeting of the Association for Computational Linguistics, pp. 21-26.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Text chunking using transformation-based learning", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Lance", |
| "suffix": "" |
| }, |
| { |
| "first": "Mitchell", |
| "middle": [ |
| "P" |
| ], |
| "last": "Ramshaw", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Marcus", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Proceedings of the 3rd Workshop on Very Large Corpora", |
| "volume": "", |
| "issue": "", |
| "pages": "82--94", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lance A. Ramshaw and Mitchell P. Marcus. 1995. Text chunking using transformation-based learning. In Proceedings of the 3rd Workshop on Very Large Corpora, pages 82-94.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Numerical optimization", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Nocedal", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Wright", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nocedal, J., and Wright, S. 1999. Numerical optimi- zation. Springer.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Chinese segmentation and new word detection using conditional random fields", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Peng", |
| "suffix": "" |
| }, |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Feng", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Mccallum", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Porceedings of the Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "562--568", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Peng, F., Feng, F., and McCallum, A. 2004. Chinese segmentation and new word detection using condi- tional random fields. In Porceedings of the Com- putational Linguistics, pp. 562-568.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Chinese Word Segmentation Based On Direct Maximum Entropy Model", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Shi", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the Fourth SIGHAN Workshop on Chinese Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shi, W. 2005. Chinese Word Segmentation Based On Direct Maximum Entropy Model. In Proceedings of the Fourth SIGHAN Workshop on Chinese Language Processing.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "A general and multi-lingual phrase chunking model based on masking method", |
| "authors": [ |
| { |
| "first": "Y", |
| "middle": [ |
| "C" |
| ], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [ |
| "H" |
| ], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [ |
| "S" |
| ], |
| "last": "Lee", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Lecture Notes in Computer Science (LNCS): Computational Linguistics and Intelligent Text Processing", |
| "volume": "3878", |
| "issue": "", |
| "pages": "144--155", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wu, Y. C., Chang, C. H. and Lee, Y. S. 2006a. A general and multi-lingual phrase chunking model based on masking method. Lecture Notes in Com- puter Science (LNCS): Computational Linguistics and Intelligent Text Processing, 3878: 144-155.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Extracting named entities using support vector machines", |
| "authors": [ |
| { |
| "first": "Y", |
| "middle": [ |
| "C" |
| ], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [ |
| "K" |
| ], |
| "last": "Fan", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [ |
| "S" |
| ], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [ |
| "J" |
| ], |
| "last": "Yen", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Lecture Notes in Bioinformatics (LNBI): Knowledge Discovery in Life Science Literature", |
| "volume": "", |
| "issue": "", |
| "pages": "91--103", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wu, Y. C., Fan, T. K., Lee Y. S. and Yen, S. J. 2006b. Extracting named entities using support vector machines,\" Lecture Notes in Bioinformatics (LNBI): Knowledge Discovery in Life Science Literature, (3886): 91-103.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "The Exploration of Deterministic and Efficient Dependency Parsing", |
| "authors": [ |
| { |
| "first": "Y", |
| "middle": [ |
| "C" |
| ], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [ |
| "S" |
| ], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "Yang", |
| "middle": [], |
| "last": "", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "C" |
| ], |
| "last": "", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the 10 th Conference on Natural Language Learning (CoNLL)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wu, Y. C., Lee, Y. S., and Yang, J. C. 2006c. The Exploration of Deterministic and Efficient De- pendency Parsing. In Proceedings of the 10 th Con- ference on Natural Language Learning (CoNLL).", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "A Hybrid Approach to Chinese Word Segmentation around CRFs", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "X", |
| "middle": [], |
| "last": "Dai", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Ni", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of the Fourth SIGHAN Workshop on Chinese Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhou, J., Dai, X., Ni, R., Chen, J. 2005. .A Hybrid Approach to Chinese Word Segmentation around CRFs. In Proceedings of the Fourth SIGHAN Workshop on Chinese Language Processing.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "uris": null, |
| "text": "word phrase LOC: Location ORG: Organization O: Non-named entity word", |
| "type_str": "figure", |
| "num": null |
| }, |
| "FIGREF1": { |
| "uris": null, |
| "text": "Sequence of word classification model", |
| "type_str": "figure", |
| "num": null |
| }, |
| "TABREF0": { |
| "text": "", |
| "content": "<table><tr><td>Feature</td><td>Examples Feature Type Examples</td></tr><tr><td>Type</td><td/></tr></table>", |
| "type_str": "table", |
| "html": null, |
| "num": null |
| }, |
| "TABREF1": { |
| "text": "", |
| "content": "<table><tr><td/><td>Dataset</td><td>F1-measure</td></tr><tr><td>Word segmentation</td><td>UPUC</td><td>91.00</td></tr><tr><td>Named entity recognition</td><td>CityU</td><td>78.71</td></tr></table>", |
| "type_str": "table", |
| "html": null, |
| "num": null |
| }, |
| "TABREF2": { |
| "text": "", |
| "content": "<table><tr><td>Closed Task</td><td>CityU</td><td>MSR</td><td>UPUC</td></tr><tr><td>Recall</td><td>0.958</td><td>0.940</td><td>0.917</td></tr><tr><td>Precision</td><td>0.926</td><td>0.906</td><td>0.904</td></tr><tr><td>F-measure</td><td>0.942</td><td>0.923</td><td>0.910</td></tr></table>", |
| "type_str": "table", |
| "html": null, |
| "num": null |
| }, |
| "TABREF3": { |
| "text": "", |
| "content": "<table><tr><td colspan=\"2\">Closed Task City (official</td><td>City</td><td>MSR</td></tr><tr><td/><td>result)</td><td>(correct)</td><td/></tr><tr><td>Recall</td><td>0.697</td><td>0.931</td><td>0.752</td></tr><tr><td>Precision</td><td>0.935</td><td>0.814</td><td>0.896</td></tr><tr><td>F-measure</td><td>0.787</td><td>0.868</td><td>0.818</td></tr></table>", |
| "type_str": "table", |
| "html": null, |
| "num": null |
| } |
| } |
| } |
| } |