| { |
| "paper_id": "U05-1026", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T03:08:21.072299Z" |
| }, |
| "title": "A Dual-Iterative Method for Concept-Word Acquisition from Large-Scale Chinese Corpora", |
| "authors": [ |
| { |
| "first": "Guogang", |
| "middle": [], |
| "last": "Tian", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Key Laboratory of Intelligent Information Processing", |
| "institution": "Chinese Academy of Sciences Graduate University of the Chinese Academy of Sciences", |
| "location": { |
| "postCode": "100080", |
| "settlement": "Beijing", |
| "country": "China" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Cungen", |
| "middle": [], |
| "last": "Cao", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "Key Laboratory of Intelligent Information Processing", |
| "institution": "", |
| "location": { |
| "postCode": "100080", |
| "country": "China" |
| } |
| }, |
| "email": "cgcao@ict.ac.cn" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper proposes a dual-iterative method, a hierarchical inner and outer iteration method (HIO), to acquire concept words from a large-scale, un-segmented Chinese corpus. It has two levels of iteration: the EM-CLS algorithm and the Viterbi-C/S algorithm constitute the inner iteration for generating concept words, and the concept word validation constitutes the outer iteration together with the concept word generation. Through multiple iterations, it integrates the concept word generation and validation into a uniform acquisition process. In the process of acquisition, the HIO method can cope with the problem of over-segmentation, over-combination and data sparseness. The experimental result shows that the HIO method is valid for concept word acquisition that can simultaneously increase the precision and recall rate of concept word acquisition.", |
| "pdf_parse": { |
| "paper_id": "U05-1026", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper proposes a dual-iterative method, a hierarchical inner and outer iteration method (HIO), to acquire concept words from a large-scale, un-segmented Chinese corpus. It has two levels of iteration: the EM-CLS algorithm and the Viterbi-C/S algorithm constitute the inner iteration for generating concept words, and the concept word validation constitutes the outer iteration together with the concept word generation. Through multiple iterations, it integrates the concept word generation and validation into a uniform acquisition process. In the process of acquisition, the HIO method can cope with the problem of over-segmentation, over-combination and data sparseness. The experimental result shows that the HIO method is valid for concept word acquisition that can simultaneously increase the precision and recall rate of concept word acquisition.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Concept word acquisition is an important research in knowledge acquisition from text (KAT) (Cao and Sui, 2003) , and it is also the foundation of ontology learning (Maedche, 2002) . Its main purpose is to acquire plentiful concept words from text corpora. It is very similar to unknown word recognition (Chen and Bai, 1998) , (Feng, Chen, et al., 2004) and term extraction (Bourigault and Jacquemin, 1999) . However, there are subtle distinctions among these three researches. Generally, concept word can be classified into three types: proper name, compound word and derived word. Except for these three word types, unknown word recognition also identifies numeric-type compounds, and it does not concern known words listed in a dictionary. Term extraction (Bourigault and Jacquemin, 1999) mainly processes domain texts, and often extracts commonly used professional terms from a specific domain text corpus. Fu and Luke (2003) proposed a two-stage Chinese segmentation system. At the first stage, it segmented the input text according to known words on the basis of 2-gram statistical model, and then identified unknown words at the second stage using a hybrid method which consisted of word context, word composition and word juncture model. Yang and Li (2003) proposed a heuristic method that it generated five rules using mutual information and significance estimation to extract unknown word. Peng and Schuurmans (2001) proposed an unsupervised training method to build probability models that accurately segmented Chinese character sequences into words. It used successive EM phases to learn a good probability model over character strings, and then prunes the model with a mutual information selection criterion to obtain a more accurate word lexicon. Wu (2000, 2002) proposed a likelihood ratio method to extract possible unknown words or phrases defined by PLUs (phrase-like-unit). The final PLU was decided by two principles of overlap competition and inclusion competition. Nagao and Mori (1994) proposed a rapid n-gram extraction method to extract adjacent substrings with same prefix in an ordered prefix table. It was noted that it was an affix method intrinsically.", |
| "cite_spans": [ |
| { |
| "start": 91, |
| "end": 110, |
| "text": "(Cao and Sui, 2003)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 164, |
| "end": 179, |
| "text": "(Maedche, 2002)", |
| "ref_id": "BIBREF14" |
| }, |
| { |
| "start": 303, |
| "end": 323, |
| "text": "(Chen and Bai, 1998)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 326, |
| "end": 352, |
| "text": "(Feng, Chen, et al., 2004)", |
| "ref_id": "BIBREF9" |
| }, |
| { |
| "start": 373, |
| "end": 405, |
| "text": "(Bourigault and Jacquemin, 1999)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 758, |
| "end": 790, |
| "text": "(Bourigault and Jacquemin, 1999)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 910, |
| "end": 928, |
| "text": "Fu and Luke (2003)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 1245, |
| "end": 1263, |
| "text": "Yang and Li (2003)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 1399, |
| "end": 1425, |
| "text": "Peng and Schuurmans (2001)", |
| "ref_id": null |
| }, |
| { |
| "start": 1760, |
| "end": 1775, |
| "text": "Wu (2000, 2002)", |
| "ref_id": null |
| }, |
| { |
| "start": 1986, |
| "end": 2007, |
| "text": "Nagao and Mori (1994)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "From these above works, we can summarize that there are two kinds of method to identify or acquire unknown words, that is, the non-iterative statistical method and the affix method. The non-iterative unknown word recognition (Fu and Luke, 2003) , (Yang and Li, 2003) , (Peng and Schuurmans, 2001) , Wu, 2000, 2002) , (Zhang, Lv, et al., 2003 ) usually adopts n-gram statistical model that is combined with segmentation and combination operation to identify unknown words. It can deal with over-segmentation, but can not tackle over-combination. In addition, the length of unknown word must be restricted in order to ensure system performance. The acquired unknown words are often 2-grams, 3-grams and 4-grams. The affix method (Nagao and Mori, 1994) has even more limits. For example, it can not deal with unknown words that have not obvious affix features, and it can not use contextual information of unknown words, either.", |
| "cite_spans": [ |
| { |
| "start": 225, |
| "end": 244, |
| "text": "(Fu and Luke, 2003)", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 247, |
| "end": 266, |
| "text": "(Yang and Li, 2003)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 269, |
| "end": 296, |
| "text": "(Peng and Schuurmans, 2001)", |
| "ref_id": null |
| }, |
| { |
| "start": 299, |
| "end": 314, |
| "text": "Wu, 2000, 2002)", |
| "ref_id": null |
| }, |
| { |
| "start": 317, |
| "end": 343, |
| "text": "(Zhang, Lv, et al., 2003 )", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "This paper, motivated by the work of Chang and Su (1997) and Liu, Zhang, et al. (2004) presents a hierarchical inner and outer iteration method to acquire concept words from a large-scale, un-segmented Chinese text corpus. It has two levels of iteration which involves concept word generation and validation. It makes some extension on EM algorithm and Viterbi algorithm which make up the concept word generation. The concept word validation combines mutual information and context entropy into a validation criterion. These two levels of iteration can simultaneously increase the precision and recall rate of concept word acquisition.", |
| "cite_spans": [ |
| { |
| "start": 37, |
| "end": 56, |
| "text": "Chang and Su (1997)", |
| "ref_id": "BIBREF10" |
| }, |
| { |
| "start": 61, |
| "end": 86, |
| "text": "Liu, Zhang, et al. (2004)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "The main contribution of this paper is that it proposes a HIO method for concept word acquisition. The HIO method unifies concept word generation and validation into a consecutively iterative process so that it can increase precision and recall simultaneously. The rest of this paper is organized as follows: Section 2 presents the HIO method. Concept word generation is discussed in section 2.1, and concept word validation is discussed in section 2.2. The whole HIO algorithm is presented in Section 2.3. The experiment result and error analysis are provided in section 3. Section 4 concludes this paper and outlines the future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1." |
| }, |
| { |
| "text": "The HIO method (a Hierarchical Inner and Outer iteration method) has two levels of iteration, that is, the inner iteration and the outer iteration. The alternation of EM-CLS and Viterbi-C/S algorithm constitutes the inner iteration of the HIO -concept word generation, and concept word validation constitutes the outer iteration of the HIO. The basic structure of the HIO method is illustrated in Fig. 1 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 397, |
| "end": 403, |
| "text": "Fig. 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "The HIO Method", |
| "sec_num": "2." |
| }, |
| { |
| "text": "The HIO method can cope with the two primary problems in concept word acquisition: over-segmentation and over-combination. Data sparseness is one of common problems in statistical language processing. Concept word acquisition is not the exception. In the acquisition process, it may produce the sparse data. Katz smoothing is applied in the HIO method to smooth sparse data and reduce their effect on concept word acquisition.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The HIO Method", |
| "sec_num": "2." |
| }, |
| { |
| "text": "The EM-CLS algorithm, which is based on EM (expectation maximization) algorithm, estimates generated terms' probability distribution and identifies their types in a large corpus. Figueiredo, 2004) , (Prescher, 2003 ) is a common method for estimating maximum-likelihood when missing data are present. It has two steps: E-step (expectation step) and M-step (maximization step). Given the observed data x and the current parameter estimation ) t ( \u03b8 , E-step computes the conditional expectation (with respect to the missing data y) of the logarithm of a complete posteriori probability function, logp (y,\u03b8|x) . Usually E-step is called as Q function, as illustrated in (1). Equation 2shows the M-step of EM algorithm. M-step chooses the parameters which can maximize Q function as the estimated parameters. Through consecutive iterations of E-Step and M-Step, EM algorithm can get stabilized parameters.", |
| "cite_spans": [ |
| { |
| "start": 179, |
| "end": 196, |
| "text": "Figueiredo, 2004)", |
| "ref_id": null |
| }, |
| { |
| "start": 199, |
| "end": 214, |
| "text": "(Prescher, 2003", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 600, |
| "end": 607, |
| "text": "(y,\u03b8|x)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "EM-CLS algorithm", |
| "sec_num": "2.1.1" |
| }, |
| { |
| "text": "Step:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "\u222b + = + \u221d \u2261 dy x y p x y p p x x y p E p x x y p E Q t t t t ) | , ( log ) , | ( ) ( log ] , | ) | , ( [log ) ( log ] , | ) | , ( [log ) | ( ) ( ) ( ) \u03b8 \u03b8 \u03b8 = + (2) An un-segmented corpus is denoted as C={C 1 , C 2 , \u2026 ,C n } where C i (1 \u2264 i \u2264 n) represents an", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "un-segmented sentence. After segmentation, C is converted into the segmented corpus denoted as S={S 1 , S 2 , \u2026, S n } where S i (1 \u2264 i \u2264 n) is a segmentation of C i . The generated candidate terms 1 are grouped into a set denoted as T={t 1 , t 2 , \u2026, t m }, where t j (1\u2264j\u2264m) is the generated candidate term.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "If C i is taken as the observed data, S i as the missing data, we can estimate the maximum-likelihood of term t j with the EM algorithm which is deemed as its probability distribution in the corpus C. Equation 3shows the probability estimation of term t j . In 3, S i * denotes the optimal segmentation of sentence C i , which can be achieved by the Viterbi-C/S algorithm (to be discussed in the next section). f(t j , S i ) denotes the frequency of term t j in sentence S i .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "} | ) ( { ), (^T t t p T t p t j j j j \u2208 = = .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "After estimating the probability of term t j , we still have to judge to which type it belongs. The candidate term has three types that are normal term, over-segmented term and over-combined term. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "i i i j n i t i i i j m j n i t i i i j n i t i i i j t j T C S p S t f T C S p S t f T C S p S t f T C S p S t f t 1 1 ) ( * 1 ) ( * 1 1 ) ( * 1 ) ( * ) 1 ( ) | , ( ) , ( ) | , ( ) , ( ) , | ( ) , ( ) , | ( ) , (", |
| "eq_num": "(3)" |
| } |
| ], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "If a concept word (or meaningful word) is segmented into several components, it is called over-segmented term. For example, \u9ad8 \u8840 \u7cd6 (hyperglycemia) is possibly spitted into \u9ad8 (high) and \u8840\u7cd6 (blood sugar).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "If a word is combined with another word, but their combination is not a concept word (or meaningful word), it is called over-combined term, such as \u4f46\u4e5f (but also).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "Equation 4can assign a type label to t j , which is denoted by CLS(t j ). H (i) denotes the sentence type label.", |
| "cite_spans": [ |
| { |
| "start": 74, |
| "end": 79, |
| "text": "H (i)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "| ) ( | | } ) ( | { | max arg ) ( ) ( ) ( j i k j k k H j t Sen H S t Sen S S t CLS i \u2208 \u2227 \u2208 = (4) In (4), Sen(t j )={S i | t j \u2208S i },", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "E-", |
| "sec_num": null |
| }, |
| { |
| "text": "The Viterbi-C/S algorithm dynamically segments a 1 The meaning of \"term\" here is different from the meaning of \"term\" in term extraction. Here term refers to ordinary word. corpus using the estimated probability of candidate terms and executes combination and segmentation operations on ambiguous terms in order to achieve the optimal segmentation. After completing corpus segmentation, it judges if a sentence contains overlap ambiguity or combination ambiguity. Given a segmented sentence S i , S i =t 1 t 2 \u2026t k (1\u2264j\u2264 k,t j \u2208T), it is assumed that terms are independent each other, the likelihood of sentence S i is defined as:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "\u220f = = k j j i t p S p 1 ) ( ) ( (5)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "Definition 1: It is the optimal segmentation that its likelihood is maximal among all segmentations of a sentence. The optimal segmentation is denoted as", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "S i * ) | , ( max arg ) , | ( max arg* T C S p T C S p S i i S i i S i i i = = (6)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "Like candidate terms, segmented sentences are also classified into three types: normal sentence (N-Sen), overlap-ambiguity sentence (OA-Sen), and combination-ambiguity sentence (CA-Sen).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "If a segmented sentence contains over-combined terms, it is considered as an OA-Sen.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "If a segmented sentence contains over-segmented terms, it is considered as a CA-Sen.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "It is observed that there is a direct correspondence between the type of candidate term and segmented sentence: normal term -N-Sen, over-segmented term -CA-Sen and over-combined term -OA-Sen.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "Definition 2: Segmented Density is defined as the number of segmented term in each length unit.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "For a sentence", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "S i , ) ( ) ( ) ( ) ( i i i i S length S NT S p S SD \u00d7 = (7). For a corpus S, \u2211 \u2211 \u2211 \u2208 \u2208 \u2208 \u00d7 = \u00d7 = S S i S S i i S S i i i i i S length S NT S p S length S NT S p S SD ) ( ) ( ) ( ) ( ) ( ) ( ) ( (8).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "In (7)-(8), NT(X) denotes the number of terms in sentence X, and length(Y) denotes the length of sentence Y.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "The type of segmented sentence is measured by (9). Setting a threshold range [r 1 , r 2 ] (r 1 <r 2 ), if CLS(S i )<r 1 , S i is a OA-Sen, if CLS(S i )>r 2 , S i is a CA-Sen, if r 1 \u2264CLS(S i )\u2264r 2 , S i is a N-Sen.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": ") ( SD ) ( SD ) ( S S S CLS i i = (9)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "We make an extension to the classical Viterbi algorithm (Rabiner, 1989) , thus get the Viterbi-C/S algorithm as illustrated in Fig. 2 .", |
| "cite_spans": [ |
| { |
| "start": 56, |
| "end": 71, |
| "text": "(Rabiner, 1989)", |
| "ref_id": "BIBREF12" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 127, |
| "end": 133, |
| "text": "Fig. 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "When segmenting a corpus, Viterbi-C/S binds combination and segmentation operations (C/S operation) on the selected terms according to their types. If over-segmented term is selected, combination operation is performed, if over-combined term is selected, segmentation operation is performed. These combination or segmentation operations on candidate term possibly causes data sparseness problem. So we use Katz Smoothing method (Goodman, 2001) as the smoothing strategy to eliminate sparse data.", |
| "cite_spans": [ |
| { |
| "start": 428, |
| "end": 443, |
| "text": "(Goodman, 2001)", |
| "ref_id": "BIBREF7" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": ") ... ( ) ... ( ) ... ( 1 2 1 1 1 1 i i n i i n i i i n i t t t p t t t t t p \u2212 + \u2212 \u2212 + \u2212 \u2212 + \u2212 = \u03b1", |
| "eq_num": "(10)" |
| } |
| ], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "In (10), t'=t i-n+1 \u2026t i-1 t i doesn't exist in candidate term set T, \u03b1 is a normalization constant.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Viterbi-C/S algorithm", |
| "sec_num": "2.1.2" |
| }, |
| { |
| "text": "The generated candidate terms need to be further validated to filter out ambiguous terms. The validation takes into considerations candidate term's composition and local context. The former is considered as a cohesion validation which adopts mutual information method (Sproat and Shih, 1990) . The latter is considered as an independence validation which adopts context entropy method (Tung and Lee, 1994) . So the concept word validation is the combination of mutual information and context entropy method.", |
| "cite_spans": [ |
| { |
| "start": 268, |
| "end": 291, |
| "text": "(Sproat and Shih, 1990)", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 385, |
| "end": 405, |
| "text": "(Tung and Lee, 1994)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Concept Word Validation", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "Viterbi-C/S Algorithm Input: un-segmented corpus C, candidate terms' probability estimation Output: the optimal segmentation and its type 1. selecting a sentence C i from corpus C; 2. selecting all possible candidate terms at the current position of sentence C i , which constitute a set denoted as T p ={t 1 p , t 2 p , t 3 p , \u2026}; 3. selecting a candidate term which has maximum-likelihood from the set T p as a possible segmented term of sentence C i , denoted as st; 4. performing the corresponding operation according to the type of term st a. if an over-segmented term, performing segmenting operation on it and re-estimating the likelihood of new term, goto (3); b. if an over-combined term, performing combining operation on it and re-estimating the likelihood of new term, goto (3); c. if a normal term, segmenting the sentence, computing the likelihood of current segmentation and moving the position pointer forwardly 5. repeating step 1-4, until all sentences in corpus are segmented 6. computing segmented density for corpus and sentences, and determining the type label of sentence Fig. 2 . The Viterbi-C/S Algorithm The basic assumptions of concept word validation are that:", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 1096, |
| "end": 1102, |
| "text": "Fig. 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Concept Word Validation", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "If a term is an over-combined term, it contains at least a division point where its cohesion degree must be low.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Concept Word Validation", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "If a term is an over-segmented term, its local context features in corpus must be weak.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Concept Word Validation", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "It is assumed that there is at most two division points in a validating term t v = c 1 c 2 \u2026c n .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Mutual Information", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "If t v l =c 1 c 2 \u2026c l \u2208T (1\u2264l<n) and t'= c 1 c 2 \u2026c l c l+1 \u2209T, t v l is called the maximal left substring of t v , and l is the left division point of t v .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Mutual Information", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "If t v r =c r c r+1 \u2026c n \u2208T (1<r\u2264n) and t'= c r-1 c r \u2026c n \u2209T, t v r is called the maximal right substring of t v , and r is the right division point of t v .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Mutual Information", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "(1) If l<r-1, t v has two division points, which is denoted as t", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Mutual Information", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "v =t v 2 =t v l t v m t v r ;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Mutual Information", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "(2) If l=r-1, t v has a division point, which is denoted as t", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Mutual Information", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "v =t v 1 =t v l t v r ;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Mutual Information", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "(3) If l\u2265r, t v has two possible divisions which are denoted as t", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Mutual Information", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "v =t v 1-L = t v l t v -l and t v 1-R = t v =t v -r t v r respectively. To case (1) ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( log ) ( ) ( 2 r v m v l v r v m v l v r v m v l v r v m v l v v v t p t t p t t p t p t p t p t p t t t p t MI t MI + + = = (11) To case (2), ) ( ) ( ) ( log ) ( ) ( 1 r v l v r v l v v v t p t p t t p t MI t MI = = (12) To case (3), )} ( ), ( min{ ) ( ; ) ( ) ( ) ( log ) ( ; ) ( ) ( ) ( log ) ( 1 1 1 1 R v L v v r v r v r v r v R v l v l v l v l v L v t MI t MI t MI t p t p t t p t MI t p t p t t p t MI \u2212 \u2212 \u2212 \u2212 \u2212 \u2212 \u2212 \u2212 = = = (13)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Mutual Information", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "Before computing the mutual information of the validating term, we above all identify to which type it belongs among the above case (1) to (3) and then adopt the corresponding equation (11) (12) (13) . Similarly, we still apply equation 10to deal with data sparseness problem.", |
| "cite_spans": [ |
| { |
| "start": 185, |
| "end": 189, |
| "text": "(11)", |
| "ref_id": null |
| }, |
| { |
| "start": 190, |
| "end": 194, |
| "text": "(12)", |
| "ref_id": null |
| }, |
| { |
| "start": 195, |
| "end": 199, |
| "text": "(13)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Mutual Information", |
| "sec_num": "2.2.1" |
| }, |
| { |
| "text": "It is assumed that t v is a validating term. Its left context is denoted as \u03b1={\u03b1 1 , \u03b1 2 , \u2026 ,\u03b1 l }, and its right context is denoted as \u03b2={\u03b2 1 , \u03b2 2 , \u2026 , \u03b2 r }. The left context entropy, right context entropy and context entropy of the validating term t v is defined in (14). ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Context Entropy", |
| "sec_num": "2.2.2" |
| }, |
| { |
| "text": ")} ( ), ( min{ ) ( ; ) ( log ) ( ) ( ; ) ( log ) ( ) ( v R v L v i v i v v R v i v i v", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Context Entropy", |
| "sec_num": "2.2.2" |
| }, |
| { |
| "text": "= \u2212 = \u2212 = \u2211 \u2211 \u2208 \u2208 \u03b2 \u03b2 \u03b1 \u03b1 \u03b2 \u03b2 \u03b1 \u03b1 (14)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Context Entropy", |
| "sec_num": "2.2.2" |
| }, |
| { |
| "text": "Table 1 lists the joint validation rules combining mutual information and context entropy criterions. th mi and th entr are thresholds we designate to mutual information and context entropy, respectively.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Context Entropy", |
| "sec_num": "2.2.2" |
| }, |
| { |
| "text": "The wrong candidates are removed from the dictionary. The other three types of terms are saved into the candidate dictionary again. After validating terms, concept word generation is restarted again and the concept word acquisition goes into the next iteration.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Context Entropy", |
| "sec_num": "2.2.2" |
| }, |
| { |
| "text": "The HIO Algorithm is illustrated in Fig.3 . 3. outputting all normal candidate terms as concept words set Fig.3 . The HIO Algorithm We adopt a 400M Chinese corpus extracted from web pages as the experimental corpus. Before running the HIO method, a series of preprocessing operations are performed, which involve recognizing special unknown words such as numeric-type words, English words, etc., acquiring concept words matching with specific context patterns, using forward maximum matching method to initially segment the corpus and estimating the initial probability of terms. We set the inner iteration to 10 times and the outer iteration to 5 times. When completing the HIO operations, we randomly select 2000 sentences from this corpus. After filtering out many common words such as auxiliary words, adjectives and adverbs, we get many concept words which have higher precision and recall rates as listed in Table 2 . We get 5387 unknown words in total 14392 terms, among which there are 833 bi-gram words, 1593 tri-gram words, 1049 four-gram words, 1196 five-gram words, 716 six-and over-six-gram words. Fig. 4 shows the effect of the inner and outer iteration on the precision and recall rate of concept word acquisition. It is observed that the precision and recall rate are both increased with the increase of iteration times.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 36, |
| "end": 41, |
| "text": "Fig.3", |
| "ref_id": null |
| }, |
| { |
| "start": 106, |
| "end": 111, |
| "text": "Fig.3", |
| "ref_id": null |
| }, |
| { |
| "start": 914, |
| "end": 921, |
| "text": "Table 2", |
| "ref_id": "TABREF3" |
| }, |
| { |
| "start": 1111, |
| "end": 1117, |
| "text": "Fig. 4", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "The HIO Algorithm", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "There are two types of errors produced in HIO: commission error and omission error (Yang and Li, 2004) . A commission error is that an acquired term is actually not a concept word, but the HIO considers it as a concept word. The reason is that every component of this term is common words and often occurs simultaneously. An omission error is that the HIO misses a concept word in the corpus. The reason is that one component of this word is more commonly used than the rest and the statistical feature of their combination is not prominent. However, the error distribution we get is contrary to the result of Yang and Li (2004) . The number of omission errors exceeds that of commission errors, especially in tri-gram concept word.", |
| "cite_spans": [ |
| { |
| "start": 83, |
| "end": 102, |
| "text": "(Yang and Li, 2004)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 610, |
| "end": 628, |
| "text": "Yang and Li (2004)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental Result and Error Analysis", |
| "sec_num": "3." |
| }, |
| { |
| "text": "This paper proposes a hierarchical inner and outer iteration method (HIO) for concept word acquisition. It can deal with the problem of over-segmentation, over-combination and data sparseness produced in the process of acquisition. Its prominent features involve:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions and Future Work", |
| "sec_num": "4." |
| }, |
| { |
| "text": "(1) The HIO method is the combination of the inner and outer iteration, which can increase the precision and recall rate of concept words acquisition simultaneously. (2) Concept word generation and validation are uniform and consistent in the HIO method. (3) The EM-CLS algorithm can classify candidate terms as well as estimate their probability distribution. (4) The Viterbi-C/S can perform segmenting and combining operations on terms while segmenting corpus. (5) HIO uses Katz smoothing to lessen data sparseness effect on concept word acquisition. Now we are going on a series of research on knowledge acquisition from text. The acquired knowledge types include concepts and their relations. Concept word acquisition is fundamental, which can provide essential support for other work in KAT research. We are also developing methods for acquiring relations, including isa, part-of and co-title.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions and Future Work", |
| "sec_num": "4." |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "This work is supported by the Natural Science Foundation (grant no. 60273019 and 60496326), and the National 973 Programme (grant no. 2003CB317008).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgement", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Segmenting Chinese Unknown Words by Heuristic Method", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Christopher", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [ |
| "W" |
| ], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "ICADL 2003", |
| "volume": "2911", |
| "issue": "", |
| "pages": "510--520", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Christopher C. Yang and K.W. Li. 2003. Segmenting Chinese Unknown Words by Heuristic Method, ICADL 2003, LNCS 2911, pp 510-520", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Error Analysis of Chinese Text Segmentation using Statistical Approach", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Christopher", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [ |
| "W" |
| ], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of 2004 Joint ACM/IEEE Conference on Digital Library (JCDL' 04)", |
| "volume": "", |
| "issue": "", |
| "pages": "256--257", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Christopher C. Yang and K.W. Li. 2004. Error Analysis of Chinese Text Segmentation using Statistical Approach. Proceedings of 2004 Joint ACM/IEEE Conference on Digital Library (JCDL' 04), Tucson, Arizona, USA, pp256-257", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Constructing Ontology and Knowledge Bases from Text", |
| "authors": [ |
| { |
| "first": "Cungen", |
| "middle": [], |
| "last": "Cao", |
| "suffix": "" |
| }, |
| { |
| "first": "Yuefei", |
| "middle": [], |
| "last": "Sui", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "20 th International Conference on Computer Processing of Oriental Languages", |
| "volume": "", |
| "issue": "", |
| "pages": "34--42", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Cungen Cao and Yuefei Sui. 2003. Constructing Ontology and Knowledge Bases from Text. 20 th International Conference on Computer Processing of Oriental Languages, Shenyang, China, pp 34-42", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "A Tutorial on the Expectation-Maximization Algorithm Including Maximum-Likelihood Estimation and EM Training of Probabilistic Context-Free Grammars, Presented at the 15th European Summer School in Logic, Language and Information", |
| "authors": [ |
| { |
| "first": "Detlef", |
| "middle": [], |
| "last": "Prescher", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Detlef Prescher. 2003. A Tutorial on the Expectation-Maximization Algorithm Including Maximum-Likelihood Estimation and EM Training of Probabilistic Context-Free Grammars, Presented at the 15th European Summer School in Logic, Language and Information (ESSLLI 2003), Vienna, Austria, August 18-29, 2003", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "TERM EXTRACTION + TERM CLUSTERING: An Integrated Platform for Computer-Aided Terminology", |
| "authors": [ |
| { |
| "first": "Didier", |
| "middle": [], |
| "last": "Bourigault", |
| "suffix": "" |
| }, |
| { |
| "first": "Christian", |
| "middle": [], |
| "last": "Jacquemin", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "Proceedings of EACL '99", |
| "volume": "", |
| "issue": "", |
| "pages": "15--22", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Didier Bourigault and Christian Jacquemin. 1999. TERM EXTRACTION + TERM CLUSTERING: An Integrated Platform for Computer-Aided Terminology. Proceedings of EACL '99, pp. 15-22", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Self-supervised Chinese Word Segmentation", |
| "authors": [], |
| "year": null, |
| "venue": "Advances in Intelligent Data Analysis (Proceedings of IDA-01)", |
| "volume": "", |
| "issue": "", |
| "pages": "238--247", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Self-supervised Chinese Word Segmentation. In Advances in Intelligent Data Analysis (Proceedings of IDA-01), pp 238-247", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "A Bit of Progress in Language Modeling", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [ |
| "T" |
| ], |
| "last": "Goodman", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Computer Speech and Language", |
| "volume": "", |
| "issue": "10", |
| "pages": "403--434", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Goodman, J. T. 2001. A Bit of Progress in Language Modeling. Computer Speech and Language, 2001 (10), pp 403-434", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "A two-stage statistical word segmentation system for Chinese", |
| "authors": [ |
| { |
| "first": "Guhong", |
| "middle": [], |
| "last": "Fu", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Luke", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Second SIGHAN Workshop on Chinese Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "156--159", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Guhong Fu and K.K Luke. 2003. A two-stage statistical word segmentation system for Chinese, Second SIGHAN Workshop on Chinese Language Processing, Sapporo, Japan, pp. 156-159", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Accessor Variety Criteria for Chinese Word Extraction, Computational Linguistics", |
| "authors": [ |
| { |
| "first": "Haodi", |
| "middle": [], |
| "last": "Feng", |
| "suffix": "" |
| }, |
| { |
| "first": "Kang", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "Xiaotie", |
| "middle": [], |
| "last": "Deng", |
| "suffix": "" |
| }, |
| { |
| "first": "Weimin", |
| "middle": [], |
| "last": "Zheng", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "", |
| "volume": "30", |
| "issue": "", |
| "pages": "75--93", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Haodi Feng, Kang Chen, Xiaotie Deng, Weimin Zheng. 2004. Accessor Variety Criteria for Chinese Word Extraction, Computational Linguistics, 30 (1), pp. 75-93", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "An Unsupervised Iterative Method for Chinese New Lexicon Extraction", |
| "authors": [ |
| { |
| "first": "Jing-Shin", |
| "middle": [], |
| "last": "Chang", |
| "suffix": "" |
| }, |
| { |
| "first": "Keh-Yih", |
| "middle": [], |
| "last": "Su", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Computational Linguistics and Chinese Language Processing", |
| "volume": "2", |
| "issue": "", |
| "pages": "97--148", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Jing-Shin Chang and Keh-Yih Su. 1997. An Unsupervised Iterative Method for Chinese New Lexicon Extraction, Computational Linguistics and Chinese Language Processing, 2 (2), pp 97-148", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Unknown word detection for Chinese by a corpus-based learning method", |
| "authors": [ |
| { |
| "first": "Ming-Hong", |
| "middle": [], |
| "last": "Keh-Jiann Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Bai", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "International Journal of Computational Linguistics and Chinese Language Processing", |
| "volume": "3", |
| "issue": "1", |
| "pages": "27--44", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Keh-Jiann Chen and Ming-Hong Bai. 1998. Unknown word detection for Chinese by a corpus-based learning method, International Journal of Computational Linguistics and Chinese Language Processing, 3(1):27-44", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition", |
| "authors": [ |
| { |
| "first": "Lawrence", |
| "middle": [ |
| "R" |
| ], |
| "last": "Rabiner", |
| "suffix": "" |
| } |
| ], |
| "year": 1989, |
| "venue": "Proceedings of the IEEE", |
| "volume": "77", |
| "issue": "2", |
| "pages": "257--286", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lawrence R. Rabiner. 1989. A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition. Proceedings of the IEEE, 77 (2), pp 257-286", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Chinese Lexical Analysis Using Cascaded Hidden Markov Model", |
| "authors": [ |
| { |
| "first": "Zhang", |
| "middle": [], |
| "last": "Liu Qun", |
| "suffix": "" |
| }, |
| { |
| "first": "Yu", |
| "middle": [], |
| "last": "Huaping", |
| "suffix": "" |
| }, |
| { |
| "first": "Cheng", |
| "middle": [], |
| "last": "Hongkui", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Xueqi", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Journal of Computer Research and Development", |
| "volume": "41", |
| "issue": "8", |
| "pages": "1421--1429", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Liu Qun, Zhang Huaping, Yu HongKui, and Cheng Xueqi. 2004. Chinese Lexical Analysis Using Cascaded Hidden Markov Model, Journal of Computer Research and Development, 41(8), pp. 1421-1429", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "A New Method of N-gram Statistics for Large Number of N and Automatic Extraction of Words and Phrases from Large Text Data of Japanese", |
| "authors": [ |
| { |
| "first": ";", |
| "middle": [], |
| "last": "Maedche", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Mori", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "Lecture Notes on the EM Algorithm", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Maedche. 2002. Ontology Learning for the Semantic Web. Kluwer Academic Publishers M\u00e1rio A. T, Figueiredo. 2004. Lecture Notes on the EM Algorithm, http://www.lx.it.pt/~mtf/learning/aboutEM.pdf Nagao, M and Mori,S. 1994. A New Method of N-gram Statistics for Large Number of N and Automatic Extraction of Words and Phrases from Large Text Data of Japanese. COLING-94", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "A Statistical Method for Finding Word Boundaries in Chinese Text", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Sproat", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Shih", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "Computer Processing of Chinese and Oriental Languages", |
| "volume": "", |
| "issue": "4", |
| "pages": "336--351", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sproat, R. and Shih, C. 1990. A Statistical Method for Finding Word Boundaries in Chinese Text, Computer Processing of Chinese and Oriental Languages, 1990 (4), pp.336-351", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Identification of Unknown Words from a Corpus", |
| "authors": [ |
| { |
| "first": "Cheng-Huang", |
| "middle": [], |
| "last": "Tung", |
| "suffix": "" |
| }, |
| { |
| "first": "Hsi-Jian", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "Computer Processing of Chinese and Oriental Languages", |
| "volume": "8", |
| "issue": "", |
| "pages": "131--145", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tung, Cheng-Huang and Hsi-Jian Lee. 1994. Identification of Unknown Words from a Corpus, Computer Processing of Chinese and Oriental Languages, Vol. 8, pp. 131-145", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Unknown Word and Phrase Extraction Using a Phrase-Like-Unit-Based Likelihood Ratio", |
| "authors": [ |
| { |
| "first": "Yusheng", |
| "middle": [], |
| "last": "Lai", |
| "suffix": "" |
| }, |
| { |
| "first": "Chunghsien", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "International Journal of Computer Processing of Oriental Languages", |
| "volume": "13", |
| "issue": "1", |
| "pages": "83--95", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yusheng Lai and ChungHsien Wu. 2000. Unknown Word and Phrase Extraction Using a Phrase-Like-Unit-Based Likelihood Ratio. International Journal of Computer Processing of Oriental Languages, 13 (1), pp 83-95", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Meaningful Term Extraction and Discriminative Term Selection in Text Categorization via Unknown-Word Methodology", |
| "authors": [ |
| { |
| "first": "Yu-Sheng", |
| "middle": [], |
| "last": "Lai", |
| "suffix": "" |
| }, |
| { |
| "first": "Chung-Hsien", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| } |
| ], |
| "year": 2002, |
| "venue": "ACM Transactions on Asian Language Information Processing", |
| "volume": "1", |
| "issue": "1", |
| "pages": "34--64", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Yu-sheng Lai and Chung-hsien Wu. 2002. Meaningful Term Extraction and Discriminative Term Selection in Text Categorization via Unknown-Word Methodology. ACM Transactions on Asian Language Information Processing, Vol. 1, No. 1, March 2002, pp 34-64", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "A Statistical Approach to Extract Chinese Chunk Candidates from Large Corpora", |
| "authors": [ |
| { |
| "first": "Zhang", |
| "middle": [], |
| "last": "Le", |
| "suffix": "" |
| }, |
| { |
| "first": "Lv", |
| "middle": [], |
| "last": "Xue-Qiang", |
| "suffix": "" |
| }, |
| { |
| "first": "Shen", |
| "middle": [], |
| "last": "Yan-Na", |
| "suffix": "" |
| }, |
| { |
| "first": "Yao", |
| "middle": [], |
| "last": "Tian-Shun", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "109--117", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Zhang Le, Lv Xue-qiang, Shen Yan-na and Yao Tian-shun. A Statistical Approach to Extract Chinese Chunk Candidates from Large Corpora. ICCPOL 2003, pp109-117, 2003", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "num": null, |
| "uris": null, |
| "text": "Fig.1. The Structure of HIO Method" |
| }, |
| "FIGREF2": { |
| "type_str": "figure", |
| "num": null, |
| "uris": null, |
| "text": "The Precision and Recall w.r.t. the Inner and Outer Iteration" |
| }, |
| "TABREF2": { |
| "content": "<table><tr><td colspan=\"3\">The Joint Validation of MI and Entropy</td></tr><tr><td>MI(t v )</td><td>Entr(t v )</td><td>Term type(t v )</td></tr><tr><td>\u2265th mi</td><td>\u2265th entr</td><td>correct candidate</td></tr><tr><td>\u2265th mi</td><td><th entr</td><td>over-segmented</td></tr><tr><td><th mi</td><td>\u2265th entr</td><td>over-combined</td></tr><tr><td><th mi</td><td><th entr</td><td>wrong candidate</td></tr></table>", |
| "html": null, |
| "text": "", |
| "type_str": "table", |
| "num": null |
| }, |
| "TABREF3": { |
| "content": "<table><tr><td/><td colspan=\"3\">The Experimental Result</td></tr><tr><td colspan=\"3\">length Count P (%)</td><td>R (%)</td></tr><tr><td>2</td><td colspan=\"2\">7782 92.2</td><td>83.4</td></tr><tr><td>3</td><td colspan=\"2\">2234 86.1</td><td>69.0</td></tr><tr><td>4</td><td colspan=\"2\">1627 89.3</td><td>70.4</td></tr><tr><td>5</td><td colspan=\"2\">1893 94.7</td><td>60.3</td></tr><tr><td>\u22656</td><td>856</td><td>91.6</td><td>51.1</td></tr><tr><td colspan=\"3\">Sum. 14392 91.2</td><td>73.1</td></tr></table>", |
| "html": null, |
| "text": "", |
| "type_str": "table", |
| "num": null |
| } |
| } |
| } |
| } |