| { |
| "paper_id": "S10-1046", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T15:28:04.557920Z" |
| }, |
| "title": "TUD: semantic relatedness for relation classification", |
| "authors": [ |
| { |
| "first": "Gy\u00f6rgy", |
| "middle": [], |
| "last": "Szarvas", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Technische Universit\u00e4t Darmstadt Hochschulstra\u00dfe", |
| "location": { |
| "addrLine": "10", |
| "postCode": "D-64289", |
| "settlement": "Darmstadt", |
| "country": "Germany" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Iryna", |
| "middle": [], |
| "last": "Gurevych", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Technische Universit\u00e4t Darmstadt Hochschulstra\u00dfe", |
| "location": { |
| "addrLine": "10", |
| "postCode": "D-64289", |
| "settlement": "Darmstadt", |
| "country": "Germany" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "In this paper, we describe the system submitted by the team TUD to Task 8 at SemEval 2010. The challenge focused on the identification of semantic relations between pairs of nominals in sentences collected from the web. We applied maximum entropy classification using both lexical and syntactic features to describe the nominals and their context. In addition, we experimented with features describing the semantic relatedness (SR) between the target nominals and a set of clue words characteristic to the relations. Our best submission with SR features achieved 69.23% macro-averaged F-measure, providing 8.73% improvement over our baseline system. Thus, we think SR can serve as a natural way to incorporate external knowledge to relation classification.", |
| "pdf_parse": { |
| "paper_id": "S10-1046", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "In this paper, we describe the system submitted by the team TUD to Task 8 at SemEval 2010. The challenge focused on the identification of semantic relations between pairs of nominals in sentences collected from the web. We applied maximum entropy classification using both lexical and syntactic features to describe the nominals and their context. In addition, we experimented with features describing the semantic relatedness (SR) between the target nominals and a set of clue words characteristic to the relations. Our best submission with SR features achieved 69.23% macro-averaged F-measure, providing 8.73% improvement over our baseline system. Thus, we think SR can serve as a natural way to incorporate external knowledge to relation classification.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Automatic extraction of typed semantic relations between sentence constituents is an important step towards deep semantic analysis and understanding the semantic content of natural language texts. Identification of relations between a nominal and the main verb, and between pairs of nominals are important steps for the extraction of structured semantic information from text, and can benefit various applications ranging from Information Extraction and Information Retrieval to Machine Translation or Question Answering.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The Multi-Way Classification of Semantic Relations Between Pairs of Nominals challenge (Hendrickx et al., 2010) focused on the identification of specific relation types between nominals (nouns or base noun phrases) in natural language sentences collected from the web. The main task of the challenge was to identify and classify instances of 9 abstract semantic relations between noun phrases, i.e. Cause-Effect, Instrument-Agency, Product-Producer, Content-Container, Entity-Origin, Entity-Destination, Component-Whole, Member-Collection, Message-Topic. That is, given two nominals (e1 and e2) in a sentence, systems had to decide whether relation(e1,e2), relation(e2,e1) holds for one of the relation types or the nominals' relation is other (falls to a category not listed above or they are unrelated). In this sense, the challenge was an important pilot task towards large scale semantic processing of text.", |
| "cite_spans": [ |
| { |
| "start": 87, |
| "end": 111, |
| "text": "(Hendrickx et al., 2010)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "In this paper, we describe the system we submitted to Semeval 2010, Task 8. We applied maximum entropy classification to the problem using both lexical and contextual features to describe the nominals themselves and their context (i.e. the sentence). In addition, we experimented with features exploiting the strength of association between the target nominals and a predefined set of clue words characteristic to the nine relation types. In order to measure the semantic relatedness (SR) of targets and clues, we used the Explicit Semantic Analysis (Gabrilovich and Markovitch, 2007) SR measure (based on Wikipedia, Wiktionary and WordNet). Our best submission, benefiting from SR features, achieved 69.23% macro-averaged Fmeasure for the 9 relation types used. Providing 8.73% improvement over our baseline system, we found the SR-based features to be beneficial for the classification of semantic relations.", |
| "cite_spans": [ |
| { |
| "start": 550, |
| "end": 584, |
| "text": "(Gabrilovich and Markovitch, 2007)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Feature set In our system, we used both lexical (1-3) and contextual features (4-8) to describe the nominals and their context (i.e. the sentence). Additionally, we experimented with a set of features (9) that exploit the co-occurrence statistics of the nominals and a set of clue words chosen manually, examining the relation definitions and examples provided by the organizers. The clues characterize the relations addressed in the task (e.g. cargo, goods, content, box, bottle characterize the Content-Container relation) 1 . Each feature type was distinguished from the others using a prefix. All but the semantic relatedness features we used were binary, denoting whether a specific word, lemma, POS tag, etc. is found in the example sentence, or not. SR features were real valued, scaled to [0, 1] for each clue word separately (on train, and the same scaling factores were applied on the test data). The feature types used: 1. Token: word unigrams in the sentence in their inflected form. 2. Lemma: word uni-and bigrams in the sentence in their lemmatized form. 3. Target Nouns: the syntactic head words of the target nouns. 4. POS: the part of speech uni-and biand trigrams in the sentence. 5. Between POS: the part of speech sequence between the target nouns. 6. Dependency Path: the dependency path (syntactic relations and directions) between the target nouns. The whole path constituted a single feature. 7. Target Distance: the distance between the target nouns (in tokens). 8. Sentence Length: the length of the sentence (in tokens). 9. Semantic Relatedness: the semantic relatedness scores measuring the strength of association between the target nominals and the set of clue words we collected. In order to measure the semantic relatedness (SR) of targets and clues, we used the Explicit Semantic Analysis (ESA) SR measure. Feature selection In order to discard uninformative features automatically, we performed feature selection on the binary features. We kept features that satisfied the following three conditions:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental setup 2.1 Feature set and selection", |
| "sec_num": "2" |
| }, |
| { |
| "text": "f req(x) > 3 (1) p = argmax y P (y|x) > t 1 (2) p 5 \u00d7 f req(x) > t 2 (3)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental setup 2.1 Feature set and selection", |
| "sec_num": "2" |
| }, |
| { |
| "text": "where f req(x) denotes the frequency of feature x observed in the training dataset, y denotes a class label, p denotes the highest posterior probability (for feature x) over the nine relations (undirected) and the other class. Finally, t 1 , t 2 are filtering thresholds chosen arbitrarily. We used t 1 = 0.25 for all features but the dependency path, where we used t 1 = 0.2. We set the thresold t 2 to 1.9 for lexical features (i.e. token and lemma features), to 0.3 for dependency path features and to 0.9 for all other features. All parameters for the feature selection process were chosen manually (crossvalidating the parameters was omitted due to lack of time during the challenge development period). The higher t 2 value for lexical features was motivated by the aim to avoid overfitting, and the lower thresholds for dependency-based features by the hypothesis that these can be most efficient to determine the direction of relationships (c.f. we disregarded direction during feature selection). As the numeric SR features were all bound to clue words selected specifically for the task, we did not perform any feature selection for that feature type.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Experimental setup 2.1 Feature set and selection", |
| "sec_num": "2" |
| }, |
| { |
| "text": "We compared three learning algorithms, using the baseline feature types (1-8), namely a C4.5 decision tree learner, a support vector classifier (SMO), and a maximum entropy (logistic regression) classifier, all implemented in the Weka package (Hall et al., 2009) . We trained the SMO model with polynomial kernel of degree 2, fitting logistic models to the output to get valid probability estimates and the C4.5 model with pruning confidence factor set to 0.33. All other parameters were set to their default values as defined in Weka. We found the maxent model to perform best in 10-fold cross validation on the training set (see Table 1 ). Thus, we used maxent in our submissions.", |
| "cite_spans": [ |
| { |
| "start": 243, |
| "end": 262, |
| "text": "(Hall et al., 2009)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 631, |
| "end": 638, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Learning models", |
| "sec_num": "2.2" |
| }, |
| { |
| "text": "We submitted 4 runs to the challenge. Baseline (Base) As our baseline system, we used the information extracted from the sentence itself (i.e. lexical and contextual features, types 1-8). Wikipedia (WP) As a first extension, we added SR features (9) exploiting term co-occurrence information, using the ESA model with Wikipedia.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Combined Semantic Relatedness (cSR) Second, we replaced the ESA measure with a combined measure developed by us, exploiting term cooccurrence not only in Wikipedia, but also in WordNet and Wiktionary glosses. We found this measure to perform better than the Wikipediabased ESA in earlier experiments. cSR threshold (cSR-t) We submitted the predictions of the cSR system, with less emphasis on the other class: we predicted other label only when the following held for the posteriors predicted by cSR:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "3" |
| }, |
| { |
| "text": "argmaxyP (y|x) p(other)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "3" |
| }, |
| { |
| "text": "< 0.7. The threshold 0.7 was chosen based on the training dataset. First, the SR features improved the performance of our system by a wide marging (see Table 2 ). The difference in performance is even more prominent on the Test dataset, which suggests that these features efficiently incorporated useful external evidence on the relation between the nominals and this not just improved the accuracy of the system, but also helped to avoid overfitting. Thus we conclude that the SR features with the encoded external knowledge helped the maxent model to learn a hypothesis that clearly generalized better.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 152, |
| "end": 160, |
| "text": "Table 2", |
| "ref_id": "TABREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Second, we notice that the combined SR measure proved to be more useful than the standard ESA measure (Gabrilovich and Markovitch, 2007) improving the performance by approximately 1 percent over ESA, both in terms of macro averaged F-measure and overall accuracy. This confirms our hypothesis that the combined measure is more robust than ESA with just Wikipedia. Table 3 : Prediction error statistics. Table 3 shows the breakdown of system predictions to different categories, and their contribution to the official ranking as true/false positives and negatives. The submission that manipulated the decision threshold for the other class improved the overall performance by a small margin. This fact, and Table 3 confirm that our approach had major difficulties in correctly discriminating the 9 relation categories from other. Since this class is an umbrella class for unrelated nominals and the numerous semantic relations not considered in the challenge, it proved to be extremely difficult to accurately characterize this class. On the other hand, the confusion of the 9 specified relations (between each other) and directionality were less prominent error types. The most frequent cross-relation confusion types were the misclassification of Component-Whole as Instrument-Agency and Member-Collection; Content-Container as Component-Whole; Instrument-Agency as Product-Producer and vice versa. Interestingly, Component-Whole and Cause-Effect relations were the most typical sources for wrong direction errors. Lowering the decision threshold for other in our system naturally resulted in more true positive relation classifications, but unfortunately not only raised the number of other instances falsely classified as being one of the valuable re-lations, but also introduced several wrong relation classification errors (see Table 3 ). That is why this step resulted only in marginal improvement.", |
| "cite_spans": [ |
| { |
| "start": 102, |
| "end": 136, |
| "text": "(Gabrilovich and Markovitch, 2007)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 364, |
| "end": 371, |
| "text": "Table 3", |
| "ref_id": null |
| }, |
| { |
| "start": 403, |
| "end": 410, |
| "text": "Table 3", |
| "ref_id": null |
| }, |
| { |
| "start": 706, |
| "end": 713, |
| "text": "Table 3", |
| "ref_id": null |
| }, |
| { |
| "start": 1833, |
| "end": 1840, |
| "text": "Table 3", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "3" |
| }, |
| { |
| "text": "In this paper, we presented our system submitted to the Multi-Way Classification of Semantic Relations Between Pairs of Nominals challenge at SemEval 2010. We submitted 4 different system runs. Our first submission was a baseline system (Base) exploiting lexical and contextual information collected solely from the sentence to be classified. A second run (WP) complemented this baseline configuration with a set of features that used Explicit Semantic Analysis (Wikipedia) to model the SR of the nominals to be classified and a set of clue words characteristic of the relations used in the challenge. Our third run (cSR) used a combined semantic relatedness measure that exploits multiple lexical semantic resources (Wikipedia, Wiktionary and WordNet) to provide more reliable relatedness estimates. Our final run (cSR-t) exploited that our system in general was inaccurate in predicting instances of the other class. Thus, it used the same predictions as cSR, but favored the prediction of one of the 9 specified classes instead of other, when a comparably high posterior for such a class was predicted by the system.", |
| "cite_spans": [ |
| { |
| "start": 717, |
| "end": 752, |
| "text": "(Wikipedia, Wiktionary and WordNet)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions & Future Work", |
| "sec_num": "4" |
| }, |
| { |
| "text": "Our approach is fairly simple, in the sense that it used mostly just local information collected from the sentence. It is clear though that encoding as much general world knowledge to the representation as possible is crucial for efficient classification of semantic relations. In the light of the above fact, the results we obtained are reasonable.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions & Future Work", |
| "sec_num": "4" |
| }, |
| { |
| "text": "As the main goal of our study, we attempted to use semantic relatedness features that exploit texts in an external knowledge source (Wikipedia, Wiktionary or WordNet in our case) to incorporate some world knowledge in the form of term cooccurrence scores. We found that our SR features significantly contribute to system performance. Thus, we think this kind of information is useful in general for relation classification. The experimental results showed that our combined SR measure performed better than the standard ESA using Wikipedia. This confirms our hypothesis that exploiting multiple resources for modeling term relatedness is beneficial in general. Obviously, our system leaves much space for improvement -the feature selection parameters and the clue word set for the SR features were chosen manually, without any cross-validation (on the training set), due to lack of time. One of the participating teams used an SVM-based system and gained a lot from manipulating the decision thresholds. Thus, despite our preliminary results, it is also an interesting option to use SVMs.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions & Future Work", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In general, we think that more features are needed to achieve significantly better performance than we reported here. Top performing systems in the challenge typically exploited web frequency information (n-gram data) and manually encoded relations from an ontology (mainly WordNet). Thus, future work is to incorporate such features.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions & Future Work", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We demonstrated that SR features are helpful to move away from lexicalized systems using tokenor lemma-based features. Probably the same holds for web-based and ontology-based features extensively used by top performing systems. This suggests that experimenting with all these to see if their value is complementary is an especially interesting piece of future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions & Future Work", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The clue list is available at:http://www.ukp.tu-darmstadt.de/research/data/ relation-classification/", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Computing Semantic Relatedness using Wikipediabased Explicit Semantic Analysis", |
| "authors": [ |
| { |
| "first": "Evgeniy", |
| "middle": [], |
| "last": "Gabrilovich", |
| "suffix": "" |
| }, |
| { |
| "first": "Shaul", |
| "middle": [], |
| "last": "Markovitch", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Proceedings of The 20th International Joint Conference on Artificial Intelligence", |
| "volume": "", |
| "issue": "", |
| "pages": "1606--1611", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Evgeniy Gabrilovich and Shaul Markovitch. 2007. Computing Semantic Relatedness using Wikipedia- based Explicit Semantic Analysis. In Proceedings of The 20th International Joint Conference on Arti- ficial Intelligence, pages 1606-1611.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "The WEKA Data Mining Software: An Update. SIGKDD Explorations", |
| "authors": [ |
| { |
| "first": "Mark", |
| "middle": [], |
| "last": "Hall", |
| "suffix": "" |
| }, |
| { |
| "first": "Eibe", |
| "middle": [], |
| "last": "Frank", |
| "suffix": "" |
| }, |
| { |
| "first": "Geoffrey", |
| "middle": [], |
| "last": "Holmes", |
| "suffix": "" |
| }, |
| { |
| "first": "Bernhard", |
| "middle": [], |
| "last": "Pfahringer", |
| "suffix": "" |
| }, |
| { |
| "first": "Peter", |
| "middle": [], |
| "last": "Reutemann", |
| "suffix": "" |
| }, |
| { |
| "first": "Ian", |
| "middle": [ |
| "H" |
| ], |
| "last": "Witten", |
| "suffix": "" |
| } |
| ], |
| "year": 2009, |
| "venue": "", |
| "volume": "11", |
| "issue": "", |
| "pages": "10--18", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mark Hall, Eibe Frank, Geoffrey Holmes, Bernhard Pfahringer, Peter Reutemann, and Ian H. Witten. 2009. The WEKA Data Mining Software: An Up- date. SIGKDD Explorations, 11(1):10-18.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Semeval-2010 task 8: Multi-way classification of semantic relations between pairs of nominals", |
| "authors": [ |
| { |
| "first": "Iris", |
| "middle": [], |
| "last": "Hendrickx", |
| "suffix": "" |
| }, |
| { |
| "first": "Su", |
| "middle": [ |
| "Nam" |
| ], |
| "last": "Kim", |
| "suffix": "" |
| }, |
| { |
| "first": "Zornitsa", |
| "middle": [], |
| "last": "Kozareva", |
| "suffix": "" |
| }, |
| { |
| "first": "Preslav", |
| "middle": [], |
| "last": "Nakov", |
| "suffix": "" |
| }, |
| { |
| "first": "Diarmuid\u00f3", |
| "middle": [], |
| "last": "S\u00e9aghdha", |
| "suffix": "" |
| }, |
| { |
| "first": "Sebastian", |
| "middle": [], |
| "last": "Pad\u00f3", |
| "suffix": "" |
| }, |
| { |
| "first": "Marco", |
| "middle": [], |
| "last": "Pennacchiotti", |
| "suffix": "" |
| }, |
| { |
| "first": "Lorenza", |
| "middle": [], |
| "last": "Romano", |
| "suffix": "" |
| }, |
| { |
| "first": "Stan", |
| "middle": [], |
| "last": "Szpakowicz", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "Proceedings of the 5th SIGLEX Workshop on Semantic Evaluation", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Iris Hendrickx, Su Nam Kim, Zornitsa Kozareva, Preslav Nakov, Diarmuid\u00d3 S\u00e9aghdha, Sebastian Pad\u00f3, Marco Pennacchiotti, Lorenza Romano, and Stan Szpakowicz. 2010. Semeval-2010 task 8: Multi-way classification of semantic relations be- tween pairs of nominals. In Proceedings of the 5th SIGLEX Workshop on Semantic Evaluation.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "TABREF1": { |
| "num": null, |
| "type_str": "table", |
| "html": null, |
| "text": "Performance of different learning methods on train (10-fold).", |
| "content": "<table/>" |
| }, |
| "TABREF2": { |
| "num": null, |
| "type_str": "table", |
| "html": null, |
| "text": "17% 78.25% 79.42% 79.10% 80.69% 81.90% 83.76% 83.38% component-whole 54.68% 58.71% 60.18% 60.79% 50.52% 57.90% 61.67% 62.15% content-container 75.09% 77.55% 78.26% 78.11% 75.27% 78.96% 78.33% 78.87% entity-destination 81.99% 82.97% 83.12% 82.90% 77.59% 82.86% 81.54% 81.12% entity-origin 68.74% 70.39% 71.14% 71.18% 67.08% 72.05% 71.03%", |
| "content": "<table><tr><td/><td/><td>Train</td><td/><td/><td/><td colspan=\"2\">Test</td><td/></tr><tr><td>relation type</td><td>Base</td><td>WP</td><td>cSR</td><td>cSR-t</td><td>Base</td><td>WP</td><td>cSR</td><td>cSR-t</td></tr><tr><td>cause-effect</td><td colspan=\"8\">78.70.36%</td></tr><tr><td>instrument-agency</td><td colspan=\"4\">47.59% 56.71% 59.60% 59.80%</td><td colspan=\"2\">31.09% 44.06%</td><td colspan=\"2\">46.78% 46.91%</td></tr><tr><td>member-collection</td><td colspan=\"4\">75.27% 79.43% 80.71% 80.89%</td><td colspan=\"4\">66.37% 71.24% 72.65% 72.65%</td></tr><tr><td>message-topic</td><td colspan=\"4\">56.40% 62.68% 64.77% 65.15%</td><td colspan=\"2\">49.88% 65.06%</td><td colspan=\"2\">68.15% 69.83%</td></tr><tr><td>product-producer</td><td colspan=\"4\">53.36% 57.98% 59.97% 60.70%</td><td colspan=\"4\">46.04% 57.94% 56.00% 57.85%</td></tr><tr><td>macro AVG F1</td><td colspan=\"4\">65.70% 69.40% 70.80% 70.96%</td><td colspan=\"2\">60.50% 68.00%</td><td colspan=\"2\">68.88% 69.23%</td></tr><tr><td colspan=\"5\">accuracy (incl. other) 62.10% 65.42% 66.83% 67.12%</td><td colspan=\"2\">56.13% 63.49%</td><td colspan=\"2\">64.63% 65.37%</td></tr></table>" |
| }, |
| "TABREF3": { |
| "num": null, |
| "type_str": "table", |
| "html": null, |
| "text": "Performance of 4 submissions on train (10-fold) and test.", |
| "content": "<table/>" |
| } |
| } |
| } |
| } |