| { |
| "paper_id": "W05-0203", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T04:45:42.287209Z" |
| }, |
| "title": "A real-time multiple-choice question generation for language testing -a preliminary study", |
| "authors": [ |
| { |
| "first": "Ayako", |
| "middle": [], |
| "last": "Hoshino", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Tokyo", |
| "location": { |
| "addrLine": "7-3-1 Hongo", |
| "postCode": "113-0033", |
| "settlement": "Bunkyo, Tokyo", |
| "country": "JAPAN" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Hiroshi", |
| "middle": [], |
| "last": "Nakagawa", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Tokyo", |
| "location": { |
| "addrLine": "7-3-1 Hongo, Bunkyo", |
| "postCode": "113-0033", |
| "settlement": "Tokyo", |
| "country": "JAPAN" |
| } |
| }, |
| "email": "nakagawa@dl.itc.u-tokyo.ac.jp" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "An automatic generation of multiplechoice questions is one of the promising examples of educational applications of NLP techniques. A machine learning approach seems to be useful for this purpose because some of the processes can be done by classification. Using basic machine learning algorithms as Naive Bayes and K-Nearest Neighbors, we have developed a real-time system which generates questions on English grammar and vocabulary from on-line news articles. This paper describes the current version of our system and discusses some of the issues on constructing this kind of system.", |
| "pdf_parse": { |
| "paper_id": "W05-0203", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "An automatic generation of multiplechoice questions is one of the promising examples of educational applications of NLP techniques. A machine learning approach seems to be useful for this purpose because some of the processes can be done by classification. Using basic machine learning algorithms as Naive Bayes and K-Nearest Neighbors, we have developed a real-time system which generates questions on English grammar and vocabulary from on-line news articles. This paper describes the current version of our system and discusses some of the issues on constructing this kind of system.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Multiple-choice question exams are widely used and are effective to assess students' knowledge, however it is costly to manually produce those questions. Naturally, this kind of task should be done with a help of computer.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Nevertheless, there have been very few attempts to generate multiple-choice questions automatically. Mitkov et al.(2003) generated questions for a linguistics exam in a semi-automatic way and evaluated that it exceeds manually made ones in cost and is at least equivalent in quality. There are some other researches that involve generating questions with multiple alternatives (Dicheva and Dimitrova, 1998) . But to the best of our knowledge, no attempt has been made to generate this kind of questions in a totally automatic way.", |
| "cite_spans": [ |
| { |
| "start": 101, |
| "end": 120, |
| "text": "Mitkov et al.(2003)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 377, |
| "end": 406, |
| "text": "(Dicheva and Dimitrova, 1998)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "This paper presents a novel approach to generate multiple-choice questions using machine learning techniques. The questions generated are those of fillin-the-blank type, so it does not involve transforming declarative sentences into question sentences as in Mitkov's work. This simplicity makes the method to be language independent.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Although this application can be very versatile, in that it can be used to test any kind of knowledge as in history exams, as a purpose of this research we limit ourselves to testing student's proficiency in a foreign language. One of the purposes of this research is to automatically extract important words or phrases in a text for a learner of the language.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The system we have implemented works in a simple pipelined manner; it takes an HTML file and turns it into the one of quiz session. The process of converting the input to multiple-choice questions includes extracting features, deciding the blank positions, and choosing the wrong alternatives (which are called distractors), which are all done in a moment when the user feeds the input. When the user submits their answer, it shows the text with the correct answers as well as an overall feed back.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "System Design", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The process of deciding blank positions in a given text follows a standard machine learning framework, which is first training a classifier on a training data , then applying it on an unseen test data, (i.e. the input text). In the current system, the mechanism of choosing distractors is implemented with the simplest algorithm, and its investigation is left to future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Methodology", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The training data is a collection of fill-in-the-blank questions from a TOEIC preparation book (Matsuno et al., 2000) . As shown in the box below, a question consists of a sentence with a missing word (or words) and four alternatives one of among which best fits into the blank.", |
| "cite_spans": [ |
| { |
| "start": 95, |
| "end": 117, |
| "text": "(Matsuno et al., 2000)", |
| "ref_id": "BIBREF2" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preparing the Training Data", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Many people showed up early to [ ] for the position that was open. 1. apply 2. appliance 3. applies 4. application", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preparing the Training Data", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The training instances are obtained from 100 questions by shifting the blank position. The original position is labeled as true, while sentences with a blank in a shifted position are at first labeled as false. The instance shown above therefore yields instances [ ] people showed up early to apply for the position that was open., Many [ ] showed up early to apply for the position that was open., and so on, all of which are labeled as false except the original blank position. 1962 (100 true and 1862 false) instances were obtained.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preparing the Training Data", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The label true here is supposed to indicate that it is possible to make a question with the sentence with a blank in the specified position, while many of the shifted positions which are labeled false can also be good blanks. A semi-supervised learning (Chakrabarti, 2003) 1 is conducted in the following manner to retrieve the instances that are potentially true among the ones initially classified as false.", |
| "cite_spans": [ |
| { |
| "start": 253, |
| "end": 272, |
| "text": "(Chakrabarti, 2003)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Preparing the Training Data", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "We retrieved the 13 instances (shown in Table 1 .) which had initially been labeled as false and classified as true in a test-on-train result with a certainty 2 of more than 0.5 with a Naive Bayes classifier 3 . The labels of those instances were changed to true before re-training the classifier. In this way, a training set with 113 true instances was obtained.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 40, |
| "end": 47, |
| "text": "Table 1", |
| "ref_id": "TABREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Preparing the Training Data", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "For the current system we use news articles from BBC.com 4 , which consist approximately 200-500 words. The test text goes through tagging and feature extraction in the same manner as the training 1 Semi-supervised learning is a method to identify the class of unclassified instances in the dataset where only some of the instances are classified.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Deciding Blank Positions", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "2 The result of a classification of a instance is obtained along with a certainty value between 0.0 to 1.0 for each class, which indicates how certain it is that an instance belongs to the class.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Deciding Blank Positions", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "3 Seven features which are word, POS, POS of the previous word, POS of the next word, position in the sentence, sentence length, word length and were used.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Deciding Blank Positions", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "4 http://news.bbc.co.uk/ data, and the instances are classified into true or false. The positions of the blanks are decided according to the certainty of the classification so the blanks (i.e. questions) are generated as many as the user has specified.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Deciding Blank Positions", |
| "sec_num": "3.2" |
| }, |
| { |
| "text": "In the current version of the system, the distractors are chosen randomly from the same article excluding punctuations and the same word as the other alternatives.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Choosing Distractors", |
| "sec_num": "3.3" |
| }, |
| { |
| "text": "The real-time system we are presenting is implemented as a Java servlet, whose one of the main screens is shown below. The tagger used here is the Tree tagger (Schmid, 1994) , which uses the Penn-Treebank tagset. The current version of the system is available at http://www.iii.u-tokyo.ac.jp/ qq36126/mcwa1/. The interface of the system consists of three sequenced web pages, namely 1)the parameter selection page, 2)the quiz session page and 3)the result page.", |
| "cite_spans": [ |
| { |
| "start": 159, |
| "end": 173, |
| "text": "(Schmid, 1994)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Current system", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The parameter selection page shows the list of the articles which are linked from the top page of the BBC website, along with the option selectors for number of blanks (5-30) and the classifier (Naive Bayes or Nearest Neighbors).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Current system", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The question session page is shown in Figure 1 . It displays the headline and the image from the chosen article under the title and a brief instruction. The alternatives are shown on option selectors, which are placed in the article text.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 38, |
| "end": 46, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Current system", |
| "sec_num": "4" |
| }, |
| { |
| "text": "The result page shows the text with the right answers shown in green when the user's choice is correct, red when it is wrong.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Current system", |
| "sec_num": "4" |
| }, |
| { |
| "text": "To examine the quality of the questions generated by the current system, we have evaluated the blank positions determined by a Naive Bayes classifier and a KNN classifier (K=3) with a certainty of more than 50 percent in 10 articles.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Among 3138 words in total, 361 blanks were made and they were manually evaluated according to their possibility of being a multiple-choice question, with an assumption of having alternatives of the same part of speech. The blank positions were categorized into three groups, which are E (possible to make a question), and D (difficult, but possible to make a question), NG (not possible or not suitable e.g. on a punctuation). The guideline for deciding E or D was if a question is on a grammar rule, or it requires more semantic understanding, for instance, a background knowledge 5 . Table 2 . shows the comparison of the number of blank positions decided by the two classifiers, each with a breakdown for each evaluation. The number in braces shows the proportion of the blanks with a certain evaluation over the total number of blanks made by the classifier. The rightmost column I shows the number of the same blank positions selected by both classifiers.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 586, |
| "end": 593, |
| "text": "Table 2", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The KNN classifier tends to be more accurate and seems to be more robust, although given the fact that it produces less blanks. The fact that an instancebased algorithm exceeds Naive Bayes, whose decision depends on the whole data, can be ascribed to a mixed nature of the training data. For example, blanks for grammar questions might have different features from ones for vocabulary questions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The result we sampled has exhibited another problem of Naive Bayes algorithm. In two articles among the data, it has shown the tendency to make a blank on be-verbs. Naive Bayes tends to choose the same word as a blank position, therefore generates many questions on the same word in one article.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "Another general problem of these methods would be that the blank positions are decided without consideration of one another; the question will be sometimes too difficult when another blank is next to or in the vicinity of the blank.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation", |
| "sec_num": "5" |
| }, |
| { |
| "text": "From the problems of the current system, we can conclude that the feature set we have used is not sufficient. It is necessary that we use larger number of features, possibly including semantic ones, so a blank position would not depend on its superficial aspects. Also, the training data should be examined in more detail.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion and Future work", |
| "sec_num": "6" |
| }, |
| { |
| "text": "As it was thought to be a criteria of evaluating generated questions, if a question requires simply a grammatical knowledge or a farther knowledge (i.e. background knowledge) can be a critical property of a generated question. We should differentiate the features from the ones which are used to generate, for example, history questions, which require rather background knowledge. Selecting suitable distractors, which is left to future work, would be a more important process in generating a question. A semantic distance between an alternative and the right answer are suggested (Mitkov and Ha, 2003) , to be a good measure to evaluate an alternative. We are investigating on a method of measuring those distances and a mechanism to retrieve best alternatives automatically.", |
| "cite_spans": [ |
| { |
| "start": 581, |
| "end": 602, |
| "text": "(Mitkov and Ha, 2003)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Discussion and Future work", |
| "sec_num": "6" |
| }, |
| { |
| "text": "We have presented a novel application of automatically generating fill-in-the-blank, multiple-choice questions using machine learning techniques, as well as a real-time system implemented. Although it is required to explore more feature settings for the process of determining blank positions, and the process of choosing distractors needs more elaboration, the system has proved to be feasible.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusion", |
| "sec_num": "7" |
| }, |
| { |
| "text": "A blank on a verbs or a part of idioms (as [according] to) was evaluated as E, most of the blanks on an adverbs, and (as [now]) were D and a blank on a punctuation or a quotation mark was NG.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Mining the Web", |
| "authors": [ |
| { |
| "first": "Soumen", |
| "middle": [], |
| "last": "Chakrabarti", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Soumen Chakrabarti. 2003. Mining the Web. Morgan Kaufmann Publishers.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "An approach to representation and extraction of terminological knowledge in icall", |
| "authors": [ |
| { |
| "first": "Darina", |
| "middle": [], |
| "last": "Dicheva", |
| "suffix": "" |
| }, |
| { |
| "first": "Vania", |
| "middle": [], |
| "last": "Dimitrova", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Journal of Computing and Information Technology", |
| "volume": "", |
| "issue": "", |
| "pages": "39--52", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Darina Dicheva and Vania Dimitrova. 1998. An ap- proach to representation and extraction of terminolog- ical knowledge in icall. In Journal of Computing and Information Technology, pages 39 -52.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "STEP-UP Bunpo mondai TOEIC TEST", |
| "authors": [ |
| { |
| "first": "Shuhou", |
| "middle": [], |
| "last": "Matsuno", |
| "suffix": "" |
| }, |
| { |
| "first": "Tomoko", |
| "middle": [], |
| "last": "Miyahara", |
| "suffix": "" |
| }, |
| { |
| "first": "Yoshi", |
| "middle": [], |
| "last": "Aoki", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Shuhou Matsuno, Tomoko Miyahara, and Yoshi Aoki. 2000. STEP-UP Bunpo mondai TOEIC TEST. Kiri- hara Publisher.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Computer-aided generation of multiple-choice tests", |
| "authors": [ |
| { |
| "first": "Ruslan", |
| "middle": [], |
| "last": "Mitkov", |
| "suffix": "" |
| }, |
| { |
| "first": "Le", |
| "middle": [ |
| "An" |
| ], |
| "last": "Ha", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of the HLT-NAACL 2003 Workshop on Building Educational Applications Using Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "17--22", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Ruslan Mitkov and Le An Ha. 2003. Computer-aided generation of multiple-choice tests. In Proceedings of the HLT-NAACL 2003 Workshop on Building Educa- tional Applications Using Natural Language Process- ing, pages 17 -22, Edmonton, Canada, May.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Probabilistic part-of-speech tagging using decision trees", |
| "authors": [ |
| { |
| "first": "Helmut", |
| "middle": [], |
| "last": "Schmid", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "Proceedings of International Conference on New Methods in Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Helmut Schmid. 1994. Probabilistic part-of-speech tag- ging using decision trees. In Proceedings of Interna- tional Conference on New Methods in Language Pro- cessing, Manchester, UK, September.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "num": null, |
| "uris": null, |
| "type_str": "figure", |
| "text": "a screen shot of the question session page with an enlarged answer selector." |
| }, |
| "TABREF0": { |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td/><td>: the full list of test instances classified as true in test-on-train</td><td/></tr><tr><td colspan=\"2\">certainty a test instance (sentence with a blank)</td><td>the answer</td></tr><tr><td>0.808</td><td>Joseph is preparing for tomorrow's big [ ] to the president.</td><td>presentation</td></tr><tr><td>0.751</td><td>Ms. Singh listened [ ] to the president's announcement.</td><td>carefully</td></tr><tr><td>0.744</td><td>The PR person is the one in charge of [ ] meetings and finding accommoda-</td><td>scheduling</td></tr><tr><td/><td>tions for our associates.</td><td/></tr><tr><td>0.73</td><td>Ms. Havlat received a memo from the CEO [ ] the employees' conduct.</td><td>regarding</td></tr><tr><td>0.718</td><td>The amount of money in the budget decreased [ ] over the past year.</td><td>significantly</td></tr><tr><td>0.692</td><td>Mr. Gomez is [ ] quickly; however it will be a log time before he gets used to</td><td>learning</td></tr><tr><td/><td>the job.</td><td/></tr><tr><td>0.689</td><td>The boss can never get around to [ ] off his desk.</td><td>cleaning</td></tr><tr><td>0.629</td><td>The interest rate has been increasingly [ ] higher.</td><td>getting</td></tr><tr><td>0.628</td><td>Employees are [ ] to comply with the rules in the handbook.</td><td>asked</td></tr><tr><td>0.62</td><td>The lawyer [ ] his case before the court.</td><td>presented</td></tr><tr><td>0.59</td><td>The secretary was [ ] to correspond with the client immediately.</td><td>supposed</td></tr><tr><td>0.576</td><td>The maintenance worker checked the machine before [ ] it on.</td><td>turning</td></tr><tr><td>0.523</td><td>The [ ] manager's office is across the corridor.</td><td>assistant</td></tr></table>", |
| "html": null, |
| "text": "" |
| }, |
| "TABREF1": { |
| "type_str": "table", |
| "num": null, |
| "content": "<table><tr><td/><td/><td>NB</td><td/><td/><td/><td colspan=\"2\">KNN</td><td>I</td><td/></tr><tr><td/><td>blanks</td><td>E(%)</td><td>D(%)</td><td colspan=\"2\">NG(%) blanks</td><td>E(%)</td><td colspan=\"3\">D(%) NG(%) blanks</td></tr><tr><td>Article1</td><td>69</td><td colspan=\"2\">44(63.8) 21(30.4)</td><td>4(5.8)</td><td colspan=\"3\">33 20(60.6) 11(33.3)</td><td>2(6.1)</td><td>18</td></tr><tr><td>Article2</td><td>22</td><td>5(22.7)</td><td colspan=\"2\">3(13.6) 14(63.6)</td><td>8</td><td>5(62.5)</td><td>3(37.5)</td><td>0(0.0)</td><td>0</td></tr><tr><td>Article3</td><td>38</td><td colspan=\"2\">21(55.3) 15(39.5)</td><td>2(5.3)</td><td colspan=\"2\">18 12(66.7)</td><td>5(27.8)</td><td>1(5.6)</td><td>8</td></tr><tr><td>Article4</td><td>19</td><td>10(52.6)</td><td>9(47.4)</td><td>0(0.0)</td><td>9</td><td>7(77.8)</td><td>2(22.2)</td><td>0(0.0)</td><td>3</td></tr><tr><td>Article5</td><td>28</td><td colspan=\"2\">18(64.3) 10(35.7)</td><td>0(0.0)</td><td colspan=\"2\">14 10(71.4)</td><td>4(28.6)</td><td>0(0.0)</td><td>6</td></tr><tr><td>Article6</td><td>26</td><td>17(65.4)</td><td>8(30.8)</td><td>1(3.8)</td><td>11</td><td>6(54.5)</td><td>5(45.5)</td><td>0(0.0)</td><td>4</td></tr><tr><td>Article7</td><td>18</td><td>9(50.0)</td><td>5(27.8)</td><td>4(22.2)</td><td>6</td><td>3(50.0)</td><td>3(50.0)</td><td>0(0.0)</td><td>3</td></tr><tr><td>Article8</td><td>24</td><td>14(58.3)</td><td>9(37.5)</td><td>1(4.2)</td><td>5</td><td>3(60.0)</td><td>2(40.0)</td><td>0(0.0)</td><td>5</td></tr><tr><td>Article9</td><td>20</td><td>16(80.0)</td><td>4(20.0)</td><td>0(0.0)</td><td>6</td><td>2(33.3)</td><td>4(66.7)</td><td>0(0.0)</td><td>4</td></tr><tr><td>Article10</td><td>30</td><td colspan=\"2\">18(60.0) 12(40.0)</td><td>0(0.0)</td><td colspan=\"2\">14 11(78.6)</td><td>3(21.4)</td><td>0(0.0)</td><td>6</td></tr><tr><td/><td colspan=\"3\">294 172(58.5) 96(32.7)</td><td>26(8.8)</td><td colspan=\"3\">124 79(63.7) 42(33.9)</td><td>3(2.4)</td><td>57</td></tr></table>", |
| "html": null, |
| "text": "The evaluation on the blank positions decided by a Naive Bayes (NB) and a KNN classifier." |
| } |
| } |
| } |
| } |