H-Huang commited on
Commit
74287ed
·
1 Parent(s): 7e36217

Delete CWI_README.txt

Browse files
Files changed (1) hide show
  1. CWI_README.txt +0 -113
CWI_README.txt DELETED
@@ -1,113 +0,0 @@
1
- CWI datasets for English, German, and Spanish
2
- ==============================================
3
-
4
- Distributed by the LT group at Universität Hamburg, Germany, October 2017.
5
- contact email: yimam@informatik.uni-hamburg.de
6
- web: http://lt.informatik.uni-hamburg.de
7
-
8
-
9
- Introduction
10
- ------------
11
-
12
- Complex word Identification (CWI) is a sub-task of lexical simplification (LS), which identifies difficult words or phrases in a text. There are very few CWI datasets available, and mostly limited to English language. To alleviate this problem, we have collected CWI datasets for English, German, and Spanish.
13
-
14
- Data set collection procedures
15
- ------------------------------
16
-
17
- We collected complex word and phrase annotations (sequences of words, up to maximum 50 characters) using the Amazon Mechanical Turk (MTurk) crowdsourcing platform, from native and non-native English, German, and Spanish speakers. We collect annotations using MTurk, on a paragraph-level HIT (Human Intelligence Task), which is 5-10 sentences long.
18
-
19
- The English datasets consists of three genres:
20
-
21
- * Professionally written news
22
- * News written by amateurs (WikiNews)
23
- * Wikipedia articles
24
-
25
- The German and Spanish datasets are compiled from German Wikipedia and Spanish Wikipedia articles.
26
-
27
- Files
28
- -----
29
- The zip file contains this README file and three sub folders, namely English, German, and Spanish.
30
- Under each sub folder, there are both training and development annotation files.
31
- The test files are currently held back, as we are preparing a shared task on multilingual CWI.
32
-
33
- For English, there are the following files:
34
-
35
- ## development and training dataset the news genre, with annotations
36
- English_News_Train.tsv (6515 lines, 1500943 bytes)
37
- English_News_Dev.tsv (824 lines 186306 bytes)
38
-
39
- ## development and training dataset for news genre, all the sentences in a HIT with out the actual annotations
40
- English_News_HITs_Train.tsv (180lines, 153004 bytes)
41
- English_News_HITs_Dev.tsv (23lines, 19225 bytes)
42
-
43
- ## development and training dataset the Wikinews genre, with annotations
44
- English_WikiNews_Train.tsv (99 lines, 89136 bytes)
45
- English_WikiNews_Dev.tsv (476 lines, 102916 bytes)
46
-
47
- ## development and training dataset for Wikinews genre, all the sentences in a HIT with out the actual annotations
48
- English_WikiNews_HITs_Train.tsv (3878 lines, 864816 bytes)
49
- English_WikiNews_HITs_Dev.tsv (13 lines, 11525 bytes)
50
-
51
- ## development and training dataset the Wikipedia genre, with annotations
52
- English_Wikipedia_Train.tsv (2903 lines, 673747 bytes)
53
- English_Wikipedia_Dev.tsv (369 lines, 79832 bytes)
54
-
55
- ## development and training dataset for Wikipedia genre, all the sentences in a HIT with out the actual annotations
56
- English_Wikipedia_HITs_Train.tsv (81 lines, 63281 bytes)
57
- English_Wikipedia_HITs_Dev.tsv (10 lines, 8014 bytes)
58
-
59
- For German, there are the following files
60
- ## development and training dataset, with annotations
61
- German_Train.tsv (2629 lines, 618781 bytes)
62
- German_Dev.tsv (325 lines, 70478 bytes)
63
-
64
- ## development and training dataset, all the sentences in a HIT with out the actual annotations
65
- German_HITs_Train.tsv (85 lines, 91609 bytes)
66
- German_HITs_Dev.tsv (11 lines, 11937 bytes)
67
-
68
- For Spanish, there are the following files
69
- ## development and training dataset, with annotations
70
- Spanish_Train.tsv (5601 lines, 2954656 bytes)
71
- Spanish_Dev.tsv (701 lines, 312785 bytes)
72
-
73
- ## development and training dataset, all the sentences in a HIT with out the actual annotations
74
- Spanish_HITs_Train.tsv (141 lines, 166885 bytes)
75
- Spanish_HITs_Dev.tsv (18 lines, 19161 bytes)
76
-
77
-
78
- Data formats
79
- ------------
80
-
81
- These datasets contain information about complex phrases annotated with some statistics. Each line represents a sentence with one complex phrase (CP) annotation and relevant information, each separated by a TAB character.
82
-
83
- * First column shows the HIT ID of the sentence. All sentences with the same ID belong to the same HIT.
84
- * Second column shows the actual sentence where there exists a complex phrase annotation.
85
- * The third and fourth columns display the start and end offsets of the complex phrase annotation in this sentence.
86
- * The fifth column represents the actual complex phrase annotation.
87
- * The sixth, seventh, and eighth columns show the number of native annotators, the number of non-native annotators and the total number of annotators who have marked this complex phrase.
88
-
89
- Examples
90
- --------
91
-
92
- ID1 Both China and the Philippines flexed their muscles on Wednesday. 31 37 flexed 2 7 9
93
-
94
- ID1 Both China and the Philippines flexed their muscles on Wednesday. 31 51 flexed their muscles 4 2 6
95
-
96
- Here, we can see that the phrase "flexed" is marked as complex phrase by 2 native and 7 non-native English speakers where as the phrase "flexed their muscles" is marked by 4 native and 2 non native English speakers.
97
-
98
- Download
99
- --------
100
-
101
- The datasets are available from
102
- https://www.inf.uni-hamburg.de/en/inst/ab/lt/resources/data/complex-word-identification-dataset.html
103
-
104
- License
105
- -------
106
- The data is distributed under CC-BY 4.0 license, see https://creativecommons.org/licenses/by/4.0/ for details
107
-
108
- Publications
109
- ------------
110
- Please cite one of these publications, if you use the data in your research:
111
-
112
- * Seid Muhie Yimam, Sanja Štajner, Martin Riedl, and Chris Biemann (2017): CWIG3G2 - Complex Word Identification Task across Three Text Genres and Two User Groups. In Proceedings of The 8th International Joint Conference on Natural Language Processing (IJCNLP 2017). Taipei, Taiwan [For English data in different genres]
113
- * Seid Muhie Yimam, Sanja Štajner, Martin Riedl, and Chris Biemann (2017): Multilingual and Cross-Lingual Complex Word Identification. In Proceedings of The 2017 International Conference on Recent Advances in Natural Language Processing (RANLP). Varna, Bulgaria [for multilingual data]