File size: 4,575 Bytes
40b0307
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
# SumeCzech Corpus

These are the accompanying materials of the paper:
```
@inproceedings{straka-etal-2018-sumeczech,
    title = "{S}ume{C}zech: Large {C}zech News-Based Summarization Dataset",
    author = "Straka, Milan  and Mediankin, Nikita  and Kocmi, Tom  and
      {\v{Z}}abokrtsk{\'y}, Zden{\v{e}}k  and Hude{\v{c}}ek, Vojt{\v{e}}ch  and Haji{\v{c}}, Jan",
    booktitle = "Proceedings of the Eleventh International Conference on Language Resources and Evaluation ({LREC}-2018)",
    month = may,
    year = "2018",
    address = "Miyazaki, Japan",
    publisher = "European Languages Resources Association (ELRA)",
}
```

## SumeCzech Download Script

To download the SumeCzech dataset, use the `downloader.py` script.
The script has several dependencies (and requires an exact version for
some of them) listed in `requirements.txt`, you can install them
using `pip3 install -r requirements.txt`.

You can start the script using `python3 downloader.py`. By default,
16 parallel processes are used to download the data (you can
override this number using the `--parallel N` option).

During download, MD5 hash of every document's headline, abstract and text
is checked with the official one, allowing to detect possible errors
during download and extraction. Although not recommended, the check
can be bypassed by using the `--no_verify_md5` option.

The validated documents are saved during download. If the download script
is interrupted and run again, it will reuse the already processed
documents and only download new ones.

### Changelog:

- 13 Feb 2018: The original download script was released.

- 25 Feb 2023: An update with the following changes:
  - use the new domain https://data.commoncrawl.org of the CC download;
  - support Python 3.10 and 3.11, where `collections.Callable` was removed.


## SumeCzech ROUGE_RAW Evaluation Metric

The RougeRAW metric is implemented in `rouge_raw.py` module, which can
compute the RougeRAW-1, RougeRAW-2, RougeRAW-L metrics either for
a single pair of documents, or for a pair of corpora.

Unfortunately, slightly different tokenization was used in the original
paper. Therefore, here we provide the results of the systems from the paper
evaluated using the `rouge_raw.py` module.

### Results for abstract-headline on test
```
            RougeRAW-1      RougeRAW-2      RougeRAW-L
Method      P    R    F     P    R    F     P    R    F
first     13.9 23.6 16.5  04.1 07.4 05.0  12.2 20.7 14.5
random    11.0 17.8 12.8  02.6 04.5 03.1  09.6 15.5 11.1
textrank  13.3 22.8 15.9  03.7 06.8 04.6  11.6 19.9 13.8
t2t       20.2 15.9 17.2  06.7 05.1 05.6  18.6 14.7 15.8
```

### Results for abstract-headline on oodtest
```
            RougeRAW-1      RougeRAW-2      RougeRAW-L
Method      P    R    F     P    R    F     P    R    F
first     13.3 26.5 16.7  04.7 10.0 06.0  11.6 23.3 14.7
random    10.6 20.7 13.1  03.2 06.9 04.1  09.3 18.2 11.5
textrank  12.8 25.9 16.3  04.5 09.6 05.7  11.3 22.7 14.2
t2t       19.4 15.1 16.3  07.1 05.2 05.7  18.1 14.1 15.2
```

### Results for text-headline on test
```
            RougeRAW-1      RougeRAW-2      RougeRAW-L
Method      P    R    F     P    R    F     P    R    F
first     07.4 13.5 08.9  01.1 02.2 01.3  06.5 11.7 07.7
random    05.9 10.3 06.9  00.5 01.0 00.6  05.2 08.9 06.0
textrank  06.0 16.5 08.3  00.8 02.3 01.1  05.0 13.8 06.9
t2t       08.8 07.0 07.5  00.8 00.6 00.7  08.1 06.5 07.0
```

### Results for text-headline on oodtest
```
            RougeRAW-1      RougeRAW-2      RougeRAW-L
Method      P    R    F     P    R    F     P    R    F
first     06.7 13.6 08.3  01.3 02.8 01.6  05.9 12.0 07.4
random    05.2 10.0 06.3  00.6 01.4 00.8  04.6 08.9 05.6
textrank  05.8 16.9 08.1  01.1 03.4 01.5  05.0 14.5 06.9
t2t       06.3 05.1 05.5  00.5 00.4 00.4  05.9 04.8 05.1
```

### Results for text-abstract on test
```
            RougeRAW-1      RougeRAW-2      RougeRAW-L
Method      P    R    F     P    R    F     P    R    F
first     13.1 17.9 14.4  01.9 02.8 02.1  08.8 12.0 09.6
random    11.7 15.5 12.7  01.2 01.7 01.3  07.7 10.3 08.4
textrank  11.1 20.8 13.8  01.6 03.1 02.0  07.1 13.4 08.9
t2t       13.2 10.5 11.3  01.2 00.9 01.0  10.2 08.1 08.7
```

### Results for text-abstract on oodtest
```
            RougeRAW-1      RougeRAW-2      RougeRAW-L
Method      P    R    F     P    R    F     P    R    F
first     11.1 17.1 12.7  01.6 02.7 01.9  07.6 11.7 08.7
random    10.1 15.1 11.4  01.0 01.7 01.2  06.9 10.3 07.8
textrank  09.8 19.9 12.5  01.5 03.3 02.0  06.6 13.3 08.4
t2t       12.5 09.4 10.3  00.8 00.6 00.6  09.8 07.5 08.1
```