| | # SumeCzech Corpus |
| |
|
| | These are the accompanying materials of the paper: |
| | ``` |
| | @inproceedings{straka-etal-2018-sumeczech, |
| | title = "{S}ume{C}zech: Large {C}zech News-Based Summarization Dataset", |
| | author = "Straka, Milan and Mediankin, Nikita and Kocmi, Tom and |
| | {\v{Z}}abokrtsk{\'y}, Zden{\v{e}}k and Hude{\v{c}}ek, Vojt{\v{e}}ch and Haji{\v{c}}, Jan", |
| | booktitle = "Proceedings of the Eleventh International Conference on Language Resources and Evaluation ({LREC}-2018)", |
| | month = may, |
| | year = "2018", |
| | address = "Miyazaki, Japan", |
| | publisher = "European Languages Resources Association (ELRA)", |
| | } |
| | ``` |
| |
|
| | ## SumeCzech Download Script |
| |
|
| | To download the SumeCzech dataset, use the `downloader.py` script. |
| | The script has several dependencies (and requires an exact version for |
| | some of them) listed in `requirements.txt`, you can install them |
| | using `pip3 install -r requirements.txt`. |
| |
|
| | You can start the script using `python3 downloader.py`. By default, |
| | 16 parallel processes are used to download the data (you can |
| | override this number using the `--parallel N` option). |
| |
|
| | During download, MD5 hash of every document's headline, abstract and text |
| | is checked with the official one, allowing to detect possible errors |
| | during download and extraction. Although not recommended, the check |
| | can be bypassed by using the `--no_verify_md5` option. |
| |
|
| | The validated documents are saved during download. If the download script |
| | is interrupted and run again, it will reuse the already processed |
| | documents and only download new ones. |
| |
|
| | ### Changelog: |
| |
|
| | - 13 Feb 2018: The original download script was released. |
| |
|
| | - 25 Feb 2023: An update with the following changes: |
| | - use the new domain https://data.commoncrawl.org of the CC download; |
| | - support Python 3.10 and 3.11, where `collections.Callable` was removed. |
| |
|
| |
|
| | ## SumeCzech ROUGE_RAW Evaluation Metric |
| | |
| | The RougeRAW metric is implemented in `rouge_raw.py` module, which can |
| | compute the RougeRAW-1, RougeRAW-2, RougeRAW-L metrics either for |
| | a single pair of documents, or for a pair of corpora. |
| |
|
| | Unfortunately, slightly different tokenization was used in the original |
| | paper. Therefore, here we provide the results of the systems from the paper |
| | evaluated using the `rouge_raw.py` module. |
| |
|
| | ### Results for abstract-headline on test |
| | ``` |
| | RougeRAW-1 RougeRAW-2 RougeRAW-L |
| | Method P R F P R F P R F |
| | first 13.9 23.6 16.5 04.1 07.4 05.0 12.2 20.7 14.5 |
| | random 11.0 17.8 12.8 02.6 04.5 03.1 09.6 15.5 11.1 |
| | textrank 13.3 22.8 15.9 03.7 06.8 04.6 11.6 19.9 13.8 |
| | t2t 20.2 15.9 17.2 06.7 05.1 05.6 18.6 14.7 15.8 |
| | ``` |
| |
|
| | ### Results for abstract-headline on oodtest |
| | ``` |
| | RougeRAW-1 RougeRAW-2 RougeRAW-L |
| | Method P R F P R F P R F |
| | first 13.3 26.5 16.7 04.7 10.0 06.0 11.6 23.3 14.7 |
| | random 10.6 20.7 13.1 03.2 06.9 04.1 09.3 18.2 11.5 |
| | textrank 12.8 25.9 16.3 04.5 09.6 05.7 11.3 22.7 14.2 |
| | t2t 19.4 15.1 16.3 07.1 05.2 05.7 18.1 14.1 15.2 |
| | ``` |
| |
|
| | ### Results for text-headline on test |
| | ``` |
| | RougeRAW-1 RougeRAW-2 RougeRAW-L |
| | Method P R F P R F P R F |
| | first 07.4 13.5 08.9 01.1 02.2 01.3 06.5 11.7 07.7 |
| | random 05.9 10.3 06.9 00.5 01.0 00.6 05.2 08.9 06.0 |
| | textrank 06.0 16.5 08.3 00.8 02.3 01.1 05.0 13.8 06.9 |
| | t2t 08.8 07.0 07.5 00.8 00.6 00.7 08.1 06.5 07.0 |
| | ``` |
| |
|
| | ### Results for text-headline on oodtest |
| | ``` |
| | RougeRAW-1 RougeRAW-2 RougeRAW-L |
| | Method P R F P R F P R F |
| | first 06.7 13.6 08.3 01.3 02.8 01.6 05.9 12.0 07.4 |
| | random 05.2 10.0 06.3 00.6 01.4 00.8 04.6 08.9 05.6 |
| | textrank 05.8 16.9 08.1 01.1 03.4 01.5 05.0 14.5 06.9 |
| | t2t 06.3 05.1 05.5 00.5 00.4 00.4 05.9 04.8 05.1 |
| | ``` |
| |
|
| | ### Results for text-abstract on test |
| | ``` |
| | RougeRAW-1 RougeRAW-2 RougeRAW-L |
| | Method P R F P R F P R F |
| | first 13.1 17.9 14.4 01.9 02.8 02.1 08.8 12.0 09.6 |
| | random 11.7 15.5 12.7 01.2 01.7 01.3 07.7 10.3 08.4 |
| | textrank 11.1 20.8 13.8 01.6 03.1 02.0 07.1 13.4 08.9 |
| | t2t 13.2 10.5 11.3 01.2 00.9 01.0 10.2 08.1 08.7 |
| | ``` |
| |
|
| | ### Results for text-abstract on oodtest |
| | ``` |
| | RougeRAW-1 RougeRAW-2 RougeRAW-L |
| | Method P R F P R F P R F |
| | first 11.1 17.1 12.7 01.6 02.7 01.9 07.6 11.7 08.7 |
| | random 10.1 15.1 11.4 01.0 01.7 01.2 06.9 10.3 07.8 |
| | textrank 09.8 19.9 12.5 01.5 03.3 02.0 06.6 13.3 08.4 |
| | t2t 12.5 09.4 10.3 00.8 00.6 00.6 09.8 07.5 08.1 |
| | ``` |
| |
|