| --- |
| pretty_name: pywsd-datasets |
| license: mit |
| task_categories: |
| - token-classification |
| language: |
| - en |
| tags: |
| - word-sense-disambiguation |
| - wsd |
| - wordnet |
| - oewn |
| - semcor |
| - semeval |
| - senseval |
| configs: |
| - config_name: en-senseval2-aw |
| data_files: |
| - split: test |
| path: data/en-senseval2-aw/test.parquet |
| - config_name: en-senseval3-aw |
| data_files: |
| - split: test |
| path: data/en-senseval3-aw/test.parquet |
| - config_name: en-semeval2007-aw |
| data_files: |
| - split: test |
| path: data/en-semeval2007-aw/test.parquet |
| - config_name: en-semeval2013-aw |
| data_files: |
| - split: test |
| path: data/en-semeval2013-aw/test.parquet |
| - config_name: en-semeval2015-aw |
| data_files: |
| - split: test |
| path: data/en-semeval2015-aw/test.parquet |
| - config_name: en-semcor |
| data_files: |
| - split: train |
| path: data/en-semcor/train.parquet |
| - config_name: en-wngt |
| data_files: |
| - split: train |
| path: data/en-wngt/train.parquet |
| - config_name: en-masc |
| data_files: |
| - split: train |
| path: data/en-masc/train.parquet |
| - config_name: en-senseval2_ls |
| data_files: |
| - split: train |
| path: data/en-senseval2_ls/train.parquet |
| - split: test |
| path: data/en-senseval2_ls/test.parquet |
| - config_name: en-senseval3_ls |
| data_files: |
| - split: train |
| path: data/en-senseval3_ls/train.parquet |
| - split: test |
| path: data/en-senseval3_ls/test.parquet |
| - config_name: en-semeval2007_t17_ls |
| data_files: |
| - split: test |
| path: data/en-semeval2007_t17_ls/test.parquet |
| --- |
| |
| # pywsd-datasets |
|
|
| Unified Word Sense Disambiguation benchmark datasets, normalized to **modern |
| `wn` lexicon sense IDs** (`oewn:2024` for English, OMW for other languages). |
|
|
| Companion to [pywsd](https://pypi.org/project/pywsd/) ≥ 1.3.0. |
|
|
| ## What's shipped (v0.2) |
|
|
| **English, test-only Raganato all-words benchmark:** |
|
|
| | Config | Instances | OEWN 2024 coverage | |
| |-----------------------|-----------|--------------------| |
| | `en-senseval2-aw` | 2,282 | 99.43 % | |
| | `en-senseval3-aw` | 1,850 | 99.51 % | |
| | `en-semeval2007-aw` | 455 | 99.78 % | |
| | `en-semeval2013-aw` | 1,644 | 100.00 % | |
| | `en-semeval2015-aw` | 1,022 | 99.32 % | |
|
|
| **English, training corpora (via UFSAC v2.1):** |
|
|
| | Config | Split | OEWN 2024 coverage | |
| |---------------------------|-------|--------------------| |
| | `en-semcor` | train | see coverage_report | |
| | `en-wngt` | train | see coverage_report | |
| | `en-masc` | train | see coverage_report | |
| | `en-senseval2_ls` | train + test | lexical-sample | |
| | `en-senseval3_ls` | train + test | lexical-sample | |
| | `en-semeval2007_t17_ls` | test | lexical-sample | |
|
|
| Run `python -m pywsd_datasets.scripts.coverage_report` locally to get |
| up-to-date OEWN resolution rates after rebuilding. |
|
|
| ## Install |
|
|
| ```bash |
| pip install pywsd-datasets |
| ``` |
|
|
| ## Use via HuggingFace `datasets` |
|
|
| ```python |
| from datasets import load_dataset |
| |
| # Raganato all-words evaluation set |
| ds = load_dataset("alvations/pywsd-datasets", "en-senseval2-aw") |
| |
| # SemCor training data |
| ds = load_dataset("alvations/pywsd-datasets", "en-semcor") |
| |
| ds["test"][0] if "test" in ds else ds["train"][0] |
| # {'instance_id': 'd000.s000.t000', 'dataset': 'senseval2_aw', |
| # 'split': 'test', 'lang': 'en', |
| # 'tokens': ['The', 'art', 'of', 'change-ringing', ...], |
| # 'target_idx': 1, 'target_lemma': 'art', 'target_pos': 'n', |
| # 'source_sense_id': 'art%1:09:00::', |
| # 'source_sense_system': 'pwn_sensekey_3.0', |
| # 'sense_ids_wordnet': ['oewn-05646832-n'], |
| # 'wordnet_lexicon': 'oewn:2024', ...} |
| ``` |
|
|
| ## Use via the loader package |
|
|
| ```python |
| from pywsd_datasets.loaders.raganato import iter_instances as iter_raganato |
| from pywsd_datasets.loaders.ufsac import iter_instances as iter_ufsac |
| |
| for inst in iter_raganato("senseval2"): |
| print(inst.target_lemma, inst.sense_ids_wordnet) |
| |
| for inst in iter_ufsac("semcor", "/path/to/ufsac-public-2.1"): |
| print(inst.target_lemma, inst.sense_ids_wordnet) |
| ``` |
|
|
| ## Rebuild locally |
|
|
| ```bash |
| pip install pywsd-datasets[dev] |
| |
| # Raganato only (always works, ~1 MB fetch from our GH release mirror) |
| python -m pywsd_datasets.scripts.build_all |
| |
| # With UFSAC corpora — download ufsac-public-2.1 separately (see below) |
| python -m pywsd_datasets.scripts.build_all \ |
| --ufsac-root ~/.cache/pywsd-datasets/ufsac/ufsac-public-2.1 |
| |
| # Coverage report across every built parquet: |
| python -m pywsd_datasets.scripts.coverage_report |
| ``` |
|
|
| ### UFSAC download |
|
|
| UFSAC v2.1 is distributed as a single Google Drive tarball |
| (`ufsac-public-2.1.tar.xz`, ~196 MB). Fetch with `gdown`: |
|
|
| ```bash |
| pip install gdown |
| mkdir -p ~/.cache/pywsd-datasets/ufsac |
| gdown 'https://drive.google.com/uc?id=1kwBMIDBTf6heRno9bdLvF-DahSLHIZyV' \ |
| -O ~/.cache/pywsd-datasets/ufsac/ufsac-public-2.1.tar.xz |
| cd ~/.cache/pywsd-datasets/ufsac && tar -xf ufsac-public-2.1.tar.xz |
| ``` |
|
|
| ## Schema |
|
|
| Every row follows [`WSDInstance`](src/pywsd_datasets/schema.py): |
|
|
| ``` |
| instance_id, dataset, split, task, lang, |
| tokens[], pos_tags[], lemmas[], |
| target_idx, target_lemma, target_pos, |
| source_sense_id, source_sense_system, |
| sense_ids_wordnet[], wordnet_lexicon, |
| doc_id, sent_id |
| ``` |
|
|
| `sense_ids_wordnet` is list-valued to handle multi-gold instances and any |
| PWN-3.0 key that splits into multiple OEWN 2024 synsets. |
|
|
| ## Multilingual / XL-WSD / BabelNet — deferred |
|
|
| `loaders/xl_wsd.py` exists as a stub and raises `NotImplementedError`. |
| `mappers/babelnet_to_wn.py` is similarly unused. **Why:** |
|
|
| * XL-WSD uses BabelNet synset IDs as gold labels; resolving them to |
| modern `wn` lexicon IDs requires the BabelNet → PWN 3.0 bridge file, |
| which is distributed **only with a BabelNet academic license**. |
| * XL-WSD itself is CC-BY-NC 4.0 — we don't redistribute the data. |
|
|
| Reviving this path requires (a) a BabelNet license, (b) loading |
| `bn_to_wn.txt` via `babelnet_to_wn.load_bn_to_pwn3_map()`, (c) selecting |
| per-language OMW lexicons via `mappers.omw_lookup.lexicon_for(lang)`, |
| then (d) chaining through `pwn3_to_oewn.pwn3_sensekey_to_wn(key, lexicon=...)`. |
| All four pieces are in place — wiring them is blocked on the BabelNet |
| mapping file. See the module docstrings for details. |
|
|
| ## Roadmap |
|
|
| * **v0.2** (this release): Raganato all-words evaluation + UFSAC training |
| corpora (SemCor, WNGT, MASC, Senseval lexical-sample). |
| * **v0.3** (planned): WiC (CC-BY-NC — loader-only), CoarseWSD-20. |
| * **Deferred:** XL-WSD multilingual (needs BabelNet academic license). |
|
|
| ## Citation |
|
|
| If you use these datasets please cite the original sources: |
|
|
| * Raganato, Camacho-Collados, Navigli (2017). *Word Sense Disambiguation: |
| A Unified Evaluation Framework and Empirical Comparison.* EACL. |
| * Vial, Lecouteux, Schwab (2018). *UFSAC: Unification of Sense Annotated |
| Corpora and Tools.* LREC. |
| * Plus the specific evaluation or training set paper (Senseval-2 / 3, |
| SemEval-2007 T17, SemEval-2013 T12, SemEval-2015 T13, SemCor, |
| WNGT/Princeton Gloss Corpus, MASC). |
|
|
| ## License |
|
|
| MIT for the code. Each dataset keeps its original license — see the source |
| papers. Raganato bundle and SemEval shared-task data are |
| research-unrestricted; UFSAC is MIT. |
|
|
| ## Sense-ID mapping details |
|
|
| PWN 3.0 sense keys are resolved against OEWN 2024 via |
| [`wn.compat.sensekey`](https://github.com/goodmami/wn). The few percent of |
| keys that fail to resolve are typically WN 3.0 synsets that OEWN split, |
| merged, or removed — those rows ship with an empty `sense_ids_wordnet` list |
| so the coverage report can flag them. Background: |
|
|
| * Kaf (2023). *Mapping Wordnets on the Fly with Permanent Sense Keys.* |
| arXiv:2303.01847. |
|
|
| ## Known issues |
|
|
| * The upstream Raganato zip at `http://lcl.uniroma1.it/wsdeval/` serves a |
| mismatched TLS cert; our loader prefers the mirror on this repo's |
| GitHub release assets and falls back to the original URL over HTTP. |
| * UFSAC v2.1 is distributed as a Google Drive tarball; the loader assumes |
| you have it unpacked locally. A future release may mirror it. |
|
|