Datasets:
Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,112 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
language:
|
| 3 |
+
- en
|
| 4 |
+
pretty_name: CANNOT
|
| 5 |
+
---
|
| 6 |
+
# Dataset Card for CANNOT
|
| 7 |
+
|
| 8 |
+
## Dataset Description
|
| 9 |
+
|
| 10 |
+
- **Homepage: https://github.com/dmlls/cannot-dataset**
|
| 11 |
+
- **Repository: https://github.com/dmlls/cannot-dataset**
|
| 12 |
+
- **Paper: tba**
|
| 13 |
+
|
| 14 |
+
### Dataset Summary
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
**CANNOT** is a dataset that focuses on negated textual pairs. It currently
|
| 18 |
+
contains **77,376 samples**, of which roughly of them are negated pairs of
|
| 19 |
+
sentences, and the other half are not (they are paraphrased versions of each
|
| 20 |
+
other).
|
| 21 |
+
|
| 22 |
+
The most frequent negation that appears in the dataset is verbal negation (e.g.,
|
| 23 |
+
will → won't), although it also contains pairs with antonyms (cold → hot).
|
| 24 |
+
|
| 25 |
+
### Languages
|
| 26 |
+
CANNOT includes exclusively texts in **English**.
|
| 27 |
+
|
| 28 |
+
## Dataset Structure
|
| 29 |
+
|
| 30 |
+
The dataset is given as a
|
| 31 |
+
[`.tsv`](https://en.wikipedia.org/wiki/Tab-separated_values) file with the
|
| 32 |
+
following structure:
|
| 33 |
+
|
| 34 |
+
| premise | hypothesis | label |
|
| 35 |
+
|:------------|:---------------------------------------------------|:-----:|
|
| 36 |
+
| A sentence. | An equivalent, non-negated sentence (paraphrased). | 0 |
|
| 37 |
+
| A sentence. | The sentence negated. | 1 |
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
The dataset can be easily loaded into a Pandas DataFrame by running:
|
| 41 |
+
|
| 42 |
+
```Python
|
| 43 |
+
import pandas as pd
|
| 44 |
+
|
| 45 |
+
dataset = pd.read_csv('negation_dataset_v1.0.tsv', sep='\t')
|
| 46 |
+
|
| 47 |
+
```
|
| 48 |
+
|
| 49 |
+
## Dataset Creation
|
| 50 |
+
|
| 51 |
+
|
| 52 |
+
The dataset has been created by cleaning up and merging the following datasets:
|
| 53 |
+
|
| 54 |
+
1. _Not another Negation Benchmark: The NaN-NLI Test Suite for Sub-clausal
|
| 55 |
+
Negation_ (see
|
| 56 |
+
[`datasets/nan-nli`](https://github.com/dmlls/cannot-dataset/tree/main/datasets/nan-nli)).
|
| 57 |
+
|
| 58 |
+
2. _GLUE Diagnostic Dataset_ (see
|
| 59 |
+
[`datasets/glue-diagnostic`](https://github.com/dmlls/cannot-dataset/tree/main/datasets/glue-diagnostic)).
|
| 60 |
+
|
| 61 |
+
3. _Automated Fact-Checking of Claims from Wikipedia_ (see
|
| 62 |
+
[`datasets/wikifactcheck-english`](https://github.com/dmlls/cannot-dataset/tree/main/datasets/wikifactcheck-english)).
|
| 63 |
+
|
| 64 |
+
4. _From Group to Individual Labels Using Deep Features_ (see
|
| 65 |
+
[`datasets/sentiment-labelled-sentences`](https://github.com/dmlls/cannot-dataset/tree/main/datasets/sentiment-labelled-sentences)).
|
| 66 |
+
In this case, the negated sentences were obtained by using the Python module
|
| 67 |
+
[`negate`](https://github.com/dmlls/negate).
|
| 68 |
+
|
| 69 |
+
|
| 70 |
+
Additionally, for each of the negated samples, another pair of non-negated
|
| 71 |
+
sentences has been added by paraphrasing them with the pre-trained model
|
| 72 |
+
[`🤗tuner007/pegasus_paraphrase`](https://huggingface.co/tuner007/pegasus_paraphrase).
|
| 73 |
+
|
| 74 |
+
Furthermore, the dataset from _It Is Not Easy To Detect Paraphrases: Analysing
|
| 75 |
+
Semantic Similarity With Antonyms and Negation Using the New SemAntoNeg
|
| 76 |
+
Benchmark_ (see
|
| 77 |
+
[`datasets/antonym-substitution`](https://github.com/dmlls/cannot-dataset/tree/main/datasets/antonym-substitution))
|
| 78 |
+
has also been included. This dataset already provides both the paraphrased and
|
| 79 |
+
negated version for each premise, so no further processing was needed.
|
| 80 |
+
|
| 81 |
+
Finally, the swapped version of each pair (premise ⇋ hypothesis) has also been
|
| 82 |
+
included, and any duplicates have been removed.
|
| 83 |
+
|
| 84 |
+
The contribution of each of these individual datasets to the final CANNOT
|
| 85 |
+
dataset is:
|
| 86 |
+
|
| 87 |
+
| Dataset | Samples |
|
| 88 |
+
|:--------------------------------------------------------------------------|-----------:|
|
| 89 |
+
| Not another Negation Benchmark | 118 |
|
| 90 |
+
| GLUE Diagnostic Dataset | 154 |
|
| 91 |
+
| Automated Fact-Checking of Claims from Wikipedia | 14,970 |
|
| 92 |
+
| From Group to Individual Labels Using Deep Features | 2,110 |
|
| 93 |
+
| It Is Not Easy To Detect Paraphrases | 8,597 |
|
| 94 |
+
| <p align="right"><b>Total</b></p> | **25,949** |
|
| 95 |
+
|
| 96 |
+
_Note_: The numbers above include only the original queries present in the
|
| 97 |
+
datasets.
|
| 98 |
+
|
| 99 |
+
|
| 100 |
+
## Additional Information
|
| 101 |
+
|
| 102 |
+
### Licensing Information
|
| 103 |
+
|
| 104 |
+
TODO
|
| 105 |
+
|
| 106 |
+
### Citation Information
|
| 107 |
+
|
| 108 |
+
tba
|
| 109 |
+
|
| 110 |
+
### Contributions
|
| 111 |
+
|
| 112 |
+
Contributions to the dataset can be submitted through the [project repository](https://github.com/dmlls/cannot-dataset).
|