id int64 599M 3.29B | url stringlengths 58 61 | html_url stringlengths 46 51 | number int64 1 7.72k | title stringlengths 1 290 | state stringclasses 2
values | comments int64 0 70 | created_at timestamp[s]date 2020-04-14 10:18:02 2025-08-05 09:28:51 | updated_at timestamp[s]date 2020-04-27 16:04:17 2025-08-05 11:39:56 | closed_at timestamp[s]date 2020-04-14 12:01:40 2025-08-01 05:15:45 ⌀ | user_login stringlengths 3 26 | labels listlengths 0 4 | body stringlengths 0 228k ⌀ | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
757,677,188 | https://api.github.com/repos/huggingface/datasets/issues/1160 | https://github.com/huggingface/datasets/pull/1160 | 1,160 | adding TabFact dataset | closed | 2 | 2020-12-05T13:05:52 | 2020-12-09T11:41:39 | 2020-12-09T09:12:41 | patil-suraj | [] | Adding TabFact: A Large-scale Dataset for Table-based Fact Verification.
https://github.com/wenhuchen/Table-Fact-Checking
- The tables are stored as individual csv files, so need to download 16,573 🤯 csv files. As a result the `datasets_infos.json` file is huge (6.62 MB).
- Original dataset has nested structur... | true |
757,661,128 | https://api.github.com/repos/huggingface/datasets/issues/1159 | https://github.com/huggingface/datasets/pull/1159 | 1,159 | Add Roman Urdu dataset | closed | 0 | 2020-12-05T11:36:43 | 2020-12-07T13:41:21 | 2020-12-07T09:59:03 | jaketae | [] | This PR adds the [Roman Urdu dataset](https://archive.ics.uci.edu/ml/datasets/Roman+Urdu+Data+Set#). | true |
757,658,926 | https://api.github.com/repos/huggingface/datasets/issues/1158 | https://github.com/huggingface/datasets/pull/1158 | 1,158 | Add BBC Hindi NLI Dataset | closed | 7 | 2020-12-05T11:25:34 | 2021-02-05T09:48:31 | 2021-02-05T09:48:31 | avinsit123 | [] | # Dataset Card for BBC Hindi NLI Dataset
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances... | true |
757,657,888 | https://api.github.com/repos/huggingface/datasets/issues/1157 | https://github.com/huggingface/datasets/pull/1157 | 1,157 | Add dataset XhosaNavy English -Xhosa | closed | 0 | 2020-12-05T11:19:54 | 2020-12-07T09:11:33 | 2020-12-07T09:11:33 | spatil6 | [] | Add dataset XhosaNavy English -Xhosa
More info : http://opus.nlpl.eu/XhosaNavy.php | true |
757,656,094 | https://api.github.com/repos/huggingface/datasets/issues/1156 | https://github.com/huggingface/datasets/pull/1156 | 1,156 | add telugu-news corpus | closed | 0 | 2020-12-05T11:07:56 | 2020-12-07T09:08:48 | 2020-12-07T09:08:48 | oostopitre | [] | Adding Telugu News Corpus to datasets. | true |
757,652,517 | https://api.github.com/repos/huggingface/datasets/issues/1155 | https://github.com/huggingface/datasets/pull/1155 | 1,155 | Add BSD | closed | 5 | 2020-12-05T10:43:48 | 2020-12-07T09:27:46 | 2020-12-07T09:27:46 | j-chim | [] | This PR adds BSD, the Japanese-English business dialogue corpus by
[Rikters et al., 2020](https://www.aclweb.org/anthology/D19-5204.pdf). | true |
757,651,669 | https://api.github.com/repos/huggingface/datasets/issues/1154 | https://github.com/huggingface/datasets/pull/1154 | 1,154 | Opus sardware | closed | 0 | 2020-12-05T10:38:02 | 2020-12-05T17:05:45 | 2020-12-05T17:05:45 | spatil6 | [] | Added Opus sardware dataset for machine translation English to Sardinian.
for more info : http://opus.nlpl.eu/sardware.php | true |
757,643,302 | https://api.github.com/repos/huggingface/datasets/issues/1153 | https://github.com/huggingface/datasets/pull/1153 | 1,153 | Adding dataset for proto_qa in huggingface datasets library | closed | 0 | 2020-12-05T09:43:28 | 2020-12-05T18:53:10 | 2020-12-05T18:53:10 | bpatidar | [] | Added dataset for ProtoQA: A Question Answering Dataset for Prototypical Common-Sense Reasoning
Followed all steps for adding a new dataset. | true |
757,640,506 | https://api.github.com/repos/huggingface/datasets/issues/1152 | https://github.com/huggingface/datasets/pull/1152 | 1,152 | hindi discourse analysis dataset commit | closed | 9 | 2020-12-05T09:24:01 | 2020-12-14T19:44:48 | 2020-12-14T19:44:48 | duttahritwik | [] | true | |
757,517,092 | https://api.github.com/repos/huggingface/datasets/issues/1151 | https://github.com/huggingface/datasets/pull/1151 | 1,151 | adding psc dataset | closed | 0 | 2020-12-05T02:40:01 | 2020-12-09T11:38:41 | 2020-12-09T11:38:41 | abecadel | [] | true | |
757,512,441 | https://api.github.com/repos/huggingface/datasets/issues/1150 | https://github.com/huggingface/datasets/pull/1150 | 1,150 | adding dyk dataset | closed | 0 | 2020-12-05T02:11:42 | 2020-12-05T16:52:19 | 2020-12-05T16:52:19 | abecadel | [] | true | |
757,504,068 | https://api.github.com/repos/huggingface/datasets/issues/1149 | https://github.com/huggingface/datasets/pull/1149 | 1,149 | Fix typo in the comment in _info function | closed | 0 | 2020-12-05T01:26:20 | 2020-12-05T16:19:26 | 2020-12-05T16:19:26 | vinaykudari | [] | true | |
757,503,918 | https://api.github.com/repos/huggingface/datasets/issues/1148 | https://github.com/huggingface/datasets/pull/1148 | 1,148 | adding polemo2 dataset | closed | 0 | 2020-12-05T01:25:29 | 2020-12-05T16:51:39 | 2020-12-05T16:51:39 | abecadel | [] | true | |
757,502,199 | https://api.github.com/repos/huggingface/datasets/issues/1147 | https://github.com/huggingface/datasets/pull/1147 | 1,147 | Vinay/add/telugu books | closed | 0 | 2020-12-05T01:17:02 | 2020-12-05T16:36:04 | 2020-12-05T16:36:04 | vinaykudari | [] | Real data tests are failing as this dataset needs to be manually downloaded | true |
757,498,565 | https://api.github.com/repos/huggingface/datasets/issues/1146 | https://github.com/huggingface/datasets/pull/1146 | 1,146 | Add LINNAEUS | closed | 0 | 2020-12-05T01:01:09 | 2020-12-05T16:35:53 | 2020-12-05T16:35:53 | edugp | [] | true | |
757,477,349 | https://api.github.com/repos/huggingface/datasets/issues/1145 | https://github.com/huggingface/datasets/pull/1145 | 1,145 | Add Species-800 | closed | 4 | 2020-12-04T23:44:51 | 2022-01-13T03:09:20 | 2020-12-05T16:35:01 | edugp | [] | true | |
757,452,831 | https://api.github.com/repos/huggingface/datasets/issues/1144 | https://github.com/huggingface/datasets/pull/1144 | 1,144 | Add JFLEG | closed | 2 | 2020-12-04T22:36:38 | 2020-12-06T18:16:04 | 2020-12-06T18:16:04 | j-chim | [] | This PR adds [JFLEG ](https://www.aclweb.org/anthology/E17-2037/), an English grammatical error correction benchmark.
The tests were successful on real data, although it would be great if I can get some guidance on the **dummy data**. Basically, **for each source sentence there are 4 possible gold standard target s... | true |
757,448,920 | https://api.github.com/repos/huggingface/datasets/issues/1143 | https://github.com/huggingface/datasets/pull/1143 | 1,143 | Add the Winograd Schema Challenge | closed | 0 | 2020-12-04T22:26:59 | 2020-12-09T15:11:31 | 2020-12-09T09:32:34 | joeddav | [] | Adds the Winograd Schema Challenge, including configs for the more canonical wsc273 as well as wsc285 with 12 new examples.
- https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html
The data format was a bit of a nightmare but I think I got it to a workable format. | true |
757,413,920 | https://api.github.com/repos/huggingface/datasets/issues/1142 | https://github.com/huggingface/datasets/pull/1142 | 1,142 | Fix PerSenT | closed | 0 | 2020-12-04T21:21:02 | 2020-12-14T13:39:34 | 2020-12-14T13:39:34 | jeromeku | [] | New PR for dataset PerSenT | true |
757,411,057 | https://api.github.com/repos/huggingface/datasets/issues/1141 | https://github.com/huggingface/datasets/pull/1141 | 1,141 | Add GitHub version of ETH Py150 Corpus | closed | 2 | 2020-12-04T21:16:08 | 2020-12-09T18:32:44 | 2020-12-07T10:00:24 | bharatr21 | [] | Add the redistributable version of **ETH Py150 Corpus** | true |
757,399,142 | https://api.github.com/repos/huggingface/datasets/issues/1140 | https://github.com/huggingface/datasets/pull/1140 | 1,140 | Add Urdu Sentiment Corpus (USC). | closed | 2 | 2020-12-04T20:55:27 | 2020-12-07T03:27:23 | 2020-12-07T03:27:23 | chaitnayabasava | [] | Added Urdu Sentiment Corpus. More details about the dataset over <a href="https://github.com/MuhammadYaseenKhan/Urdu-Sentiment-Corpus">here</a>. | true |
757,393,158 | https://api.github.com/repos/huggingface/datasets/issues/1139 | https://github.com/huggingface/datasets/pull/1139 | 1,139 | Add ReFreSD dataset | closed | 3 | 2020-12-04T20:45:11 | 2020-12-16T16:01:18 | 2020-12-16T16:01:18 | mpariente | [] | This PR adds the **ReFreSD dataset**.
The original data is hosted [on this github repo](https://github.com/Elbria/xling-SemDiv) and we use the `REFreSD_rationale` to expose all the data.
Need feedback on:
- I couldn't generate the dummy data. The file we download is a tsv file, but without extension, I suppose... | true |
757,378,406 | https://api.github.com/repos/huggingface/datasets/issues/1138 | https://github.com/huggingface/datasets/pull/1138 | 1,138 | updated after the class name update | closed | 0 | 2020-12-04T20:19:43 | 2020-12-05T15:43:32 | 2020-12-05T15:43:32 | timpal0l | [] | @lhoestq <--- | true |
757,358,145 | https://api.github.com/repos/huggingface/datasets/issues/1137 | https://github.com/huggingface/datasets/pull/1137 | 1,137 | add wmt mlqe 2020 shared task | closed | 1 | 2020-12-04T19:45:34 | 2020-12-06T19:59:44 | 2020-12-06T19:53:46 | VictorSanh | [] | First commit for Shared task 1 (wmt_mlqw_task1) of WMT20 MLQE (quality estimation of machine translation)
Note that I copied the tags in the README for only one (of the 7 configurations): `en-de`.
There is one configuration for each pair of languages. | true |
757,341,607 | https://api.github.com/repos/huggingface/datasets/issues/1136 | https://github.com/huggingface/datasets/pull/1136 | 1,136 | minor change in description in paws-x.py and updated dataset_infos | closed | 0 | 2020-12-04T19:17:49 | 2020-12-06T18:02:57 | 2020-12-06T18:02:57 | bhavitvyamalik | [] | true | |
757,325,741 | https://api.github.com/repos/huggingface/datasets/issues/1135 | https://github.com/huggingface/datasets/pull/1135 | 1,135 | added paws | closed | 0 | 2020-12-04T18:52:38 | 2020-12-09T17:17:13 | 2020-12-09T17:17:13 | bhavitvyamalik | [] | Updating README and tags for dataset card in a while | true |
757,317,651 | https://api.github.com/repos/huggingface/datasets/issues/1134 | https://github.com/huggingface/datasets/pull/1134 | 1,134 | adding xquad-r dataset | closed | 0 | 2020-12-04T18:39:13 | 2020-12-05T16:50:47 | 2020-12-05T16:50:47 | manandey | [] | true | |
757,307,660 | https://api.github.com/repos/huggingface/datasets/issues/1133 | https://github.com/huggingface/datasets/pull/1133 | 1,133 | Adding XQUAD-R Dataset | closed | 0 | 2020-12-04T18:22:29 | 2020-12-04T18:28:54 | 2020-12-04T18:28:49 | manandey | [] | true | |
757,301,368 | https://api.github.com/repos/huggingface/datasets/issues/1132 | https://github.com/huggingface/datasets/pull/1132 | 1,132 | Add Urdu Sentiment Corpus (USC). | closed | 0 | 2020-12-04T18:12:24 | 2020-12-04T20:52:48 | 2020-12-04T20:52:48 | chaitnayabasava | [] | Added Urdu Sentiment Corpus. More details about the dataset over <a href="https://github.com/MuhammadYaseenKhan/Urdu-Sentiment-Corpus">here</a>. | true |
757,278,341 | https://api.github.com/repos/huggingface/datasets/issues/1131 | https://github.com/huggingface/datasets/pull/1131 | 1,131 | Adding XQUAD-R Dataset | closed | 0 | 2020-12-04T17:35:43 | 2020-12-04T18:27:22 | 2020-12-04T18:27:22 | manandey | [] | true | |
757,265,075 | https://api.github.com/repos/huggingface/datasets/issues/1130 | https://github.com/huggingface/datasets/pull/1130 | 1,130 | adding discovery | closed | 1 | 2020-12-04T17:16:54 | 2020-12-14T13:03:14 | 2020-12-14T13:03:14 | sileod | [] | true | |
757,255,492 | https://api.github.com/repos/huggingface/datasets/issues/1129 | https://github.com/huggingface/datasets/pull/1129 | 1,129 | Adding initial version of cord-19 dataset | closed | 5 | 2020-12-04T17:03:17 | 2021-02-09T10:22:35 | 2021-02-09T10:18:06 | ggdupont | [] | Initial version only reading the metadata in CSV.
### Checklist:
- [x] Create the dataset script /datasets/my_dataset/my_dataset.py using the template
- [x] Fill the _DESCRIPTION and _CITATION variables
- [x] Implement _infos(), _split_generators() and _generate_examples()
- [x] Make sure that the BUILDER_CONFIG... | true |
757,245,404 | https://api.github.com/repos/huggingface/datasets/issues/1128 | https://github.com/huggingface/datasets/pull/1128 | 1,128 | Add xquad-r dataset | closed | 0 | 2020-12-04T16:48:53 | 2020-12-04T18:14:30 | 2020-12-04T18:14:26 | manandey | [] | true | |
757,229,684 | https://api.github.com/repos/huggingface/datasets/issues/1127 | https://github.com/huggingface/datasets/pull/1127 | 1,127 | Add wikiqaar dataset | closed | 0 | 2020-12-04T16:26:18 | 2020-12-07T16:39:41 | 2020-12-07T16:39:41 | zaidalyafeai | [] | Arabic Wiki Question Answering Corpus. | true |
757,197,735 | https://api.github.com/repos/huggingface/datasets/issues/1126 | https://github.com/huggingface/datasets/pull/1126 | 1,126 | Adding babi dataset | closed | 3 | 2020-12-04T15:42:34 | 2021-03-30T09:44:04 | 2021-03-30T09:44:04 | thomwolf | [] | Adding the English version of bAbI.
Samples are taken from ParlAI for consistency with the main users at the moment.
Supersede #945 (problem with the rebase) and adresses the issues mentioned in the review (dummy data are smaller now and code comments are fixed). | true |
757,194,531 | https://api.github.com/repos/huggingface/datasets/issues/1125 | https://github.com/huggingface/datasets/pull/1125 | 1,125 | Add Urdu fake news dataset. | closed | 3 | 2020-12-04T15:38:17 | 2020-12-07T03:21:05 | 2020-12-07T03:21:05 | chaitnayabasava | [] | Added Urdu fake news dataset. More information about the dataset can be found <a href="https://github.com/MaazAmjad/Datasets-for-Urdu-news">here</a>. | true |
757,186,983 | https://api.github.com/repos/huggingface/datasets/issues/1124 | https://github.com/huggingface/datasets/pull/1124 | 1,124 | Add Xitsonga Ner | closed | 1 | 2020-12-04T15:27:44 | 2020-12-06T18:31:35 | 2020-12-06T18:31:35 | yvonnegitau | [] | Clean Xitsonga Ner PR | true |
757,181,014 | https://api.github.com/repos/huggingface/datasets/issues/1123 | https://github.com/huggingface/datasets/pull/1123 | 1,123 | adding cdt dataset | closed | 2 | 2020-12-04T15:19:36 | 2020-12-04T17:05:56 | 2020-12-04T17:05:56 | abecadel | [] | true | |
757,176,172 | https://api.github.com/repos/huggingface/datasets/issues/1122 | https://github.com/huggingface/datasets/pull/1122 | 1,122 | Add Urdu fake news. | closed | 0 | 2020-12-04T15:13:10 | 2020-12-04T15:20:07 | 2020-12-04T15:20:07 | chaitnayabasava | [] | Added Urdu fake news dataset. More information about the dataset can be found <a href="https://github.com/MaazAmjad/Datasets-for-Urdu-news">here</a>. | true |
757,169,944 | https://api.github.com/repos/huggingface/datasets/issues/1121 | https://github.com/huggingface/datasets/pull/1121 | 1,121 | adding cdt dataset | closed | 0 | 2020-12-04T15:04:33 | 2020-12-04T15:16:49 | 2020-12-04T15:16:49 | abecadel | [] | true | |
757,166,342 | https://api.github.com/repos/huggingface/datasets/issues/1120 | https://github.com/huggingface/datasets/pull/1120 | 1,120 | Add conda environment activation | closed | 0 | 2020-12-04T14:59:43 | 2020-12-04T18:34:48 | 2020-12-04T16:40:57 | parmarsuraj99 | [] | Added activation of Conda environment before installing. | true |
757,156,781 | https://api.github.com/repos/huggingface/datasets/issues/1119 | https://github.com/huggingface/datasets/pull/1119 | 1,119 | Add Google Great Code Dataset | closed | 0 | 2020-12-04T14:46:28 | 2020-12-06T17:33:14 | 2020-12-06T17:33:13 | abhishekkrthakur | [] | true | |
757,142,350 | https://api.github.com/repos/huggingface/datasets/issues/1118 | https://github.com/huggingface/datasets/pull/1118 | 1,118 | Add Tashkeela dataset | closed | 2 | 2020-12-04T14:26:18 | 2020-12-04T15:47:01 | 2020-12-04T15:46:51 | zaidalyafeai | [] | Arabic Vocalized Words Dataset. | true |
757,133,789 | https://api.github.com/repos/huggingface/datasets/issues/1117 | https://github.com/huggingface/datasets/pull/1117 | 1,117 | Fix incorrect MRQA train+SQuAD URL | closed | 3 | 2020-12-04T14:14:26 | 2020-12-06T17:14:11 | 2020-12-06T17:14:10 | yuxiang-wu | [] | Fix issue #1115
| true |
757,133,502 | https://api.github.com/repos/huggingface/datasets/issues/1116 | https://github.com/huggingface/datasets/pull/1116 | 1,116 | add dbpedia_14 dataset | closed | 5 | 2020-12-04T14:13:59 | 2020-12-07T10:06:54 | 2020-12-05T15:36:23 | hfawaz | [] | This dataset corresponds to the DBpedia dataset requested in https://github.com/huggingface/datasets/issues/353. | true |
757,127,527 | https://api.github.com/repos/huggingface/datasets/issues/1115 | https://github.com/huggingface/datasets/issues/1115 | 1,115 | Incorrect URL for MRQA SQuAD train subset | closed | 1 | 2020-12-04T14:05:24 | 2020-12-06T17:14:22 | 2020-12-06T17:14:22 | yuxiang-wu | [] | https://github.com/huggingface/datasets/blob/4ef4c8f8b7a60e35c6fa21115fca9faae91c9f74/datasets/mrqa/mrqa.py#L53
The URL for `train+SQuAD` subset of MRQA points to the dev set instead of train set. It should be `https://s3.us-east-2.amazonaws.com/mrqa/release/v2/train/SQuAD.jsonl.gz`. | false |
757,123,638 | https://api.github.com/repos/huggingface/datasets/issues/1114 | https://github.com/huggingface/datasets/pull/1114 | 1,114 | Add sesotho ner corpus | closed | 0 | 2020-12-04T13:59:41 | 2020-12-04T15:02:07 | 2020-12-04T15:02:07 | yvonnegitau | [] | Clean Sesotho PR | true |
757,115,557 | https://api.github.com/repos/huggingface/datasets/issues/1113 | https://github.com/huggingface/datasets/pull/1113 | 1,113 | add qed | closed | 0 | 2020-12-04T13:47:57 | 2020-12-05T15:46:21 | 2020-12-05T15:41:57 | patil-suraj | [] | adding QED: Dataset for Explanations in Question Answering
https://github.com/google-research-datasets/QED
https://arxiv.org/abs/2009.06354 | true |
757,108,151 | https://api.github.com/repos/huggingface/datasets/issues/1112 | https://github.com/huggingface/datasets/pull/1112 | 1,112 | Initial version of cord-19 dataset from AllenAI with only the abstract | closed | 1 | 2020-12-04T13:36:39 | 2020-12-04T16:16:40 | 2020-12-04T16:16:24 | ggdupont | [] | Initial version only reading the metadata in CSV.
### Checklist:
- [x] Create the dataset script /datasets/my_dataset/my_dataset.py using the template
- [x] Fill the _DESCRIPTION and _CITATION variables
- [x] Implement _infos(), _split_generators() and _generate_examples()
- [x] Make sure that the BUILDER_CONFIG... | true |
757,083,266 | https://api.github.com/repos/huggingface/datasets/issues/1111 | https://github.com/huggingface/datasets/pull/1111 | 1,111 | Add Siswati Ner corpus | closed | 0 | 2020-12-04T12:57:31 | 2020-12-04T14:43:01 | 2020-12-04T14:43:00 | yvonnegitau | [] | Clean Siswati PR | true |
757,082,677 | https://api.github.com/repos/huggingface/datasets/issues/1110 | https://github.com/huggingface/datasets/issues/1110 | 1,110 | Using a feature named "_type" fails with certain operations | closed | 1 | 2020-12-04T12:56:33 | 2022-01-14T18:07:00 | 2022-01-14T18:07:00 | dcfidalgo | [] | A column named `_type` leads to a `TypeError: unhashable type: 'dict'` for certain operations:
```python
from datasets import Dataset, concatenate_datasets
ds = Dataset.from_dict({"_type": ["whatever"]}).map()
concatenate_datasets([ds])
# or simply
Dataset(ds._data)
```
Context: We are using datasets to persi... | false |
757,055,702 | https://api.github.com/repos/huggingface/datasets/issues/1109 | https://github.com/huggingface/datasets/pull/1109 | 1,109 | add woz_dialogue | closed | 0 | 2020-12-04T12:13:07 | 2020-12-05T15:41:23 | 2020-12-05T15:40:18 | patil-suraj | [] | Adding Wizard-of-Oz task oriented dialogue dataset
https://github.com/nmrksic/neural-belief-tracker/tree/master/data/woz
https://arxiv.org/abs/1604.04562 | true |
757,054,732 | https://api.github.com/repos/huggingface/datasets/issues/1108 | https://github.com/huggingface/datasets/pull/1108 | 1,108 | Add Sepedi NER corpus | closed | 0 | 2020-12-04T12:11:24 | 2020-12-04T14:39:00 | 2020-12-04T14:39:00 | yvonnegitau | [] | Finally a clean PR for Sepedi | true |
757,031,179 | https://api.github.com/repos/huggingface/datasets/issues/1107 | https://github.com/huggingface/datasets/pull/1107 | 1,107 | Add arsentd_lev dataset | closed | 1 | 2020-12-04T11:31:04 | 2020-12-05T15:38:09 | 2020-12-05T15:38:09 | moussaKam | [] | Add The Arabic Sentiment Twitter Dataset for Levantine dialect (ArSenTD-LEV)
Paper: [ArSentD-LEV: A Multi-Topic Corpus for Target-based Sentiment Analysis in Arabic Levantine Tweets](https://arxiv.org/abs/1906.01830)
Homepage: http://oma-project.com/ | true |
757,027,158 | https://api.github.com/repos/huggingface/datasets/issues/1106 | https://github.com/huggingface/datasets/pull/1106 | 1,106 | Add Urdu fake news | closed | 0 | 2020-12-04T11:24:14 | 2020-12-04T14:21:12 | 2020-12-04T14:21:12 | chaitnayabasava | [] | Added Urdu fake news dataset. More information about the dataset can be found <a href="https://github.com/MaazAmjad/Datasets-for-Urdu-news">here</a>. | true |
757,024,162 | https://api.github.com/repos/huggingface/datasets/issues/1105 | https://github.com/huggingface/datasets/pull/1105 | 1,105 | add xquad_r dataset | closed | 2 | 2020-12-04T11:19:35 | 2020-12-04T16:37:00 | 2020-12-04T16:37:00 | manandey | [] | true | |
757,020,934 | https://api.github.com/repos/huggingface/datasets/issues/1104 | https://github.com/huggingface/datasets/pull/1104 | 1,104 | add TLC | closed | 0 | 2020-12-04T11:14:58 | 2020-12-04T14:29:23 | 2020-12-04T14:29:23 | chameleonTK | [] | Added TLC dataset | true |
757,016,820 | https://api.github.com/repos/huggingface/datasets/issues/1103 | https://github.com/huggingface/datasets/issues/1103 | 1,103 | Add support to download kaggle datasets | closed | 2 | 2020-12-04T11:08:37 | 2023-07-20T15:22:24 | 2023-07-20T15:22:23 | abhishekkrthakur | [
"enhancement"
] | We can use API key | false |
757,016,515 | https://api.github.com/repos/huggingface/datasets/issues/1102 | https://github.com/huggingface/datasets/issues/1102 | 1,102 | Add retries to download manager | closed | 0 | 2020-12-04T11:08:11 | 2020-12-22T15:34:06 | 2020-12-22T15:34:06 | abhishekkrthakur | [
"enhancement"
] | false | |
757,009,226 | https://api.github.com/repos/huggingface/datasets/issues/1101 | https://github.com/huggingface/datasets/pull/1101 | 1,101 | Add Wikicorpus dataset | closed | 1 | 2020-12-04T10:57:26 | 2020-12-09T18:13:10 | 2020-12-09T18:13:09 | albertvillanova | [] | Add dataset. | true |
756,998,433 | https://api.github.com/repos/huggingface/datasets/issues/1100 | https://github.com/huggingface/datasets/pull/1100 | 1,100 | Urdu fake news | closed | 0 | 2020-12-04T10:41:20 | 2020-12-04T11:19:00 | 2020-12-04T11:19:00 | chaitnayabasava | [] | Added Bend the Truth urdu fake news dataset. More inforation <a href="https://github.com/MaazAmjad/Datasets-for-Urdu-news">here</a>. | true |
756,993,540 | https://api.github.com/repos/huggingface/datasets/issues/1099 | https://github.com/huggingface/datasets/pull/1099 | 1,099 | Add tamilmixsentiment data | closed | 0 | 2020-12-04T10:34:07 | 2020-12-06T06:32:22 | 2020-12-05T16:48:33 | jamespaultg | [] | true | |
756,975,414 | https://api.github.com/repos/huggingface/datasets/issues/1098 | https://github.com/huggingface/datasets/pull/1098 | 1,098 | Add ToTTo Dataset | closed | 0 | 2020-12-04T10:07:25 | 2020-12-04T13:38:20 | 2020-12-04T13:38:19 | abhishekkrthakur | [] | Adds a brand new table to text dataset: https://github.com/google-research-datasets/ToTTo | true |
756,955,729 | https://api.github.com/repos/huggingface/datasets/issues/1097 | https://github.com/huggingface/datasets/pull/1097 | 1,097 | Add MSRA NER labels | closed | 0 | 2020-12-04T09:38:16 | 2020-12-04T13:31:59 | 2020-12-04T13:31:58 | JetRunner | [] | Fixes #940 | true |
756,952,461 | https://api.github.com/repos/huggingface/datasets/issues/1096 | https://github.com/huggingface/datasets/pull/1096 | 1,096 | FIX matinf link in ADD_NEW_DATASET.md | closed | 0 | 2020-12-04T09:33:25 | 2020-12-04T14:25:35 | 2020-12-04T14:25:35 | moussaKam | [] | true | |
756,934,964 | https://api.github.com/repos/huggingface/datasets/issues/1095 | https://github.com/huggingface/datasets/pull/1095 | 1,095 | Add TupleInf Open IE Dataset | closed | 2 | 2020-12-04T09:08:07 | 2020-12-04T15:40:54 | 2020-12-04T15:40:54 | mattbui | [] | For more information: https://allenai.org/data/tuple-ie | true |
756,927,060 | https://api.github.com/repos/huggingface/datasets/issues/1094 | https://github.com/huggingface/datasets/pull/1094 | 1,094 | add urdu fake news dataset | closed | 0 | 2020-12-04T08:57:38 | 2020-12-04T09:20:56 | 2020-12-04T09:20:56 | chaitnayabasava | [] | Added Urdu fake news dataset. The dataset can be found <a href="https://github.com/MaazAmjad/Datasets-for-Urdu-news">here</a>. | true |
756,916,565 | https://api.github.com/repos/huggingface/datasets/issues/1093 | https://github.com/huggingface/datasets/pull/1093 | 1,093 | Add NCBI Disease Corpus dataset | closed | 0 | 2020-12-04T08:42:32 | 2020-12-04T11:15:12 | 2020-12-04T11:15:12 | edugp | [] | true | |
756,913,134 | https://api.github.com/repos/huggingface/datasets/issues/1092 | https://github.com/huggingface/datasets/pull/1092 | 1,092 | Add Coached Conversation Preference Dataset | closed | 0 | 2020-12-04T08:36:49 | 2020-12-20T13:34:00 | 2020-12-04T13:49:50 | vineeths96 | [] | Adding [Coached Conversation Preference Dataset](https://research.google/tools/datasets/coached-conversational-preference-elicitation/)
| true |
756,841,254 | https://api.github.com/repos/huggingface/datasets/issues/1091 | https://github.com/huggingface/datasets/pull/1091 | 1,091 | Add Google wellformed query dataset | closed | 1 | 2020-12-04T06:25:54 | 2020-12-06T17:43:03 | 2020-12-06T17:43:02 | thevasudevgupta | [] | This pull request will add Google wellformed_query dataset. Link of dataset is https://github.com/google-research-datasets/query-wellformedness | true |
756,825,941 | https://api.github.com/repos/huggingface/datasets/issues/1090 | https://github.com/huggingface/datasets/pull/1090 | 1,090 | add thaisum | closed | 0 | 2020-12-04T05:54:48 | 2020-12-04T11:16:06 | 2020-12-04T11:16:06 | cstorm125 | [] | ThaiSum, a large-scale corpus for Thai text summarization obtained from several online news websites namely Thairath, ThaiPBS, Prachathai, and The Standard. This dataset consists of over 350,000 article and summary pairs written by journalists. We evaluate the performance of various existing summarization models on Tha... | true |
756,823,690 | https://api.github.com/repos/huggingface/datasets/issues/1089 | https://github.com/huggingface/datasets/pull/1089 | 1,089 | add sharc_modified | closed | 0 | 2020-12-04T05:49:49 | 2020-12-04T10:41:30 | 2020-12-04T10:31:44 | patil-suraj | [] | Adding modified ShARC dataset https://github.com/nikhilweee/neural-conv-qa | true |
756,822,017 | https://api.github.com/repos/huggingface/datasets/issues/1088 | https://github.com/huggingface/datasets/pull/1088 | 1,088 | add xquad_r dataset | closed | 0 | 2020-12-04T05:45:55 | 2020-12-04T10:58:13 | 2020-12-04T10:47:01 | manandey | [] | true | |
756,794,430 | https://api.github.com/repos/huggingface/datasets/issues/1087 | https://github.com/huggingface/datasets/pull/1087 | 1,087 | Add Big Patent dataset | closed | 2 | 2020-12-04T04:37:30 | 2020-12-06T17:21:00 | 2020-12-06T17:20:59 | mattbui | [] | * More info on the dataset: https://evasharma.github.io/bigpatent/
* There's another raw version of the dataset available from tfds. However, they're quite large so I don't have the resources to fully test all the configs for that version yet. We'll try to add it later. | true |
756,720,643 | https://api.github.com/repos/huggingface/datasets/issues/1086 | https://github.com/huggingface/datasets/pull/1086 | 1,086 | adding cdt dataset | closed | 2 | 2020-12-04T01:28:11 | 2020-12-04T15:04:02 | 2020-12-04T15:04:02 | abecadel | [] | - **Name:** *Cyberbullying Detection Task*
- **Description:** *The Cyberbullying Detection task was part of 2019 edition of PolEval competition. The goal is to predict if a given Twitter message contains a cyberbullying (harmful) content.*
- **Data:** *https://github.com/ptaszynski/cyberbullying-Polish*
- **Motivati... | true |
756,704,563 | https://api.github.com/repos/huggingface/datasets/issues/1085 | https://github.com/huggingface/datasets/pull/1085 | 1,085 | add mutual friends conversational dataset | closed | 1 | 2020-12-04T00:48:21 | 2020-12-16T15:58:31 | 2020-12-16T15:58:30 | VictorSanh | [] | Mutual friends dataset
WIP
TODO:
- scenario_kbs (bug with pyarrow conversion)
- download from codalab checksums bug | true |
756,688,727 | https://api.github.com/repos/huggingface/datasets/issues/1084 | https://github.com/huggingface/datasets/pull/1084 | 1,084 | adding cdsc dataset | closed | 0 | 2020-12-04T00:10:05 | 2020-12-04T10:41:26 | 2020-12-04T10:41:26 | abecadel | [] | - **Name**: *cdsc (domains: cdsc-e & cdsc-r)*
- **Description**: *Polish CDSCorpus consists of 10K Polish sentence pairs which are human-annotated for semantic relatedness and entailment. The dataset may be used for the evaluation of compositional distributional semantics models of Polish. The dataset was presented at... | true |
756,687,101 | https://api.github.com/repos/huggingface/datasets/issues/1083 | https://github.com/huggingface/datasets/pull/1083 | 1,083 | Add the multilingual Exams dataset | closed | 1 | 2020-12-04T00:06:04 | 2020-12-04T17:12:00 | 2020-12-04T17:12:00 | yjernite | [] | https://github.com/mhardalov/exams-qa
`multilingual` configs have all languages mixed together
`crosslingual` mixes the languages for test but separates them for train and dec, so I've made one config per language for train/dev data and one config with the joint test set | true |
756,676,218 | https://api.github.com/repos/huggingface/datasets/issues/1082 | https://github.com/huggingface/datasets/pull/1082 | 1,082 | Myanmar news dataset | closed | 1 | 2020-12-03T23:39:00 | 2020-12-04T10:13:38 | 2020-12-04T10:13:38 | mapmeld | [] | Add news topic classification dataset in Myanmar / Burmese languagess
This data was collected in 2017 by Aye Hninn Khine , and published on GitHub with a GPL license
https://github.com/ayehninnkhine/MyanmarNewsClassificationSystem
| true |
756,672,527 | https://api.github.com/repos/huggingface/datasets/issues/1081 | https://github.com/huggingface/datasets/pull/1081 | 1,081 | Add Knowledge-Enhanced Language Model Pre-training (KELM) | closed | 0 | 2020-12-03T23:30:09 | 2020-12-04T16:36:28 | 2020-12-04T16:36:28 | joeddav | [] | Adds the KELM dataset.
- Webpage/repo: https://github.com/google-research-datasets/KELM-corpus
- Paper: https://arxiv.org/pdf/2010.12688.pdf | true |
756,663,464 | https://api.github.com/repos/huggingface/datasets/issues/1080 | https://github.com/huggingface/datasets/pull/1080 | 1,080 | Add WikiANN NER dataset | closed | 1 | 2020-12-03T23:09:24 | 2020-12-06T17:18:55 | 2020-12-06T17:18:55 | lewtun | [] | This PR adds the full set of 176 languages from the balanced train/dev/test splits of WikiANN / PAN-X from: https://github.com/afshinrahimi/mmner
Until now, only 40 of these languages were available in `datasets` as part of the XTREME benchmark
Courtesy of the dataset author, we can now download this dataset from... | true |
756,652,427 | https://api.github.com/repos/huggingface/datasets/issues/1079 | https://github.com/huggingface/datasets/pull/1079 | 1,079 | nkjp-ner | closed | 0 | 2020-12-03T22:47:26 | 2020-12-04T09:42:06 | 2020-12-04T09:42:06 | abecadel | [] | - **Name:** *nkjp-ner*
- **Description:** *The NKJP-NER is based on a human-annotated part of NKJP. We extracted sentences with named entities of exactly one type. The task is to predict the type of the named entity.*
- **Data:** *https://klejbenchmark.com/tasks/*
- **Motivation:** *The KLEJ benchmark (Kompleksowa L... | true |
756,633,215 | https://api.github.com/repos/huggingface/datasets/issues/1078 | https://github.com/huggingface/datasets/pull/1078 | 1,078 | add AJGT dataset | closed | 0 | 2020-12-03T22:16:31 | 2020-12-04T09:55:15 | 2020-12-04T09:55:15 | zaidalyafeai | [] | Arabic Jordanian General Tweets. | true |
756,617,964 | https://api.github.com/repos/huggingface/datasets/issues/1077 | https://github.com/huggingface/datasets/pull/1077 | 1,077 | Added glucose dataset | closed | 0 | 2020-12-03T21:49:01 | 2020-12-04T09:55:53 | 2020-12-04T09:55:52 | TevenLeScao | [] | This PR adds the [Glucose](https://github.com/ElementalCognition/glucose) dataset. | true |
756,584,328 | https://api.github.com/repos/huggingface/datasets/issues/1076 | https://github.com/huggingface/datasets/pull/1076 | 1,076 | quac quac / coin coin | closed | 1 | 2020-12-03T20:55:29 | 2020-12-04T16:36:39 | 2020-12-04T09:15:20 | VictorSanh | [] | Add QUAC (Question Answering in Context)
I linearized most of the dictionnaries to lists.
Referenced to the authors' datasheet for the dataset card.
🦆🦆🦆
Coin coin | true |
756,501,235 | https://api.github.com/repos/huggingface/datasets/issues/1075 | https://github.com/huggingface/datasets/pull/1075 | 1,075 | adding cleaned verion of E2E NLG | closed | 0 | 2020-12-03T19:21:07 | 2020-12-03T19:43:56 | 2020-12-03T19:43:56 | yjernite | [] | Found at: https://github.com/tuetschek/e2e-cleaning | true |
756,483,172 | https://api.github.com/repos/huggingface/datasets/issues/1074 | https://github.com/huggingface/datasets/pull/1074 | 1,074 | Swedish MT STS-B | closed | 0 | 2020-12-03T19:06:25 | 2020-12-04T20:22:27 | 2020-12-03T20:44:28 | timpal0l | [] | Added a Swedish machine translated version of the well known STS-B Corpus | true |
756,468,034 | https://api.github.com/repos/huggingface/datasets/issues/1073 | https://github.com/huggingface/datasets/pull/1073 | 1,073 | Add DialogRE dataset | closed | 0 | 2020-12-03T18:56:40 | 2020-12-20T13:34:48 | 2020-12-04T13:41:51 | vineeths96 | [] | Adding the [DialogRE](https://github.com/nlpdata/dialogre) dataset Version 2.
- All tests passed successfully. | true |
756,454,511 | https://api.github.com/repos/huggingface/datasets/issues/1072 | https://github.com/huggingface/datasets/pull/1072 | 1,072 | actually uses the previously declared VERSION on the configs in the template | closed | 0 | 2020-12-03T18:44:27 | 2020-12-03T19:35:46 | 2020-12-03T19:35:46 | yjernite | [] | true | |
756,447,296 | https://api.github.com/repos/huggingface/datasets/issues/1071 | https://github.com/huggingface/datasets/pull/1071 | 1,071 | add xlrd to test package requirements | closed | 0 | 2020-12-03T18:32:47 | 2020-12-03T18:47:16 | 2020-12-03T18:47:16 | yjernite | [] | Adds `xlrd` package to the test requirements to handle scripts that use `pandas` to load excel files | true |
756,442,481 | https://api.github.com/repos/huggingface/datasets/issues/1070 | https://github.com/huggingface/datasets/pull/1070 | 1,070 | add conv_ai | closed | 2 | 2020-12-03T18:25:20 | 2020-12-04T07:58:35 | 2020-12-04T06:44:34 | patil-suraj | [] | Adding ConvAI dataset https://github.com/DeepPavlov/convai/tree/master/2017 | true |
756,425,737 | https://api.github.com/repos/huggingface/datasets/issues/1069 | https://github.com/huggingface/datasets/pull/1069 | 1,069 | Test | closed | 0 | 2020-12-03T18:01:45 | 2020-12-04T04:24:18 | 2020-12-04T04:24:11 | manandey | [] | true | |
756,417,337 | https://api.github.com/repos/huggingface/datasets/issues/1068 | https://github.com/huggingface/datasets/pull/1068 | 1,068 | Add Pubmed (citation + abstract) dataset (2020). | closed | 4 | 2020-12-03T17:54:10 | 2020-12-23T09:52:07 | 2020-12-23T09:52:07 | Narsil | [] | null | true |
756,414,212 | https://api.github.com/repos/huggingface/datasets/issues/1067 | https://github.com/huggingface/datasets/pull/1067 | 1,067 | add xquad-r dataset | closed | 0 | 2020-12-03T17:50:01 | 2020-12-03T17:53:21 | 2020-12-03T17:53:15 | manandey | [] | true | |
756,391,957 | https://api.github.com/repos/huggingface/datasets/issues/1066 | https://github.com/huggingface/datasets/pull/1066 | 1,066 | Add ChrEn | closed | 3 | 2020-12-03T17:17:48 | 2020-12-03T21:49:39 | 2020-12-03T21:49:39 | yjernite | [] | Adding the Cherokee English machine translation dataset of https://github.com/ZhangShiyue/ChrEn | true |
756,383,414 | https://api.github.com/repos/huggingface/datasets/issues/1065 | https://github.com/huggingface/datasets/pull/1065 | 1,065 | add xquad-r dataset | closed | 0 | 2020-12-03T17:06:23 | 2020-12-03T17:42:21 | 2020-12-03T17:42:03 | manandey | [] | true | |
756,382,186 | https://api.github.com/repos/huggingface/datasets/issues/1064 | https://github.com/huggingface/datasets/issues/1064 | 1,064 | Not support links with 302 redirect | closed | 2 | 2020-12-03T17:04:43 | 2021-01-14T02:51:25 | 2021-01-14T02:51:25 | chameleonTK | [
"bug",
"enhancement"
] | I have an issue adding this download link https://github.com/jitkapat/thailitcorpus/releases/download/v.2.0/tlc_v.2.0.tar.gz
it might be because it is not a direct link (it returns 302 and redirects to aws that returns 403 for head requests).
```
r.head("https://github.com/jitkapat/thailitcorpus/releases/downlo... | false |
756,376,374 | https://api.github.com/repos/huggingface/datasets/issues/1063 | https://github.com/huggingface/datasets/pull/1063 | 1,063 | Add the Ud treebank | closed | 1 | 2020-12-03T16:56:41 | 2020-12-04T16:11:54 | 2020-12-04T15:51:46 | jplu | [] | This PR adds the 183 datasets in 104 languages of the UD Treebank. | true |
756,373,187 | https://api.github.com/repos/huggingface/datasets/issues/1062 | https://github.com/huggingface/datasets/pull/1062 | 1,062 | Add KorNLU dataset | closed | 1 | 2020-12-03T16:52:39 | 2020-12-04T11:05:19 | 2020-12-04T11:05:19 | sumanthd17 | [] | Added Korean NLU datasets. The link to the dataset can be found [here](https://github.com/kakaobrain/KorNLUDatasets) and the paper can be found [here](https://arxiv.org/abs/2004.03289)
**Note**: The MNLI tsv file is broken, so this code currently excludes the file. Please suggest other alternative if any @lhoestq
... | true |
756,362,661 | https://api.github.com/repos/huggingface/datasets/issues/1061 | https://github.com/huggingface/datasets/pull/1061 | 1,061 | add labr dataset | closed | 0 | 2020-12-03T16:38:57 | 2020-12-03T18:25:44 | 2020-12-03T18:25:44 | zaidalyafeai | [] | Arabic Book Reviews dataset. | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.