url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
3.28B
node_id
stringlengths
18
32
number
int64
1
7.71k
title
stringlengths
1
290
user
dict
labels
listlengths
0
4
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
listlengths
0
30
created_at
timestamp[us, tz=UTC]date
2020-04-14 10:18:02
2025-07-30 11:34:53
updated_at
timestamp[us, tz=UTC]date
2020-04-27 16:04:17
2025-07-31 05:22:35
closed_at
timestamp[us, tz=UTC]date
2020-04-14 12:01:40
2025-07-30 14:22:21
author_association
stringclasses
4 values
type
null
active_lock_reason
null
sub_issues_summary
dict
body
stringlengths
0
228k
closed_by
dict
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
state_reason
stringclasses
4 values
draft
float64
0
1
pull_request
dict
created_at_dt
timestamp[us, tz=UTC]date
2020-04-14 10:18:02
2025-07-30 11:34:53
closed_at_dt
timestamp[us, tz=UTC]date
2020-04-14 12:01:40
2025-07-30 14:22:21
time_to_close
duration[us]
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/336
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/336/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/336/comments
https://api.github.com/repos/huggingface/datasets/issues/336/events
https://github.com/huggingface/datasets/issues/336
649,914,203
MDU6SXNzdWU2NDk5MTQyMDM=
336
[Dataset requests] New datasets for Open Question Answering
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[ { "color": "008672", "default": true, "description": "Extra attention is needed", "id": 1935892884, "name": "help wanted", "node_id": "MDU6TGFiZWwxOTM1ODkyODg0", "url": "https://api.github.com/repos/huggingface/datasets/labels/help%20wanted" }, { "color": "e99695", "default":...
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_...
null
[]
2020-07-02T13:03:03Z
2020-07-16T09:04:22Z
2020-07-16T09:04:22Z
MEMBER
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
We are still a few datasets missing for Open-Question Answering which is currently a field in strong development. Namely, it would be really nice to add: - WebQuestions (Berant et al., 2013) [done] - CuratedTrec (Baudis et al. 2015) [not open-source] - MS-MARCO (NGuyen et al. 2016) [done] - SearchQA (Dunn et al....
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/336/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/336/timeline
null
completed
null
null
2020-07-02T13:03:03Z
2020-07-16T09:04:22Z
13 days, 20:01:19
false
https://api.github.com/repos/huggingface/datasets/issues/335
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/335/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/335/comments
https://api.github.com/repos/huggingface/datasets/issues/335/events
https://github.com/huggingface/datasets/pull/335
649,765,179
MDExOlB1bGxSZXF1ZXN0NDQzMzgwMjI1
335
BioMRC Dataset presented in BioNLP 2020 ACL Workshop
{ "avatar_url": "https://avatars.githubusercontent.com/u/15162021?v=4", "events_url": "https://api.github.com/users/PetrosStav/events{/privacy}", "followers_url": "https://api.github.com/users/PetrosStav/followers", "following_url": "https://api.github.com/users/PetrosStav/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "I fixed the issues that you pointed out, re-run all the test and pushed the fixed code :-)", "created_at": "2020-07-02T20:01:34Z", "html_url": "https://github.com/huggingface/datasets/pull/335#issuecomment-653196722", "id": 653196722, "issue_ur...
2020-07-02T09:03:41Z
2020-07-15T08:02:07Z
2020-07-15T08:02:07Z
CONTRIBUTOR
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/335/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/335/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/335.diff", "html_url": "https://github.com/huggingface/datasets/pull/335", "merged_at": "2020-07-15T08:02:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/335.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/335...
2020-07-02T09:03:41Z
2020-07-15T08:02:07Z
12 days, 22:58:26
true
https://api.github.com/repos/huggingface/datasets/issues/334
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/334/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/334/comments
https://api.github.com/repos/huggingface/datasets/issues/334/events
https://github.com/huggingface/datasets/pull/334
649,661,791
MDExOlB1bGxSZXF1ZXN0NDQzMjk1NjQ0
334
Add dataset.shard() method
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Great, done!", "created_at": "2020-07-02T15:05:12Z", "html_url": "https://github.com/huggingface/datasets/pull/334#issuecomment-653061322", "id": 653061322, "issue_url": "https://api.github.com/repos/huggingface/datasets/issues/334", "node_...
2020-07-02T06:05:19Z
2020-07-06T12:35:36Z
2020-07-06T12:35:36Z
CONTRIBUTOR
null
null
null
Fixes https://github.com/huggingface/nlp/issues/312
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/334/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/334/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/334.diff", "html_url": "https://github.com/huggingface/datasets/pull/334", "merged_at": "2020-07-06T12:35:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/334.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/334...
2020-07-02T06:05:19Z
2020-07-06T12:35:36Z
4 days, 6:30:17
true
https://api.github.com/repos/huggingface/datasets/issues/333
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/333/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/333/comments
https://api.github.com/repos/huggingface/datasets/issues/333/events
https://github.com/huggingface/datasets/pull/333
649,236,516
MDExOlB1bGxSZXF1ZXN0NDQyOTE1NDQ0
333
fix variable name typo
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://a...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Good catch :)\r\nI think there is another occurence that needs to be fixed in the second gist (line 4924 of the notebook file):\r\n```python\r\nbleu = nlp.load_metric(...)\r\n```", "created_at": "2020-07-02T08:27:48Z", "html_url": "https://github.com/huggin...
2020-07-01T19:13:50Z
2020-07-24T15:43:31Z
2020-07-24T08:32:16Z
CONTRIBUTOR
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/333/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/333/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/333.diff", "html_url": "https://github.com/huggingface/datasets/pull/333", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/333.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/333" }
2020-07-01T19:13:50Z
2020-07-24T08:32:16Z
22 days, 13:18:26
true
https://api.github.com/repos/huggingface/datasets/issues/332
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/332/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/332/comments
https://api.github.com/repos/huggingface/datasets/issues/332/events
https://github.com/huggingface/datasets/pull/332
649,140,135
MDExOlB1bGxSZXF1ZXN0NDQyODMwMzMz
332
Add wiki_dpr
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "The two configurations don't have the same sizes, I may change that so that they both have 21015300 examples for convenience, even though it's supposed to have 21015324 examples in total.\r\n\r\nOne configuration only has 21015300 examples because it seems that the...
2020-07-01T17:12:00Z
2020-07-06T12:21:17Z
2020-07-06T12:21:16Z
MEMBER
null
null
null
Presented in the [Dense Passage Retrieval paper](https://arxiv.org/pdf/2004.04906.pdf), this dataset consists in 21M passages from the english wikipedia along with their 768-dim embeddings computed using DPR's context encoder. Note on the implementation: - There are two configs: with and without the embeddings (73G...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/332/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/332/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/332.diff", "html_url": "https://github.com/huggingface/datasets/pull/332", "merged_at": "2020-07-06T12:21:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/332.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/332...
2020-07-01T17:12:00Z
2020-07-06T12:21:16Z
4 days, 19:09:16
true
https://api.github.com/repos/huggingface/datasets/issues/331
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/331/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/331/comments
https://api.github.com/repos/huggingface/datasets/issues/331/events
https://github.com/huggingface/datasets/issues/331
648,533,199
MDU6SXNzdWU2NDg1MzMxOTk=
331
Loading CNN/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "I couldn't reproduce on my side.\r\nIt looks like you were not able to generate all the examples, and you have the problem for each split train-test-validation.\r\nCould you try to enable logging, try again and send the logs ?\r\n```python\r\nimport logging\r\nlogg...
2020-06-30T22:21:33Z
2020-07-09T13:03:40Z
2020-07-09T13:03:40Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
``` >>> import nlp >>> nlp.load_dataset('cnn_dailymail', '3.0.0') Downloading and preparing dataset cnn_dailymail/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to /u/jm8wx/.cache/huggingface/datasets/cnn_dailymail/3.0.0/3.0.0... Traceback (most recent call last): File "<stdin>", line 1, in...
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/331/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/331/timeline
null
completed
null
null
2020-06-30T22:21:33Z
2020-07-09T13:03:40Z
8 days, 14:42:07
false
https://api.github.com/repos/huggingface/datasets/issues/330
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/330/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/330/comments
https://api.github.com/repos/huggingface/datasets/issues/330/events
https://github.com/huggingface/datasets/pull/330
648,525,720
MDExOlB1bGxSZXF1ZXN0NDQyMzIxMjEw
330
Doc red
{ "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "events_url": "https://api.github.com/users/ghomasHudson/events{/privacy}", "followers_url": "https://api.github.com/users/ghomasHudson/followers", "following_url": "https://api.github.com/users/ghomasHudson/following{/other_user}", "gist...
[]
closed
false
null
[]
null
[]
2020-06-30T22:05:31Z
2020-07-06T12:10:39Z
2020-07-05T12:27:29Z
CONTRIBUTOR
null
null
null
Adding [DocRED](https://github.com/thunlp/DocRED) - a relation extraction dataset which tests document-level RE. A few implementation notes: - There are 2 separate versions of the training set - *annotated* and *distant*. Instead of `nlp.Split.Train` I've used the splits `"train_annotated"` and `"train_distant"` to ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/330/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/330/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/330.diff", "html_url": "https://github.com/huggingface/datasets/pull/330", "merged_at": "2020-07-05T12:27:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/330.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/330...
2020-06-30T22:05:31Z
2020-07-05T12:27:29Z
4 days, 14:21:58
true
https://api.github.com/repos/huggingface/datasets/issues/329
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/329/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/329/comments
https://api.github.com/repos/huggingface/datasets/issues/329/events
https://github.com/huggingface/datasets/issues/329
648,446,979
MDU6SXNzdWU2NDg0NDY5Nzk=
329
[Bug] FileLock dependency incompatible with filesystem
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Hi, can you give details on your environment/os/packages versions/etc?", "created_at": "2020-06-30T19:47:25Z", "html_url": "https://github.com/huggingface/datasets/issues/329#issuecomment-652006157", "id": 652006157, "issue_url": "https://api.github...
2020-06-30T19:45:31Z
2024-12-26T15:13:39Z
2020-06-30T21:33:06Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I'm downloading a dataset successfully with `load_dataset("wikitext", "wikitext-2-raw-v1")` But when I attempt to cache it on an external volume, it hangs indefinitely: `load_dataset("wikitext", "wikitext-2-raw-v1", cache_dir="/fsx") # /fsx is an external volume mount` The filesystem when hanging looks like thi...
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/329/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/329/timeline
null
completed
null
null
2020-06-30T19:45:31Z
2020-06-30T21:33:06Z
1:47:35
false
https://api.github.com/repos/huggingface/datasets/issues/328
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/328/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/328/comments
https://api.github.com/repos/huggingface/datasets/issues/328/events
https://github.com/huggingface/datasets/issues/328
648,326,841
MDU6SXNzdWU2NDgzMjY4NDE=
328
Fork dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "events_url": "https://api.github.com/users/timothyjlaurent/events{/privacy}", "followers_url": "https://api.github.com/users/timothyjlaurent/followers", "following_url": "https://api.github.com/users/timothyjlaurent/following{/other_user}",...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "To be able to generate the Arrow dataset you need to either use our csv or json utilities `load_dataset(\"json\", data_files=my_json_files)` OR write your own custom dataset script (you can find some inspiration from the [squad](https://github.com/huggingface/nlp/b...
2020-06-30T16:42:53Z
2020-07-06T21:43:59Z
2020-07-06T21:43:59Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
We have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. We're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers. Our preprocessing flow parses raw text and...
{ "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "events_url": "https://api.github.com/users/timothyjlaurent/events{/privacy}", "followers_url": "https://api.github.com/users/timothyjlaurent/followers", "following_url": "https://api.github.com/users/timothyjlaurent/following{/other_user}",...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/328/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/328/timeline
null
completed
null
null
2020-06-30T16:42:53Z
2020-07-06T21:43:59Z
6 days, 5:01:06
false
https://api.github.com/repos/huggingface/datasets/issues/327
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/327/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/327/comments
https://api.github.com/repos/huggingface/datasets/issues/327/events
https://github.com/huggingface/datasets/pull/327
648,312,858
MDExOlB1bGxSZXF1ZXN0NDQyMTQyOTQw
327
set seed for suffling tests
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[]
2020-06-30T16:21:34Z
2020-07-02T08:34:05Z
2020-07-02T08:34:04Z
MEMBER
null
null
null
Some tests were randomly failing because of a missing seed in a test for `train_test_split(shuffle=True)`
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/327/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/327/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/327.diff", "html_url": "https://github.com/huggingface/datasets/pull/327", "merged_at": "2020-07-02T08:34:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/327.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/327...
2020-06-30T16:21:34Z
2020-07-02T08:34:04Z
1 day, 16:12:30
true
https://api.github.com/repos/huggingface/datasets/issues/326
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/326/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/326/comments
https://api.github.com/repos/huggingface/datasets/issues/326/events
https://github.com/huggingface/datasets/issues/326
648,126,103
MDU6SXNzdWU2NDgxMjYxMDM=
326
Large dataset in Squad2-format
{ "avatar_url": "https://avatars.githubusercontent.com/u/47894090?v=4", "events_url": "https://api.github.com/users/flozi00/events{/privacy}", "followers_url": "https://api.github.com/users/flozi00/followers", "following_url": "https://api.github.com/users/flozi00/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "I'm pretty sure you can get some inspiration from the squad_v2 script. It looks like the dataset is quite big so it will take some time for the users to generate it, but it should be reasonable.\r\n\r\nAlso you are saying that you are still making the dataset grow ...
2020-06-30T12:18:59Z
2020-07-09T09:01:50Z
2020-07-09T09:01:50Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
At the moment we are building an large question answering dataset and think about sharing it with the huggingface community. Caused the computing power we splitted it into multiple tiles, but they are all in the same format. Right now the most important facts about are this: - Contexts: 1.047.671 - questions: 1.677...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/326/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/326/timeline
null
completed
null
null
2020-06-30T12:18:59Z
2020-07-09T09:01:50Z
8 days, 20:42:51
false
https://api.github.com/repos/huggingface/datasets/issues/324
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/324/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/324/comments
https://api.github.com/repos/huggingface/datasets/issues/324/events
https://github.com/huggingface/datasets/issues/324
647,525,725
MDU6SXNzdWU2NDc1MjU3MjU=
324
Error when calculating glue score
{ "avatar_url": "https://avatars.githubusercontent.com/u/47185867?v=4", "events_url": "https://api.github.com/users/D-i-l-r-u-k-s-h-i/events{/privacy}", "followers_url": "https://api.github.com/users/D-i-l-r-u-k-s-h-i/followers", "following_url": "https://api.github.com/users/D-i-l-r-u-k-s-h-i/following{/other_...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "The glue metric for cola is a metric for classification. It expects label ids as integers as inputs.", "created_at": "2020-06-30T16:13:06Z", "html_url": "https://github.com/huggingface/datasets/issues/324#issuecomment-651896268", "id": 651896268, "i...
2020-06-29T16:53:48Z
2020-07-09T09:13:34Z
2020-07-09T09:13:34Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I was trying glue score along with other metrics here. But glue gives me this error; ``` import nlp glue_metric = nlp.load_metric('glue',name="cola") glue_score = glue_metric.compute(predictions, references) ``` ``` --------------------------------------------------------------------------- --------------...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/324/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/324/timeline
null
completed
null
null
2020-06-29T16:53:48Z
2020-07-09T09:13:34Z
9 days, 16:19:46
false
https://api.github.com/repos/huggingface/datasets/issues/325
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/325/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/325/comments
https://api.github.com/repos/huggingface/datasets/issues/325/events
https://github.com/huggingface/datasets/pull/325
647,601,592
MDExOlB1bGxSZXF1ZXN0NDQxNTk3NTgw
325
Add SQuADShifts dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/8953195?v=4", "events_url": "https://api.github.com/users/millerjohnp/events{/privacy}", "followers_url": "https://api.github.com/users/millerjohnp/followers", "following_url": "https://api.github.com/users/millerjohnp/following{/other_user}", "gists_ur...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Very cool to have this dataset, thank you for adding it :)", "created_at": "2020-06-30T17:07:25Z", "html_url": "https://github.com/huggingface/datasets/pull/325#issuecomment-651925598", "id": 651925598, "issue_url": "https://api.github.com/repos/hug...
2020-06-29T19:11:16Z
2020-06-30T17:07:31Z
2020-06-30T17:07:31Z
CONTRIBUTOR
null
null
null
This PR adds the four new variants of the SQuAD dataset used in [The Effect of Natural Distribution Shift on Question Answering Models](https://arxiv.org/abs/2004.14444) to facilitate evaluating model robustness to distribution shift.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/325/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/325/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/325.diff", "html_url": "https://github.com/huggingface/datasets/pull/325", "merged_at": "2020-06-30T17:07:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/325.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/325...
2020-06-29T19:11:16Z
2020-06-30T17:07:31Z
21:56:15
true
https://api.github.com/repos/huggingface/datasets/issues/323
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/323/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/323/comments
https://api.github.com/repos/huggingface/datasets/issues/323/events
https://github.com/huggingface/datasets/pull/323
647,521,308
MDExOlB1bGxSZXF1ZXN0NDQxNTMxOTY3
323
Add package path to sys when downloading package as github archive
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Sorry for the long diff, everything after the imports comes from `black` for code quality :/ ", "created_at": "2020-06-29T16:47:07Z", "html_url": "https://github.com/huggingface/datasets/pull/323#issuecomment-651237304", "id": 651237304, "issue_url"...
2020-06-29T16:46:01Z
2020-07-30T14:00:23Z
2020-07-30T14:00:23Z
MEMBER
null
null
null
This fixes the `coval.py` metric so that imports within the downloaded module work correctly. We can use a similar trick to add the BLEURT metric (@ankparikh) @thomwolf not sure how you feel about adding to the `PYTHONPATH` from the script. This is the only way I could make it work with my understanding of `importli...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/323/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/323/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/323.diff", "html_url": "https://github.com/huggingface/datasets/pull/323", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/323.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/323" }
2020-06-29T16:46:01Z
2020-07-30T14:00:23Z
30 days, 21:14:22
true
https://api.github.com/repos/huggingface/datasets/issues/321
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/321/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/321/comments
https://api.github.com/repos/huggingface/datasets/issues/321/events
https://github.com/huggingface/datasets/issues/321
647,271,526
MDU6SXNzdWU2NDcyNzE1MjY=
321
ERROR:root:mwparserfromhell
{ "avatar_url": "https://avatars.githubusercontent.com/u/26505641?v=4", "events_url": "https://api.github.com/users/Shiro-LK/events{/privacy}", "followers_url": "https://api.github.com/users/Shiro-LK/followers", "following_url": "https://api.github.com/users/Shiro-LK/following{/other_user}", "gists_url": "htt...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "It looks like it comes from `mwparserfromhell`.\r\n\r\nWould it be possible to get the bad `section` that causes this issue ? The `section` string is from `datasets/wikipedia.py:L548` ? You could just add a `try` statement and print the section if the line `section...
2020-06-29T11:10:43Z
2022-02-14T15:21:46Z
2022-02-14T15:21:46Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi, I am trying to download some wikipedia data but I got this error for spanish "es" (but there are maybe some others languages which have the same error I haven't tried all of them ). `ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token sta...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/321/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/321/timeline
null
completed
null
null
2020-06-29T11:10:43Z
2022-02-14T15:21:46Z
595 days, 4:11:03
false
https://api.github.com/repos/huggingface/datasets/issues/322
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/322/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/322/comments
https://api.github.com/repos/huggingface/datasets/issues/322/events
https://github.com/huggingface/datasets/pull/322
647,483,850
MDExOlB1bGxSZXF1ZXN0NDQxNTAyMjc2
322
output nested dict in get_nearest_examples
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[]
2020-06-29T15:47:47Z
2020-07-02T08:33:33Z
2020-07-02T08:33:32Z
MEMBER
null
null
null
As we are using a columnar format like arrow as the backend for datasets, we expect to have a dictionary of columns when we slice a dataset like in this example: ```python my_examples = dataset[0:10] print(type(my_examples)) # >>> dict print(my_examples["my_column"][0] # >>> this is the first element of the colum...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/322/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/322/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/322.diff", "html_url": "https://github.com/huggingface/datasets/pull/322", "merged_at": "2020-07-02T08:33:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/322.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/322...
2020-06-29T15:47:47Z
2020-07-02T08:33:32Z
2 days, 16:45:45
true
https://api.github.com/repos/huggingface/datasets/issues/320
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/320/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/320/comments
https://api.github.com/repos/huggingface/datasets/issues/320/events
https://github.com/huggingface/datasets/issues/320
647,188,167
MDU6SXNzdWU2NDcxODgxNjc=
320
Blog Authorship Corpus, Non Matching Splits Sizes Error, nlp viewer
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[ { "color": "94203D", "default": false, "description": "", "id": 2107841032, "name": "nlp-viewer", "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer" } ]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "I wonder if this means downloading failed? That corpus has a really slow server.", "created_at": "2020-06-29T14:21:45Z", "html_url": "https://github.com/huggingface/datasets/issues/320#issuecomment-651154382", "id": 651154382, "issue_url": "htt...
2020-06-29T07:36:35Z
2020-06-29T14:44:42Z
2020-06-29T14:44:42Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Selecting `blog_authorship_corpus` in the nlp viewer throws the following error: ``` NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dat...
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/320/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/320/timeline
null
completed
null
null
2020-06-29T07:36:35Z
2020-06-29T14:44:42Z
7:08:07
false
https://api.github.com/repos/huggingface/datasets/issues/319
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/319/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/319/comments
https://api.github.com/repos/huggingface/datasets/issues/319/events
https://github.com/huggingface/datasets/issues/319
646,792,487
MDU6SXNzdWU2NDY3OTI0ODc=
319
Nested sequences with dicts
{ "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "events_url": "https://api.github.com/users/ghomasHudson/events{/privacy}", "followers_url": "https://api.github.com/users/ghomasHudson/followers", "following_url": "https://api.github.com/users/ghomasHudson/following{/other_user}", "gist...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Oh yes, this is a backward compatibility feature with tensorflow_dataset in which a `Sequence` or `dict` is converted in a `dict` of `lists`, unfortunately it is not very intuitive, see here: https://github.com/huggingface/nlp/blob/master/src/nlp/features.py#L409\r...
2020-06-27T23:45:17Z
2020-07-03T10:22:00Z
2020-07-03T10:22:00Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Am pretty much finished [adding a dataset](https://github.com/ghomasHudson/nlp/blob/DocRED/datasets/docred/docred.py) for [DocRED](https://github.com/thunlp/DocRED), but am getting an error when trying to add a nested `nlp.features.sequence(nlp.features.sequence({key:value,...}))`. The original data is in this form...
{ "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "events_url": "https://api.github.com/users/ghomasHudson/events{/privacy}", "followers_url": "https://api.github.com/users/ghomasHudson/followers", "following_url": "https://api.github.com/users/ghomasHudson/following{/other_user}", "gist...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/319/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/319/timeline
null
completed
null
null
2020-06-27T23:45:17Z
2020-07-03T10:22:00Z
5 days, 10:36:43
false
https://api.github.com/repos/huggingface/datasets/issues/318
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/318/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/318/comments
https://api.github.com/repos/huggingface/datasets/issues/318/events
https://github.com/huggingface/datasets/pull/318
646,682,840
MDExOlB1bGxSZXF1ZXN0NDQwOTExOTYy
318
Multitask
{ "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "events_url": "https://api.github.com/users/ghomasHudson/events{/privacy}", "followers_url": "https://api.github.com/users/ghomasHudson/followers", "following_url": "https://api.github.com/users/ghomasHudson/following{/other_user}", "gist...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "It's definitely going in the right direction ! Thanks for giving it a try\r\n\r\nI really like the API.\r\nIMO it's fine right now if we don't have all the dataset transforms (map, filter, etc.) as it can be done before building the multitask dataset, but it will b...
2020-06-27T13:27:29Z
2022-07-06T15:19:57Z
2022-07-06T15:19:57Z
CONTRIBUTOR
null
null
null
Following our discussion in #217, I've implemented a first working version of `MultiDataset`. There's a function `build_multitask()` which takes either individual `nlp.Dataset`s or `dicts` of splits and constructs `MultiDataset`(s). I've added a notebook with example usage. I've implemented many of the `nlp.Datas...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/318/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/318/timeline
null
null
1
{ "diff_url": "https://github.com/huggingface/datasets/pull/318.diff", "html_url": "https://github.com/huggingface/datasets/pull/318", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/318.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/318" }
2020-06-27T13:27:29Z
2022-07-06T15:19:57Z
739 days, 1:52:28
true
https://api.github.com/repos/huggingface/datasets/issues/316
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/316/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/316/comments
https://api.github.com/repos/huggingface/datasets/issues/316/events
https://github.com/huggingface/datasets/pull/316
646,366,450
MDExOlB1bGxSZXF1ZXN0NDQwNjY5NzY5
316
add AG News dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Thanks @jxmorris12 for adding this adding. \r\nCan you please add a small description of the PR?", "created_at": "2020-06-26T17:01:33Z", "html_url": "https://github.com/huggingface/datasets/pull/316#issuecomment-650288419", "id": 650288419, "is...
2020-06-26T16:11:58Z
2020-06-30T09:58:08Z
2020-06-30T08:31:55Z
CONTRIBUTOR
null
null
null
adds support for the AG-News topic classification dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/316/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/316/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/316.diff", "html_url": "https://github.com/huggingface/datasets/pull/316", "merged_at": "2020-06-30T08:31:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/316.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/316...
2020-06-26T16:11:58Z
2020-06-30T08:31:55Z
3 days, 16:19:57
true
https://api.github.com/repos/huggingface/datasets/issues/317
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/317/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/317/comments
https://api.github.com/repos/huggingface/datasets/issues/317/events
https://github.com/huggingface/datasets/issues/317
646,555,384
MDU6SXNzdWU2NDY1NTUzODQ=
317
Adding a dataset with multiple subtasks
{ "avatar_url": "https://avatars.githubusercontent.com/u/294483?v=4", "events_url": "https://api.github.com/users/erickrf/events{/privacy}", "followers_url": "https://api.github.com/users/erickrf/followers", "following_url": "https://api.github.com/users/erickrf/following{/other_user}", "gists_url": "https://...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "For one dataset you can have different configurations that each have their own `nlp.Features`.\r\nWe imagine having one configuration per subtask for example.\r\nThey are loaded with `nlp.load_dataset(\"my_dataset\", \"my_config\")`.\r\n\r\nFor example the `glue` d...
2020-06-26T23:14:19Z
2020-10-27T15:36:52Z
2020-10-27T15:36:52Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I intent to add the datasets of the MT Quality Estimation shared tasks to `nlp`. However, they have different subtasks -- such as word-level, sentence-level and document-level quality estimation, each of which having different language pairs, and some of the data reused in different subtasks. For example, in [QE 201...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/317/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/317/timeline
null
completed
null
null
2020-06-26T23:14:19Z
2020-10-27T15:36:52Z
122 days, 16:22:33
false
https://api.github.com/repos/huggingface/datasets/issues/314
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/314/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/314/comments
https://api.github.com/repos/huggingface/datasets/issues/314/events
https://github.com/huggingface/datasets/pull/314
645,461,174
MDExOlB1bGxSZXF1ZXN0NDM5OTM4MTMw
314
Fixed singlular very minor spelling error
{ "avatar_url": "https://avatars.githubusercontent.com/u/40696362?v=4", "events_url": "https://api.github.com/users/SchizoidBat/events{/privacy}", "followers_url": "https://api.github.com/users/SchizoidBat/followers", "following_url": "https://api.github.com/users/SchizoidBat/following{/other_user}", "gists_u...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Thank you BatJeti! The storm-joker, aka the typo, finally got caught!", "created_at": "2020-06-25T12:01:31Z", "html_url": "https://github.com/huggingface/datasets/pull/314#issuecomment-649498218", "id": 649498218, "issue_url": "https://api.gith...
2020-06-25T10:45:59Z
2020-06-26T08:46:41Z
2020-06-25T12:43:59Z
CONTRIBUTOR
null
null
null
An instance of "independantly" was changed to "independently". That's all.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/314/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/314/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/314.diff", "html_url": "https://github.com/huggingface/datasets/pull/314", "merged_at": "2020-06-25T12:43:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/314.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/314...
2020-06-25T10:45:59Z
2020-06-25T12:43:59Z
1:58:00
true
https://api.github.com/repos/huggingface/datasets/issues/315
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/315/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/315/comments
https://api.github.com/repos/huggingface/datasets/issues/315/events
https://github.com/huggingface/datasets/issues/315
645,888,943
MDU6SXNzdWU2NDU4ODg5NDM=
315
[Question] Best way to batch a large dataset?
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
[ { "color": "c5def5", "default": false, "description": "Generic discussion on the library", "id": 2067400324, "name": "generic discussion", "node_id": "MDU6TGFiZWwyMDY3NDAwMzI0", "url": "https://api.github.com/repos/huggingface/datasets/labels/generic%20discussion" } ]
open
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Update: I think I've found a solution.\r\n\r\n```python\r\noutput_types = {\"input_ids\": tf.int64, \"token_type_ids\": tf.int64, \"attention_mask\": tf.int64}\r\ndef train_dataset_gen():\r\n for i in range(len(train_dataset)):\r\n yield train_datase...
2020-06-25T22:30:20Z
2020-10-27T15:38:17Z
null
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I'm training on large datasets such as Wikipedia and BookCorpus. Following the instructions in [the tutorial notebook](https://colab.research.google.com/github/huggingface/nlp/blob/master/notebooks/Overview.ipynb), I see the following recommended for TensorFlow: ```python train_tf_dataset = train_tf_dataset.filter(...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 1, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/315/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/315/timeline
null
null
null
null
2020-06-25T22:30:20Z
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/312
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/312/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/312/comments
https://api.github.com/repos/huggingface/datasets/issues/312/events
https://github.com/huggingface/datasets/issues/312
645,025,561
MDU6SXNzdWU2NDUwMjU1NjE=
312
[Feature request] Add `shard()` method to dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Hi Jared,\r\nInteresting, thanks for raising this question. You can also do that after loading with `dataset.select()` or `dataset.filter()` which let you keep only a specific subset of rows in a dataset.\r\nWhat is your use-case for sharding?", "created_at": "...
2020-06-24T22:48:33Z
2020-07-06T12:35:36Z
2020-07-06T12:35:36Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Currently, to shard a dataset into 10 pieces on different ranks, you can run ```python rank = 3 # for example size = 10 dataset = nlp.load_dataset('wikitext', 'wikitext-2-raw-v1', split=f"train[{rank*10}%:{(rank+1)*10}%]") ``` However, this breaks down if you have a number of ranks that doesn't divide cleanly...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/312/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/312/timeline
null
completed
null
null
2020-06-24T22:48:33Z
2020-07-06T12:35:36Z
11 days, 13:47:03
false
https://api.github.com/repos/huggingface/datasets/issues/313
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/313/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/313/comments
https://api.github.com/repos/huggingface/datasets/issues/313/events
https://github.com/huggingface/datasets/pull/313
645,390,088
MDExOlB1bGxSZXF1ZXN0NDM5ODc4MDg5
313
Add MWSC
{ "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "events_url": "https://api.github.com/users/ghomasHudson/events{/privacy}", "followers_url": "https://api.github.com/users/ghomasHudson/followers", "following_url": "https://api.github.com/users/ghomasHudson/following{/other_user}", "gist...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/followin...
null
[ { "author_association": "CONTRIBUTOR", "body": "Looks good to me", "created_at": "2020-06-27T10:28:16Z", "html_url": "https://github.com/huggingface/datasets/pull/313#issuecomment-650538771", "id": 650538771, "issue_url": "https://api.github.com/repos/huggingface/datasets/issues/313", "n...
2020-06-25T09:22:02Z
2020-06-30T08:28:11Z
2020-06-30T08:28:11Z
CONTRIBUTOR
null
null
null
Adding the [Modified Winograd Schema Challenge](https://github.com/salesforce/decaNLP/blob/master/local_data/schema.txt) dataset which formed part of the [decaNLP](http://decanlp.com/) benchmark. Not sure how much use people would find for it it outside of the benchmark, but it is general purpose. Code is heavily bo...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 3, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/313/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/313/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/313.diff", "html_url": "https://github.com/huggingface/datasets/pull/313", "merged_at": "2020-06-30T08:28:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/313.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/313...
2020-06-25T09:22:02Z
2020-06-30T08:28:11Z
4 days, 23:06:09
true
https://api.github.com/repos/huggingface/datasets/issues/311
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/311/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/311/comments
https://api.github.com/repos/huggingface/datasets/issues/311/events
https://github.com/huggingface/datasets/pull/311
645,013,131
MDExOlB1bGxSZXF1ZXN0NDM5NTQ3OTg0
311
Add qa_zre
{ "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "events_url": "https://api.github.com/users/ghomasHudson/events{/privacy}", "followers_url": "https://api.github.com/users/ghomasHudson/followers", "following_url": "https://api.github.com/users/ghomasHudson/following{/other_user}", "gist...
[]
closed
false
null
[]
null
[]
2020-06-24T22:17:22Z
2020-06-29T16:37:38Z
2020-06-29T16:37:38Z
CONTRIBUTOR
null
null
null
Adding the QA-ZRE dataset from ["Zero-Shot Relation Extraction via Reading Comprehension"](http://nlp.cs.washington.edu/zeroshot/). A common processing step seems to be replacing the `XXX` placeholder with the `subject`. I've left this out as it's something you could easily do with `map`.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/311/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/311/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/311.diff", "html_url": "https://github.com/huggingface/datasets/pull/311", "merged_at": "2020-06-29T16:37:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/311.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/311...
2020-06-24T22:17:22Z
2020-06-29T16:37:38Z
4 days, 18:20:16
true
https://api.github.com/repos/huggingface/datasets/issues/310
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/310/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/310/comments
https://api.github.com/repos/huggingface/datasets/issues/310/events
https://github.com/huggingface/datasets/pull/310
644,806,720
MDExOlB1bGxSZXF1ZXN0NDM5MzY1MDg5
310
add wikisql
{ "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "events_url": "https://api.github.com/users/ghomasHudson/events{/privacy}", "followers_url": "https://api.github.com/users/ghomasHudson/followers", "following_url": "https://api.github.com/users/ghomasHudson/following{/other_user}", "gist...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "That's great work @ghomasHudson !", "created_at": "2020-06-25T09:17:38Z", "html_url": "https://github.com/huggingface/datasets/pull/310#issuecomment-649413914", "id": 649413914, "issue_url": "https://api.github.com/repos/huggingface/datasets/is...
2020-06-24T18:00:35Z
2020-06-25T12:32:25Z
2020-06-25T12:32:25Z
CONTRIBUTOR
null
null
null
Adding the [WikiSQL](https://github.com/salesforce/WikiSQL) dataset. Interesting things to note: - Have copied the function (`_convert_to_human_readable`) which converts the SQL query to a human-readable (string) format as this is what most people will want when actually using this dataset for NLP applications. - ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/310/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/310/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/310.diff", "html_url": "https://github.com/huggingface/datasets/pull/310", "merged_at": "2020-06-25T12:32:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/310.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/310...
2020-06-24T18:00:35Z
2020-06-25T12:32:25Z
18:31:50
true
https://api.github.com/repos/huggingface/datasets/issues/308
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/308/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/308/comments
https://api.github.com/repos/huggingface/datasets/issues/308/events
https://github.com/huggingface/datasets/pull/308
644,195,251
MDExOlB1bGxSZXF1ZXN0NDM4ODYyMzYy
308
Specify utf-8 encoding for MRPC files
{ "avatar_url": "https://avatars.githubusercontent.com/u/15801338?v=4", "events_url": "https://api.github.com/users/patpizio/events{/privacy}", "followers_url": "https://api.github.com/users/patpizio/followers", "following_url": "https://api.github.com/users/patpizio/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[]
2020-06-23T22:44:36Z
2020-06-25T12:52:21Z
2020-06-25T12:16:10Z
CONTRIBUTOR
null
null
null
Fixes #307, again probably a Windows-related issue.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/308/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/308/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/308.diff", "html_url": "https://github.com/huggingface/datasets/pull/308", "merged_at": "2020-06-25T12:16:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/308.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/308...
2020-06-23T22:44:36Z
2020-06-25T12:16:10Z
1 day, 13:31:34
true
https://api.github.com/repos/huggingface/datasets/issues/307
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/307/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/307/comments
https://api.github.com/repos/huggingface/datasets/issues/307/events
https://github.com/huggingface/datasets/issues/307
644,187,262
MDU6SXNzdWU2NDQxODcyNjI=
307
Specify encoding for MRPC
{ "avatar_url": "https://avatars.githubusercontent.com/u/15801338?v=4", "events_url": "https://api.github.com/users/patpizio/events{/privacy}", "followers_url": "https://api.github.com/users/patpizio/followers", "following_url": "https://api.github.com/users/patpizio/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[]
2020-06-23T22:24:49Z
2020-06-25T12:16:09Z
2020-06-25T12:16:09Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Same as #242, but with MRPC: on Windows, I get a `UnicodeDecodeError` when I try to download the dataset: ```python dataset = nlp.load_dataset('glue', 'mrpc') ``` ```python Downloading and preparing dataset glue/mrpc (download: Unknown size, generated: Unknown size, total: Unknown size) to C:\Users\Python\.cache...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/307/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/307/timeline
null
completed
null
null
2020-06-23T22:24:49Z
2020-06-25T12:16:09Z
1 day, 13:51:20
false
https://api.github.com/repos/huggingface/datasets/issues/309
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/309/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/309/comments
https://api.github.com/repos/huggingface/datasets/issues/309/events
https://github.com/huggingface/datasets/pull/309
644,783,822
MDExOlB1bGxSZXF1ZXN0NDM5MzQ1NzYz
309
Add narrative qa
{ "avatar_url": "https://avatars.githubusercontent.com/u/8019486?v=4", "events_url": "https://api.github.com/users/Varal7/events{/privacy}", "followers_url": "https://api.github.com/users/Varal7/followers", "following_url": "https://api.github.com/users/Varal7/following{/other_user}", "gists_url": "https://ap...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Does it make sense to download the full stories? I remember attempting to implement this dataset a while ago and ended up with something like:\r\n```python\r\n def _split_generators(self, dl_manager):\r\n \"\"\"Returns SplitGenerators.\"\"\"\r\n\r\n ...
2020-06-24T17:26:18Z
2020-09-03T09:02:10Z
2020-09-03T09:02:09Z
NONE
null
null
null
Test cases for dummy data don't pass Only contains data for summaries (not whole story)
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/309/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/309/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/309.diff", "html_url": "https://github.com/huggingface/datasets/pull/309", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/309.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/309" }
2020-06-24T17:26:18Z
2020-09-03T09:02:09Z
70 days, 15:35:51
true
https://api.github.com/repos/huggingface/datasets/issues/306
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/306/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/306/comments
https://api.github.com/repos/huggingface/datasets/issues/306/events
https://github.com/huggingface/datasets/pull/306
644,176,078
MDExOlB1bGxSZXF1ZXN0NDM4ODQ2MTI3
306
add pg19 dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/108653?v=4", "events_url": "https://api.github.com/users/lucidrains/events{/privacy}", "followers_url": "https://api.github.com/users/lucidrains/followers", "following_url": "https://api.github.com/users/lucidrains/following{/other_user}", "gists_url": ...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "@lucidrains - Thanks a lot for making the PR - PG19 is a super important dataset! Thanks for making it. Many people are asking for PG-19, so it would be great to have that in the library as soon as possible @thomwolf .", "created_at": "2020-06-25T09:27:09Z...
2020-06-23T22:03:52Z
2020-07-06T07:55:59Z
2020-07-06T07:55:59Z
CONTRIBUTOR
null
null
null
https://github.com/huggingface/nlp/issues/274 Add functioning PG19 dataset with dummy data `cos_e.py` was just auto-linted by `make style`
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/306/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/306/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/306.diff", "html_url": "https://github.com/huggingface/datasets/pull/306", "merged_at": "2020-07-06T07:55:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/306.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/306...
2020-06-23T22:03:52Z
2020-07-06T07:55:59Z
12 days, 9:52:07
true
https://api.github.com/repos/huggingface/datasets/issues/305
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/305/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/305/comments
https://api.github.com/repos/huggingface/datasets/issues/305/events
https://github.com/huggingface/datasets/issues/305
644,148,149
MDU6SXNzdWU2NDQxNDgxNDk=
305
Importing downloaded package repository fails
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
[ { "color": "25b21e", "default": false, "description": "A bug in a metric script", "id": 2067393914, "name": "metric bug", "node_id": "MDU6TGFiZWwyMDY3MzkzOTE0", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug" } ]
closed
false
null
[]
null
[]
2020-06-23T21:09:05Z
2020-07-30T16:44:23Z
2020-07-30T16:44:23Z
MEMBER
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
The `get_imports` function in `src/nlp/load.py` has a feature to download a package as a zip archive of the github repository and import functions from the unpacked directory. This is used for example in the `metrics/coval.py` file, and would be useful to add BLEURT (@ankparikh). Currently however, the code seems to...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/305/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/305/timeline
null
completed
null
null
2020-06-23T21:09:05Z
2020-07-30T16:44:23Z
36 days, 19:35:18
false
https://api.github.com/repos/huggingface/datasets/issues/304
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/304/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/304/comments
https://api.github.com/repos/huggingface/datasets/issues/304/events
https://github.com/huggingface/datasets/issues/304
644,091,970
MDU6SXNzdWU2NDQwOTE5NzA=
304
Problem while printing doc string when instantiating multiple metrics.
{ "avatar_url": "https://avatars.githubusercontent.com/u/51091425?v=4", "events_url": "https://api.github.com/users/codehunk628/events{/privacy}", "followers_url": "https://api.github.com/users/codehunk628/followers", "following_url": "https://api.github.com/users/codehunk628/following{/other_user}", "gists_u...
[ { "color": "25b21e", "default": false, "description": "A bug in a metric script", "id": 2067393914, "name": "metric bug", "node_id": "MDU6TGFiZWwyMDY3MzkzOTE0", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug" } ]
closed
false
null
[]
null
[]
2020-06-23T19:32:05Z
2020-07-22T09:50:58Z
2020-07-22T09:50:58Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
When I load more than one metric and try to print doc string of a particular metric,. It shows the doc strings of all imported metric one after the other which looks quite confusing and clumsy. Attached [Colab](https://colab.research.google.com/drive/13H0ZgyQ2se0mqJ2yyew0bNEgJuHaJ8H3?usp=sharing) Notebook for problem ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/304/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/304/timeline
null
completed
null
null
2020-06-23T19:32:05Z
2020-07-22T09:50:58Z
28 days, 14:18:53
false
https://api.github.com/repos/huggingface/datasets/issues/302
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/302/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/302/comments
https://api.github.com/repos/huggingface/datasets/issues/302/events
https://github.com/huggingface/datasets/issues/302
643,910,418
MDU6SXNzdWU2NDM5MTA0MTg=
302
Question - Sign Language Datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "events_url": "https://api.github.com/users/AmitMY/events{/privacy}", "followers_url": "https://api.github.com/users/AmitMY/followers", "following_url": "https://api.github.com/users/AmitMY/following{/other_user}", "gists_url": "https://ap...
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "c5def5", "default": fals...
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Even more complicating - \r\n\r\nAs I see it, datasets can have \"addons\".\r\nFor example, the WebNLG dataset is a dataset for data-to-text. However, a work of mine and other works enriched this dataset with text plans / underlying text structures. In that ca...
2020-06-23T14:53:40Z
2020-11-25T11:25:33Z
2020-11-25T11:25:33Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
An emerging field in NLP is SLP - sign language processing. I was wondering about adding datasets here, specifically because it's shaping up to be large and easily usable. The metrics for sign language to text translation are the same. So, what do you think about (me, or others) adding datasets here? An exa...
{ "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "events_url": "https://api.github.com/users/AmitMY/events{/privacy}", "followers_url": "https://api.github.com/users/AmitMY/followers", "following_url": "https://api.github.com/users/AmitMY/following{/other_user}", "gists_url": "https://ap...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/302/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/302/timeline
null
completed
null
null
2020-06-23T14:53:40Z
2020-11-25T11:25:33Z
154 days, 20:31:53
false
https://api.github.com/repos/huggingface/datasets/issues/303
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/303/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/303/comments
https://api.github.com/repos/huggingface/datasets/issues/303/events
https://github.com/huggingface/datasets/pull/303
643,912,464
MDExOlB1bGxSZXF1ZXN0NDM4NjI3Nzcw
303
allow to move files across file systems
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[]
2020-06-23T14:56:08Z
2020-06-23T15:08:44Z
2020-06-23T15:08:43Z
MEMBER
null
null
null
Users are allowed to use the `cache_dir` that they want. Therefore it can happen that we try to move files across filesystems. We were using `os.rename` that doesn't allow that, so I changed some of them to `shutil.move`. This should fix #301
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/303/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/303/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/303.diff", "html_url": "https://github.com/huggingface/datasets/pull/303", "merged_at": "2020-06-23T15:08:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/303.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/303...
2020-06-23T14:56:08Z
2020-06-23T15:08:43Z
0:12:35
true
https://api.github.com/repos/huggingface/datasets/issues/301
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/301/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/301/comments
https://api.github.com/repos/huggingface/datasets/issues/301/events
https://github.com/huggingface/datasets/issues/301
643,763,525
MDU6SXNzdWU2NDM3NjM1MjU=
301
Setting cache_dir gives error on wikipedia download
{ "avatar_url": "https://avatars.githubusercontent.com/u/33862536?v=4", "events_url": "https://api.github.com/users/hallvagi/events{/privacy}", "followers_url": "https://api.github.com/users/hallvagi/followers", "following_url": "https://api.github.com/users/hallvagi/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Whoops didn't mean to close this one.\r\nI did some changes, could you try to run it from the master branch ?", "created_at": "2020-06-23T15:42:45Z", "html_url": "https://github.com/huggingface/datasets/issues/301#issuecomment-648246960", "id": 64824696...
2020-06-23T11:31:44Z
2020-06-24T07:05:07Z
2020-06-24T07:05:07Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
First of all thank you for a super handy library! I'd like to download large files to a specific drive so I set `cache_dir=my_path`. This works fine with e.g. imdb and squad. But on wikipedia I get an error: ``` nlp.load_dataset('wikipedia', '20200501.de', split = 'train', cache_dir=my_path) ``` ``` OSError ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/33862536?v=4", "events_url": "https://api.github.com/users/hallvagi/events{/privacy}", "followers_url": "https://api.github.com/users/hallvagi/followers", "following_url": "https://api.github.com/users/hallvagi/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/301/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/301/timeline
null
completed
null
null
2020-06-23T11:31:44Z
2020-06-24T07:05:07Z
19:33:23
false
https://api.github.com/repos/huggingface/datasets/issues/300
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/300/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/300/comments
https://api.github.com/repos/huggingface/datasets/issues/300/events
https://github.com/huggingface/datasets/pull/300
643,688,304
MDExOlB1bGxSZXF1ZXN0NDM4NDQ4Mjk1
300
Fix bertscore references
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[]
2020-06-23T09:38:59Z
2020-06-23T14:47:38Z
2020-06-23T14:47:37Z
MEMBER
null
null
null
I added some type checking for metrics. There was an issue where a metric could interpret a string a a list. A `ValueError` is raised if a string is given instead of a list. Moreover I added support for both strings and lists of strings for `references` in `bertscore`, as it is the case in the original code. Both...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/300/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/300/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/300.diff", "html_url": "https://github.com/huggingface/datasets/pull/300", "merged_at": "2020-06-23T14:47:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/300.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/300...
2020-06-23T09:38:59Z
2020-06-23T14:47:37Z
5:08:38
true
https://api.github.com/repos/huggingface/datasets/issues/299
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/299/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/299/comments
https://api.github.com/repos/huggingface/datasets/issues/299/events
https://github.com/huggingface/datasets/pull/299
643,611,557
MDExOlB1bGxSZXF1ZXN0NDM4Mzg0NDgw
299
remove some print in snli file
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "I guess you can just rebase from master to fix the CI", "created_at": "2020-06-23T07:56:08Z", "html_url": "https://github.com/huggingface/datasets/pull/299#issuecomment-647976464", "id": 647976464, "issue_url": "https://api.github.com/repos/huggingf...
2020-06-23T07:46:06Z
2020-06-23T08:10:46Z
2020-06-23T08:10:44Z
CONTRIBUTOR
null
null
null
This PR removes unwanted `print` statements in some files such as `snli.py`
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/299/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/299/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/299.diff", "html_url": "https://github.com/huggingface/datasets/pull/299", "merged_at": "2020-06-23T08:10:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/299.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/299...
2020-06-23T07:46:06Z
2020-06-23T08:10:44Z
0:24:38
true
https://api.github.com/repos/huggingface/datasets/issues/298
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/298/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/298/comments
https://api.github.com/repos/huggingface/datasets/issues/298/events
https://github.com/huggingface/datasets/pull/298
643,603,804
MDExOlB1bGxSZXF1ZXN0NDM4Mzc4MDM4
298
Add searchable datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Looks very cool! Only looked at it superficially though", "created_at": "2020-06-23T10:50:43Z", "html_url": "https://github.com/huggingface/datasets/pull/298#issuecomment-648066503", "id": 648066503, "issue_url": "https://api.github.com/repos/h...
2020-06-23T07:33:03Z
2020-06-26T07:50:44Z
2020-06-26T07:50:43Z
MEMBER
null
null
null
# Better support for Numpy format + Add Indexed Datasets I was working on adding Indexed Datasets but in the meantime I had to also add more support for Numpy arrays in the lib. ## Better support for Numpy format New features: - New fast method to convert Numpy arrays from Arrow structure (up to x100 speed up...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/298/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/298/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/298.diff", "html_url": "https://github.com/huggingface/datasets/pull/298", "merged_at": "2020-06-26T07:50:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/298.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/298...
2020-06-23T07:33:03Z
2020-06-26T07:50:43Z
3 days, 0:17:40
true
https://api.github.com/repos/huggingface/datasets/issues/297
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/297/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/297/comments
https://api.github.com/repos/huggingface/datasets/issues/297/events
https://github.com/huggingface/datasets/issues/297
643,444,625
MDU6SXNzdWU2NDM0NDQ2MjU=
297
Error in Demo for Specific Datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/60150701?v=4", "events_url": "https://api.github.com/users/s-jse/events{/privacy}", "followers_url": "https://api.github.com/users/s-jse/followers", "following_url": "https://api.github.com/users/s-jse/following{/other_user}", "gists_url": "https://api....
[ { "color": "94203D", "default": false, "description": "", "id": 2107841032, "name": "nlp-viewer", "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer" } ]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Thanks for reporting these errors :)\r\n\r\nI can actually see two issues here.\r\n\r\nFirst, datasets like `natural_questions` require apache_beam to be processed. Right now the import is not at the right place so we have this error message. However, even the impo...
2020-06-23T00:38:42Z
2020-07-17T17:43:06Z
2020-07-17T17:43:06Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Selecting `natural_questions` or `newsroom` dataset in the online demo results in an error similar to the following. ![image](https://user-images.githubusercontent.com/60150701/85347842-ac861900-b4ae-11ea-98c4-a53a00934783.png)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/297/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/297/timeline
null
completed
null
null
2020-06-23T00:38:42Z
2020-07-17T17:43:06Z
24 days, 17:04:24
false
https://api.github.com/repos/huggingface/datasets/issues/296
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/296/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/296/comments
https://api.github.com/repos/huggingface/datasets/issues/296/events
https://github.com/huggingface/datasets/issues/296
643,423,717
MDU6SXNzdWU2NDM0MjM3MTc=
296
snli -1 labels
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "@jxmorris12 , we use `-1` to label examples for which `gold label` is missing (`gold label = -` in the original dataset). ", "created_at": "2020-06-23T07:41:56Z", "html_url": "https://github.com/huggingface/datasets/issues/296#issuecomment-647969288"...
2020-06-22T23:33:30Z
2020-06-23T14:41:59Z
2020-06-23T14:41:58Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I'm trying to train a model on the SNLI dataset. Why does it have so many -1 labels? ``` import nlp from collections import Counter data = nlp.load_dataset('snli')['train'] print(Counter(data['label'])) Counter({0: 183416, 2: 183187, 1: 182764, -1: 785}) ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/296/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/296/timeline
null
completed
null
null
2020-06-22T23:33:30Z
2020-06-23T14:41:58Z
15:08:28
false
https://api.github.com/repos/huggingface/datasets/issues/295
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/295/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/295/comments
https://api.github.com/repos/huggingface/datasets/issues/295/events
https://github.com/huggingface/datasets/issues/295
643,245,412
MDU6SXNzdWU2NDMyNDU0MTI=
295
Improve input warning for evaluation metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/19514537?v=4", "events_url": "https://api.github.com/users/Tiiiger/events{/privacy}", "followers_url": "https://api.github.com/users/Tiiiger/followers", "following_url": "https://api.github.com/users/Tiiiger/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[]
2020-06-22T17:28:57Z
2020-06-23T14:47:37Z
2020-06-23T14:47:37Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi, I am the author of `bert_score`. Recently, we received [ an issue ](https://github.com/Tiiiger/bert_score/issues/62) reporting a problem in using `bert_score` from the `nlp` package (also see #238 in this repo). After looking into this, I realized that the problem arises from the format `nlp.Metric` takes inpu...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/295/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/295/timeline
null
completed
null
null
2020-06-22T17:28:57Z
2020-06-23T14:47:37Z
21:18:40
false
https://api.github.com/repos/huggingface/datasets/issues/294
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/294/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/294/comments
https://api.github.com/repos/huggingface/datasets/issues/294/events
https://github.com/huggingface/datasets/issues/294
643,181,179
MDU6SXNzdWU2NDMxODExNzk=
294
Cannot load arxiv dataset on MacOS?
{ "avatar_url": "https://avatars.githubusercontent.com/u/8917831?v=4", "events_url": "https://api.github.com/users/JohnGiorgi/events{/privacy}", "followers_url": "https://api.github.com/users/JohnGiorgi/followers", "following_url": "https://api.github.com/users/JohnGiorgi/following{/other_user}", "gists_url":...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "I couldn't replicate this issue on my macbook :/\r\nCould you try to play with different encodings in `with open(path, encoding=...) as f` in scientific_papers.py:L108 ?", "created_at": "2020-06-23T10:08:55Z", "html_url": "https://github.com/huggingface/dat...
2020-06-22T15:46:55Z
2020-06-30T15:25:10Z
2020-06-30T15:25:10Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I am having trouble loading the `"arxiv"` config from the `"scientific_papers"` dataset on MacOS. When I try loading the dataset with: ```python arxiv = nlp.load_dataset("scientific_papers", "arxiv") ``` I get the following stack trace: ```bash JSONDecodeError Traceback (most recen...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8917831?v=4", "events_url": "https://api.github.com/users/JohnGiorgi/events{/privacy}", "followers_url": "https://api.github.com/users/JohnGiorgi/followers", "following_url": "https://api.github.com/users/JohnGiorgi/following{/other_user}", "gists_url":...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/294/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/294/timeline
null
completed
null
null
2020-06-22T15:46:55Z
2020-06-30T15:25:10Z
7 days, 23:38:15
false
https://api.github.com/repos/huggingface/datasets/issues/293
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/293/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/293/comments
https://api.github.com/repos/huggingface/datasets/issues/293/events
https://github.com/huggingface/datasets/pull/293
642,942,182
MDExOlB1bGxSZXF1ZXN0NDM3ODM1ODI4
293
Don't test community datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[]
2020-06-22T10:15:33Z
2020-06-22T11:07:00Z
2020-06-22T11:06:59Z
MEMBER
null
null
null
This PR disables testing for community datasets on aws. It should fix the CI that is currently failing.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/293/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/293/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/293.diff", "html_url": "https://github.com/huggingface/datasets/pull/293", "merged_at": "2020-06-22T11:06:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/293.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/293...
2020-06-22T10:15:33Z
2020-06-22T11:06:59Z
0:51:26
true
https://api.github.com/repos/huggingface/datasets/issues/292
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/292/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/292/comments
https://api.github.com/repos/huggingface/datasets/issues/292/events
https://github.com/huggingface/datasets/pull/292
642,897,797
MDExOlB1bGxSZXF1ZXN0NDM3Nzk4NTM2
292
Update metadata for x_stance dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/5830820?v=4", "events_url": "https://api.github.com/users/jvamvas/events{/privacy}", "followers_url": "https://api.github.com/users/jvamvas/followers", "following_url": "https://api.github.com/users/jvamvas/following{/other_user}", "gists_url": "https:/...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Great! Thanks @jvamvas for these updates.\r\n", "created_at": "2020-06-22T10:25:57Z", "html_url": "https://github.com/huggingface/datasets/pull/292#issuecomment-647429164", "id": 647429164, "issue_url": "https://api.github.com/repos/huggingface...
2020-06-22T09:13:26Z
2020-06-23T08:07:24Z
2020-06-23T08:07:24Z
CONTRIBUTOR
null
null
null
Thank you for featuring the x_stance dataset in your library. This PR updates some metadata: - Citation: Replace preprint with proceedings - URL: Use a URL with long-term availability
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/292/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/292/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/292.diff", "html_url": "https://github.com/huggingface/datasets/pull/292", "merged_at": "2020-06-23T08:07:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/292.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/292...
2020-06-22T09:13:26Z
2020-06-23T08:07:24Z
22:53:58
true
https://api.github.com/repos/huggingface/datasets/issues/291
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/291/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/291/comments
https://api.github.com/repos/huggingface/datasets/issues/291/events
https://github.com/huggingface/datasets/pull/291
642,688,450
MDExOlB1bGxSZXF1ZXN0NDM3NjM1NjMy
291
break statement not required
{ "avatar_url": "https://avatars.githubusercontent.com/u/12967587?v=4", "events_url": "https://api.github.com/users/mayurnewase/events{/privacy}", "followers_url": "https://api.github.com/users/mayurnewase/followers", "following_url": "https://api.github.com/users/mayurnewase/following{/other_user}", "gists_u...
[]
closed
false
null
[]
null
[ { "author_association": "NONE", "body": "I guess,test failing due to connection error?", "created_at": "2020-06-22T01:51:10Z", "html_url": "https://github.com/huggingface/datasets/pull/291#issuecomment-647222037", "id": 647222037, "issue_url": "https://api.github.com/repos/huggingface/datase...
2020-06-22T01:40:55Z
2020-06-23T17:57:58Z
2020-06-23T09:37:02Z
NONE
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/291/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/291/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/291.diff", "html_url": "https://github.com/huggingface/datasets/pull/291", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/291.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/291" }
2020-06-22T01:40:55Z
2020-06-23T09:37:02Z
1 day, 7:56:07
true
https://api.github.com/repos/huggingface/datasets/issues/290
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/290/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/290/comments
https://api.github.com/repos/huggingface/datasets/issues/290/events
https://github.com/huggingface/datasets/issues/290
641,978,286
MDU6SXNzdWU2NDE5NzgyODY=
290
ConnectionError - Eli5 dataset download
{ "avatar_url": "https://avatars.githubusercontent.com/u/8490096?v=4", "events_url": "https://api.github.com/users/JovanNj/events{/privacy}", "followers_url": "https://api.github.com/users/JovanNj/followers", "following_url": "https://api.github.com/users/JovanNj/following{/other_user}", "gists_url": "https:/...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "It should ne fixed now, thanks for reporting this one :)\r\nIt was an issue on our google storage.\r\n\r\nLet me now if you're still facing this issue.", "created_at": "2020-06-19T14:41:17Z", "html_url": "https://github.com/huggingface/datasets/issues/290#i...
2020-06-19T13:40:33Z
2020-06-20T13:22:24Z
2020-06-20T13:22:24Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi, I have a problem with downloading Eli5 dataset. When typing `nlp.load_dataset('eli5')`, I get ConnectionError: Couldn't reach https://storage.googleapis.com/huggingface-nlp/cache/datasets/eli5/LFQA_reddit/1.0.0/explain_like_im_five-train_eli5.arrow I would appreciate if you could help me with this issue.
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/290/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/290/timeline
null
completed
null
null
2020-06-19T13:40:33Z
2020-06-20T13:22:24Z
23:41:51
false
https://api.github.com/repos/huggingface/datasets/issues/289
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/289/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/289/comments
https://api.github.com/repos/huggingface/datasets/issues/289/events
https://github.com/huggingface/datasets/pull/289
641,934,194
MDExOlB1bGxSZXF1ZXN0NDM3MDc0MTM3
289
update xsum
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Looks cool!\r\n@mariamabarham can you add a detailed description here what exactly is changed and how the user can load xsum now?", "created_at": "2020-06-19T16:45:38Z", "html_url": "https://github.com/huggingface/datasets/pull/289#issuecomment-6467434...
2020-06-19T12:28:32Z
2020-06-22T13:27:26Z
2020-06-22T07:20:07Z
CONTRIBUTOR
null
null
null
This PR makes the following update to the xsum dataset: - Manual download is not required anymore - dataset can be loaded as follow: `nlp.load_dataset('xsum')` **Important** Instead of using on outdated url to download the data: "https://raw.githubusercontent.com/EdinburghNLP/XSum/master/XSum-Dataset/XSum...
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/289/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/289/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/289.diff", "html_url": "https://github.com/huggingface/datasets/pull/289", "merged_at": "2020-06-22T07:20:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/289.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/289...
2020-06-19T12:28:32Z
2020-06-22T07:20:07Z
2 days, 18:51:35
true
https://api.github.com/repos/huggingface/datasets/issues/288
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/288/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/288/comments
https://api.github.com/repos/huggingface/datasets/issues/288/events
https://github.com/huggingface/datasets/issues/288
641,888,610
MDU6SXNzdWU2NDE4ODg2MTA=
288
Error at the first example in README: AttributeError: module 'dill' has no attribute '_dill'
{ "avatar_url": "https://avatars.githubusercontent.com/u/14964542?v=4", "events_url": "https://api.github.com/users/wutong8023/events{/privacy}", "followers_url": "https://api.github.com/users/wutong8023/followers", "following_url": "https://api.github.com/users/wutong8023/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "It looks like the bug comes from `dill`. Which version of `dill` are you using ?", "created_at": "2020-06-19T11:52:19Z", "html_url": "https://github.com/huggingface/datasets/issues/288#issuecomment-646594079", "id": 646594079, "issue_url": "https://...
2020-06-19T11:01:22Z
2020-06-21T09:05:11Z
2020-06-21T09:05:11Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
/Users/parasol_tree/anaconda3/lib/python3.6/site-packages/tensorflow/python/framework/dtypes.py:469: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'. _np_qint8 = np.dtype([("qint8", np.int8, 1)]) /Users/...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/288/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/288/timeline
null
completed
null
null
2020-06-19T11:01:22Z
2020-06-21T09:05:11Z
1 day, 22:03:49
false
https://api.github.com/repos/huggingface/datasets/issues/287
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/287/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/287/comments
https://api.github.com/repos/huggingface/datasets/issues/287/events
https://github.com/huggingface/datasets/pull/287
641,800,227
MDExOlB1bGxSZXF1ZXN0NDM2OTY0NTg0
287
fix squad_v2 metric
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[]
2020-06-19T08:24:46Z
2020-06-19T08:33:43Z
2020-06-19T08:33:41Z
MEMBER
null
null
null
Fix #280 The imports were wrong
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/287/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/287/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/287.diff", "html_url": "https://github.com/huggingface/datasets/pull/287", "merged_at": "2020-06-19T08:33:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/287.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/287...
2020-06-19T08:24:46Z
2020-06-19T08:33:41Z
0:08:55
true
https://api.github.com/repos/huggingface/datasets/issues/286
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/286/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/286/comments
https://api.github.com/repos/huggingface/datasets/issues/286/events
https://github.com/huggingface/datasets/pull/286
641,585,758
MDExOlB1bGxSZXF1ZXN0NDM2NzkzMjI4
286
Add ANLI dataset.
{ "avatar_url": "https://avatars.githubusercontent.com/u/11016329?v=4", "events_url": "https://api.github.com/users/easonnie/events{/privacy}", "followers_url": "https://api.github.com/users/easonnie/followers", "following_url": "https://api.github.com/users/easonnie/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Awesome!! Thanks @easonnie.\r\nLet's wait for additional reviews maybe from @lhoestq @patrickvonplaten @jplu", "created_at": "2020-06-19T08:41:11Z", "html_url": "https://github.com/huggingface/datasets/pull/286#issuecomment-646516979", "id": 646516...
2020-06-18T22:27:30Z
2020-06-22T12:23:27Z
2020-06-22T12:23:27Z
CONTRIBUTOR
null
null
null
I completed all the steps in https://github.com/huggingface/nlp/blob/master/CONTRIBUTING.md#how-to-add-a-dataset and push the code for ANLI. Please let me know if there are any errors.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/286/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/286/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/286.diff", "html_url": "https://github.com/huggingface/datasets/pull/286", "merged_at": "2020-06-22T12:23:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/286.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/286...
2020-06-18T22:27:30Z
2020-06-22T12:23:27Z
3 days, 13:55:57
true
https://api.github.com/repos/huggingface/datasets/issues/285
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/285/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/285/comments
https://api.github.com/repos/huggingface/datasets/issues/285/events
https://github.com/huggingface/datasets/pull/285
641,360,702
MDExOlB1bGxSZXF1ZXN0NDM2NjAyMjk4
285
Consistent formatting of citations
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Circle CI shuold be green :-) ", "created_at": "2020-06-19T15:44:08Z", "html_url": "https://github.com/huggingface/datasets/pull/285#issuecomment-646705096", "id": 646705096, "issue_url": "https://api.github.com/repos/huggingface/datasets/issue...
2020-06-18T16:25:23Z
2020-06-22T08:09:25Z
2020-06-22T08:09:24Z
CONTRIBUTOR
null
null
null
#283
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/285/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/285/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/285.diff", "html_url": "https://github.com/huggingface/datasets/pull/285", "merged_at": "2020-06-22T08:09:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/285.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/285...
2020-06-18T16:25:23Z
2020-06-22T08:09:24Z
3 days, 15:44:01
true
https://api.github.com/repos/huggingface/datasets/issues/284
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/284/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/284/comments
https://api.github.com/repos/huggingface/datasets/issues/284/events
https://github.com/huggingface/datasets/pull/284
641,337,217
MDExOlB1bGxSZXF1ZXN0NDM2NTgxODQ2
284
Fix manual download instructions
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Verified that this works, thanks!", "created_at": "2020-06-18T17:01:08Z", "html_url": "https://github.com/huggingface/datasets/pull/284#issuecomment-646175431", "id": 646175431, "issue_url": "https://api.github.com/repos/huggingface/datasets/is...
2020-06-18T15:59:57Z
2020-06-19T08:24:21Z
2020-06-19T08:24:19Z
CONTRIBUTOR
null
null
null
This PR replaces the static `DatasetBulider` variable `MANUAL_DOWNLOAD_INSTRUCTIONS` by a property function `manual_download_instructions()`. Some datasets like XTREME and all WMT need the manual data dir only for a small fraction of the possible configs. After some brainstorming with @mariamabarham and @lhoestq...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/284/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/284/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/284.diff", "html_url": "https://github.com/huggingface/datasets/pull/284", "merged_at": "2020-06-19T08:24:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/284.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/284...
2020-06-18T15:59:57Z
2020-06-19T08:24:19Z
16:24:22
true
https://api.github.com/repos/huggingface/datasets/issues/282
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/282/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/282/comments
https://api.github.com/repos/huggingface/datasets/issues/282/events
https://github.com/huggingface/datasets/pull/282
641,217,759
MDExOlB1bGxSZXF1ZXN0NDM2NDgxNzMy
282
Update dataset_info from gcs
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[]
2020-06-18T13:41:15Z
2020-06-18T16:24:52Z
2020-06-18T16:24:51Z
MEMBER
null
null
null
Some datasets are hosted on gcs (wikipedia for example). In this PR I make sure that, when a user loads such datasets, the file_instructions are built using the dataset_info.json from gcs and not from the info extracted from the local `dataset_infos.json` (the one that contain the info for each config). Indeed local fi...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/282/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/282/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/282.diff", "html_url": "https://github.com/huggingface/datasets/pull/282", "merged_at": "2020-06-18T16:24:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/282.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/282...
2020-06-18T13:41:15Z
2020-06-18T16:24:51Z
2:43:36
true
https://api.github.com/repos/huggingface/datasets/issues/283
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/283/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/283/comments
https://api.github.com/repos/huggingface/datasets/issues/283/events
https://github.com/huggingface/datasets/issues/283
641,270,439
MDU6SXNzdWU2NDEyNzA0Mzk=
283
Consistent formatting of citations
{ "avatar_url": "https://avatars.githubusercontent.com/u/35882?v=4", "events_url": "https://api.github.com/users/srush/events{/privacy}", "followers_url": "https://api.github.com/users/srush/followers", "following_url": "https://api.github.com/users/srush/following{/other_user}", "gists_url": "https://api.git...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_...
null
[]
2020-06-18T14:48:45Z
2020-06-22T17:30:46Z
2020-06-22T17:30:46Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
The citations are all of a different format, some have "```" and have text inside, others are proper bibtex. Can we make it so that they all are proper citations, i.e. parse by the bibtex spec: https://bibtexparser.readthedocs.io/en/master/
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/283/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/283/timeline
null
completed
null
null
2020-06-18T14:48:45Z
2020-06-22T17:30:46Z
4 days, 2:42:01
false
https://api.github.com/repos/huggingface/datasets/issues/281
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/281/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/281/comments
https://api.github.com/repos/huggingface/datasets/issues/281/events
https://github.com/huggingface/datasets/issues/281
641,067,856
MDU6SXNzdWU2NDEwNjc4NTY=
281
Private/sensitive data
{ "avatar_url": "https://avatars.githubusercontent.com/u/6368040?v=4", "events_url": "https://api.github.com/users/MFreidank/events{/privacy}", "followers_url": "https://api.github.com/users/MFreidank/followers", "following_url": "https://api.github.com/users/MFreidank/following{/other_user}", "gists_url": "h...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Hi @MFreidank, you should already be able to load a dataset from local sources, indeed. (ping @lhoestq and @jplu)\r\n\r\nWe're also thinking about the ability to host private datasets on a hosted bucket with permission management, but that's further down the road."...
2020-06-18T09:47:27Z
2020-06-20T13:15:12Z
2020-06-20T13:15:12Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi all, Thanks for this fantastic library, it makes it very easy to do prototyping for NLP projects interchangeably between TF/Pytorch. Unfortunately, there is data that cannot easily be shared publicly as it may contain sensitive information. Is there support/a plan to support such data with NLP, e.g. by readin...
{ "avatar_url": "https://avatars.githubusercontent.com/u/6368040?v=4", "events_url": "https://api.github.com/users/MFreidank/events{/privacy}", "followers_url": "https://api.github.com/users/MFreidank/followers", "following_url": "https://api.github.com/users/MFreidank/following{/other_user}", "gists_url": "h...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/281/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/281/timeline
null
completed
null
null
2020-06-18T09:47:27Z
2020-06-20T13:15:12Z
2 days, 3:27:45
false
https://api.github.com/repos/huggingface/datasets/issues/280
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/280/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/280/comments
https://api.github.com/repos/huggingface/datasets/issues/280/events
https://github.com/huggingface/datasets/issues/280
640,677,615
MDU6SXNzdWU2NDA2Nzc2MTU=
280
Error with SquadV2 Metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/32203792?v=4", "events_url": "https://api.github.com/users/avinregmi/events{/privacy}", "followers_url": "https://api.github.com/users/avinregmi/followers", "following_url": "https://api.github.com/users/avinregmi/following{/other_user}", "gists_url": "...
[]
closed
false
null
[]
null
[]
2020-06-17T19:10:54Z
2020-06-19T08:33:41Z
2020-06-19T08:33:41Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I can't seem to import squad v2 metrics. **squad_metric = nlp.load_metric('squad_v2')** **This throws me an error.:** ``` ImportError Traceback (most recent call last) <ipython-input-8-170b6a170555> in <module> ----> 1 squad_metric = nlp.load_metric('squad_v2') ~/env/lib6...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/280/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/280/timeline
null
completed
null
null
2020-06-17T19:10:54Z
2020-06-19T08:33:41Z
1 day, 13:22:47
false
https://api.github.com/repos/huggingface/datasets/issues/278
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/278/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/278/comments
https://api.github.com/repos/huggingface/datasets/issues/278/events
https://github.com/huggingface/datasets/issues/278
640,518,917
MDU6SXNzdWU2NDA1MTg5MTc=
278
MemoryError when loading German Wikipedia
{ "avatar_url": "https://avatars.githubusercontent.com/u/4698028?v=4", "events_url": "https://api.github.com/users/gregburman/events{/privacy}", "followers_url": "https://api.github.com/users/gregburman/followers", "following_url": "https://api.github.com/users/gregburman/following{/other_user}", "gists_url":...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Hi !\r\n\r\nAs you noticed, \"big\" datasets like Wikipedia require apache beam to be processed.\r\nHowever users usually don't have an apache beam runtime available (spark, dataflow, etc.) so our goal for this library is to also make available processed versions o...
2020-06-17T15:06:21Z
2020-06-19T12:53:02Z
2020-06-19T12:53:02Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi, first off let me say thank you for all the awesome work you're doing at Hugging Face across all your projects (NLP, Transformers, Tokenizers) - they're all amazing contributions to us working with NLP models :) I'm trying to download the German Wikipedia dataset as follows: ``` wiki = nlp.load_dataset("wikip...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/278/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/278/timeline
null
completed
null
null
2020-06-17T15:06:21Z
2020-06-19T12:53:02Z
1 day, 21:46:41
false
https://api.github.com/repos/huggingface/datasets/issues/279
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/279/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/279/comments
https://api.github.com/repos/huggingface/datasets/issues/279/events
https://github.com/huggingface/datasets/issues/279
640,611,692
MDU6SXNzdWU2NDA2MTE2OTI=
279
Dataset Preprocessing Cache with .map() function not working as expected
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "When you're processing a dataset with `.map`, it checks whether it has already done this computation using a hash based on the function and the input (using some fancy serialization with `dill`). If you found that it doesn't work as expected in some cases, let us k...
2020-06-17T17:17:21Z
2021-07-06T21:43:28Z
2021-04-18T23:43:49Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I've been having issues with reproducibility when loading and processing datasets with the `.map` function. I was only able to resolve them by clearing all of the cache files on my system. Is there a way to disable using the cache when processing a dataset? As I make minor processing changes on the same dataset, I ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/279/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/279/timeline
null
completed
null
null
2020-06-17T17:17:21Z
2021-04-18T23:43:49Z
305 days, 6:26:28
false
https://api.github.com/repos/huggingface/datasets/issues/277
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/277/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/277/comments
https://api.github.com/repos/huggingface/datasets/issues/277/events
https://github.com/huggingface/datasets/issues/277
640,163,053
MDU6SXNzdWU2NDAxNjMwNTM=
277
Empty samples in glue/qqp
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "We are only wrapping the original dataset.\r\n\r\nMaybe try to ask on the GLUE mailing list or reach out to the original authors?", "created_at": "2020-06-20T13:26:12Z", "html_url": "https://github.com/huggingface/datasets/issues/277#issuecomment-646994358"...
2020-06-17T05:54:52Z
2020-06-21T00:21:45Z
2020-06-21T00:21:45Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
``` qqp = nlp.load_dataset('glue', 'qqp') print(qqp['train'][310121]) print(qqp['train'][362225]) ``` ``` {'question1': 'How can I create an Android app?', 'question2': '', 'label': 0, 'idx': 310137} {'question1': 'How can I develop android app?', 'question2': '', 'label': 0, 'idx': 362246} ``` Notice that que...
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/277/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/277/timeline
null
completed
null
null
2020-06-17T05:54:52Z
2020-06-21T00:21:45Z
3 days, 18:26:53
false
https://api.github.com/repos/huggingface/datasets/issues/275
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/275/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/275/comments
https://api.github.com/repos/huggingface/datasets/issues/275/events
https://github.com/huggingface/datasets/issues/275
639,439,052
MDU6SXNzdWU2Mzk0MzkwNTI=
275
NonMatchingChecksumError when loading pubmed dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/48441753?v=4", "events_url": "https://api.github.com/users/DavideStenner/events{/privacy}", "followers_url": "https://api.github.com/users/DavideStenner/followers", "following_url": "https://api.github.com/users/DavideStenner/following{/other_user}", "g...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "For some reason the files are not available for unauthenticated users right now (like the download service of this package). Instead of downloading the right files, it downloads the html of the error.\r\nAccording to the error it should be back again in 24h.\r\n\r\...
2020-06-16T07:31:51Z
2020-06-19T07:37:07Z
2020-06-19T07:37:07Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I get this error when i run `nlp.load_dataset('scientific_papers', 'pubmed', split = 'train[:50%]')`. The error is: ``` --------------------------------------------------------------------------- NonMatchingChecksumError Traceback (most recent call last) <ipython-input-2-7742dea167d0> in <module...
{ "avatar_url": "https://avatars.githubusercontent.com/u/48441753?v=4", "events_url": "https://api.github.com/users/DavideStenner/events{/privacy}", "followers_url": "https://api.github.com/users/DavideStenner/followers", "following_url": "https://api.github.com/users/DavideStenner/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/275/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/275/timeline
null
completed
null
null
2020-06-16T07:31:51Z
2020-06-19T07:37:07Z
3 days, 0:05:16
false
https://api.github.com/repos/huggingface/datasets/issues/276
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/276/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/276/comments
https://api.github.com/repos/huggingface/datasets/issues/276/events
https://github.com/huggingface/datasets/pull/276
639,490,858
MDExOlB1bGxSZXF1ZXN0NDM1MDY5Nzg5
276
Fix metric compute (original_instructions missing)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Awesome! This is working now:\r\n\r\n```python\r\nimport nlp \r\nseqeval = nlp.load_metric(\"seqeval\") \r\ny_true = [['O', 'O', 'O', 'B-MISC', 'I-MISC', 'I-MISC', 'O'], ['B-PER', 'I-PER', 'O']] \r\ny_pred = [['O', 'O', 'B-MISC', 'I-MISC', 'I-MISC', 'I-MISC', ...
2020-06-16T08:52:01Z
2020-06-18T07:41:45Z
2020-06-18T07:41:44Z
MEMBER
null
null
null
When loading arrow data we added in cc8d250 a way to specify the instructions that were used to store them with the loaded dataset. However metrics load data the same way but don't need instructions (we use one single file). In this PR I just make `original_instructions` optional when reading files to load a `Datas...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/276/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/276/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/276.diff", "html_url": "https://github.com/huggingface/datasets/pull/276", "merged_at": "2020-06-18T07:41:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/276.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/276...
2020-06-16T08:52:01Z
2020-06-18T07:41:44Z
1 day, 22:49:43
true
https://api.github.com/repos/huggingface/datasets/issues/274
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/274/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/274/comments
https://api.github.com/repos/huggingface/datasets/issues/274/events
https://github.com/huggingface/datasets/issues/274
639,156,625
MDU6SXNzdWU2MzkxNTY2MjU=
274
PG-19
{ "avatar_url": "https://avatars.githubusercontent.com/u/108653?v=4", "events_url": "https://api.github.com/users/lucidrains/events{/privacy}", "followers_url": "https://api.github.com/users/lucidrains/followers", "following_url": "https://api.github.com/users/lucidrains/following{/other_user}", "gists_url": ...
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Sounds good! Do you want to give it a try?", "created_at": "2020-06-16T06:46:25Z", "html_url": "https://github.com/huggingface/datasets/issues/274#issuecomment-644566626", "id": 644566626, "issue_url": "https://api.github.com/repos/huggingface/datas...
2020-06-15T21:02:26Z
2020-07-06T15:35:02Z
2020-07-06T15:35:02Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi, and thanks for all your open-sourced work, as always! I was wondering if you would be open to adding PG-19 to your collection of datasets. https://github.com/deepmind/pg19 It is often used for benchmarking long-range language modeling.
{ "avatar_url": "https://avatars.githubusercontent.com/u/108653?v=4", "events_url": "https://api.github.com/users/lucidrains/events{/privacy}", "followers_url": "https://api.github.com/users/lucidrains/followers", "following_url": "https://api.github.com/users/lucidrains/following{/other_user}", "gists_url": ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/274/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/274/timeline
null
completed
null
null
2020-06-15T21:02:26Z
2020-07-06T15:35:02Z
20 days, 18:32:36
false
https://api.github.com/repos/huggingface/datasets/issues/273
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/273/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/273/comments
https://api.github.com/repos/huggingface/datasets/issues/273/events
https://github.com/huggingface/datasets/pull/273
638,968,054
MDExOlB1bGxSZXF1ZXN0NDM0NjM0MzU4
273
update cos_e to add cos_e v1.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
[]
2020-06-15T16:03:22Z
2020-06-16T08:25:54Z
2020-06-16T08:25:52Z
CONTRIBUTOR
null
null
null
This PR updates the cos_e dataset to add v1.0 as requested here #163 @nazneenrajani
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/273/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/273/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/273.diff", "html_url": "https://github.com/huggingface/datasets/pull/273", "merged_at": "2020-06-16T08:25:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/273.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/273...
2020-06-15T16:03:22Z
2020-06-16T08:25:52Z
16:22:30
true
https://api.github.com/repos/huggingface/datasets/issues/272
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/272/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/272/comments
https://api.github.com/repos/huggingface/datasets/issues/272/events
https://github.com/huggingface/datasets/pull/272
638,307,313
MDExOlB1bGxSZXF1ZXN0NDM0MTExOTQ3
272
asd
{ "avatar_url": "https://avatars.githubusercontent.com/u/66900970?v=4", "events_url": "https://api.github.com/users/sn696/events{/privacy}", "followers_url": "https://api.github.com/users/sn696/followers", "following_url": "https://api.github.com/users/sn696/following{/other_user}", "gists_url": "https://api....
[]
closed
false
null
[]
null
[]
2020-06-14T08:20:38Z
2020-06-14T09:16:41Z
2020-06-14T09:16:41Z
NONE
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/272/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/272/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/272.diff", "html_url": "https://github.com/huggingface/datasets/pull/272", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/272.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/272" }
2020-06-14T08:20:38Z
2020-06-14T09:16:41Z
0:56:03
true
https://api.github.com/repos/huggingface/datasets/issues/271
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/271/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/271/comments
https://api.github.com/repos/huggingface/datasets/issues/271/events
https://github.com/huggingface/datasets/pull/271
638,135,754
MDExOlB1bGxSZXF1ZXN0NDMzOTg3NDkw
271
Fix allociné dataset configuration
{ "avatar_url": "https://avatars.githubusercontent.com/u/37028092?v=4", "events_url": "https://api.github.com/users/TheophileBlard/events{/privacy}", "followers_url": "https://api.github.com/users/TheophileBlard/followers", "following_url": "https://api.github.com/users/TheophileBlard/following{/other_user}", ...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Actually when there is only one configuration, then you don't need to specify the configuration in `load_dataset`. You can run:\r\n```python\r\ndataset = load_dataset('allocine')\r\n```\r\nand it works.\r\n\r\nMaybe we should take that into account in the nlp viewe...
2020-06-13T10:12:10Z
2020-06-18T07:41:21Z
2020-06-18T07:41:20Z
CONTRIBUTOR
null
null
null
This is a patch for #244. According to the [live nlp viewer](url), the Allociné dataset must be loaded with : ```python dataset = load_dataset('allocine', 'allocine') ``` This is redundant, as there is only one "dataset configuration", and should only be: ```python dataset = load_dataset('allocine') ``` This ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/271/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/271/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/271.diff", "html_url": "https://github.com/huggingface/datasets/pull/271", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/271.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/271" }
2020-06-13T10:12:10Z
2020-06-18T07:41:20Z
4 days, 21:29:10
true
https://api.github.com/repos/huggingface/datasets/issues/270
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/270/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/270/comments
https://api.github.com/repos/huggingface/datasets/issues/270/events
https://github.com/huggingface/datasets/issues/270
638,121,617
MDU6SXNzdWU2MzgxMjE2MTc=
270
c4 dataset is not viewable in nlpviewer demo
{ "avatar_url": "https://avatars.githubusercontent.com/u/6441313?v=4", "events_url": "https://api.github.com/users/rajarsheem/events{/privacy}", "followers_url": "https://api.github.com/users/rajarsheem/followers", "following_url": "https://api.github.com/users/rajarsheem/following{/other_user}", "gists_url":...
[ { "color": "94203D", "default": false, "description": "", "id": 2107841032, "name": "nlp-viewer", "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer" } ]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "C4 is too large to be shown in the viewer", "created_at": "2020-10-27T15:35:29Z", "html_url": "https://github.com/huggingface/datasets/issues/270#issuecomment-717328214", "id": 717328214, "issue_url": "https://api.github.com/repos/huggingface/datase...
2020-06-13T08:26:16Z
2020-10-27T15:35:29Z
2020-10-27T15:35:13Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I get the following error when I try to view the c4 dataset in [nlpviewer](https://huggingface.co/nlp/viewer/) ```python ModuleNotFoundError: No module named 'langdetect' Traceback: File "/home/sasha/.local/lib/python3.7/site-packages/streamlit/ScriptRunner.py", line 322, in _run_script exec(code, module.__d...
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/270/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/270/timeline
null
completed
null
null
2020-06-13T08:26:16Z
2020-10-27T15:35:13Z
136 days, 7:08:57
false
https://api.github.com/repos/huggingface/datasets/issues/269
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/269/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/269/comments
https://api.github.com/repos/huggingface/datasets/issues/269/events
https://github.com/huggingface/datasets/issues/269
638,106,774
MDU6SXNzdWU2MzgxMDY3NzQ=
269
Error in metric.compute: missing `original_instructions` argument
{ "avatar_url": "https://avatars.githubusercontent.com/u/1668462?v=4", "events_url": "https://api.github.com/users/zphang/events{/privacy}", "followers_url": "https://api.github.com/users/zphang/followers", "following_url": "https://api.github.com/users/zphang/following{/other_user}", "gists_url": "https://ap...
[ { "color": "25b21e", "default": false, "description": "A bug in a metric script", "id": 2067393914, "name": "metric bug", "node_id": "MDU6TGFiZWwyMDY3MzkzOTE0", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
[]
2020-06-13T06:26:54Z
2020-06-18T07:41:44Z
2020-06-18T07:41:44Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I'm running into an error using metrics for computation in the latest master as well as version 0.2.1. Here is a minimal example: ```python import nlp rte_metric = nlp.load_metric('glue', name="rte") rte_metric.compute( [0, 0, 1, 1], [0, 1, 0, 1], ) ``` ``` 181 # Read the predictio...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/269/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/269/timeline
null
completed
null
null
2020-06-13T06:26:54Z
2020-06-18T07:41:44Z
5 days, 1:14:50
false
https://api.github.com/repos/huggingface/datasets/issues/268
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/268/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/268/comments
https://api.github.com/repos/huggingface/datasets/issues/268/events
https://github.com/huggingface/datasets/pull/268
637,848,056
MDExOlB1bGxSZXF1ZXN0NDMzNzU5NzQ1
268
add Rotten Tomatoes Movie Review sentences sentiment dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "@jplu @thomwolf @patrickvonplaten @lhoestq -- How do I request reviewers? Thanks.", "created_at": "2020-06-15T18:42:46Z", "html_url": "https://github.com/huggingface/datasets/pull/268#issuecomment-644310478", "id": 644310478, "issue_url": "http...
2020-06-12T15:53:59Z
2020-06-18T07:46:24Z
2020-06-18T07:46:23Z
CONTRIBUTOR
null
null
null
Sentence-level movie reviews v1.0 from here: http://www.cs.cornell.edu/people/pabo/movie-review-data/
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/268/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/268/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/268.diff", "html_url": "https://github.com/huggingface/datasets/pull/268", "merged_at": "2020-06-18T07:46:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/268.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/268...
2020-06-12T15:53:59Z
2020-06-18T07:46:23Z
5 days, 15:52:24
true
https://api.github.com/repos/huggingface/datasets/issues/267
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/267/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/267/comments
https://api.github.com/repos/huggingface/datasets/issues/267/events
https://github.com/huggingface/datasets/issues/267
637,415,545
MDU6SXNzdWU2Mzc0MTU1NDU=
267
How can I load/find WMT en-romanian?
{ "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "events_url": "https://api.github.com/users/sshleifer/events{/privacy}", "followers_url": "https://api.github.com/users/sshleifer/followers", "following_url": "https://api.github.com/users/sshleifer/following{/other_user}", "gists_url": "h...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/followin...
null
[ { "author_association": "CONTRIBUTOR", "body": "I will take a look :-) ", "created_at": "2020-06-15T17:11:35Z", "html_url": "https://github.com/huggingface/datasets/issues/267#issuecomment-644259729", "id": 644259729, "issue_url": "https://api.github.com/repos/huggingface/datasets/issues/267...
2020-06-12T01:09:37Z
2020-06-19T08:24:19Z
2020-06-19T08:24:19Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I believe it is from `wmt16` When I run ```python wmt = nlp.load_dataset('wmt16') ``` I get: ```python AssertionError: The dataset wmt16 with config cs-en requires manual data. Please follow the manual download instructions: Some of the wmt configs here, require a manual download. Please look into wm...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/267/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/267/timeline
null
completed
null
null
2020-06-12T01:09:37Z
2020-06-19T08:24:19Z
7 days, 7:14:42
false
https://api.github.com/repos/huggingface/datasets/issues/266
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/266/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/266/comments
https://api.github.com/repos/huggingface/datasets/issues/266/events
https://github.com/huggingface/datasets/pull/266
637,156,392
MDExOlB1bGxSZXF1ZXN0NDMzMTk1NDgw
266
Add sort, shuffle, test_train_split and select methods
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Nice !\r\n\r\nAlso it looks like we can have a train_test_split method for free:\r\n```python\r\ntrain_indices, test_indices = train_test_split(range(len(dataset)))\r\ntrain = dataset.sort(indices=train_indices)\r\ntest = dataset.sort(indices=test_indices)\r\n```\r...
2020-06-11T16:22:20Z
2020-06-18T16:23:25Z
2020-06-18T16:23:24Z
MEMBER
null
null
null
Add a bunch of methods to reorder/split/select rows in a dataset: - `dataset.select(indices)`: Create a new dataset with rows selected following the list/array of indices (which can have a different size than the dataset and contain duplicated indices, the only constrain is that all the integers in the list must be sm...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/266/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/266/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/266.diff", "html_url": "https://github.com/huggingface/datasets/pull/266", "merged_at": "2020-06-18T16:23:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/266.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/266...
2020-06-11T16:22:20Z
2020-06-18T16:23:24Z
7 days, 0:01:04
true
https://api.github.com/repos/huggingface/datasets/issues/263
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/263/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/263/comments
https://api.github.com/repos/huggingface/datasets/issues/263/events
https://github.com/huggingface/datasets/issues/263
637,028,015
MDU6SXNzdWU2MzcwMjgwMTU=
263
[Feature request] Support for external modality for language datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/1479733?v=4", "events_url": "https://api.github.com/users/aleSuglia/events{/privacy}", "followers_url": "https://api.github.com/users/aleSuglia/followers", "following_url": "https://api.github.com/users/aleSuglia/following{/other_user}", "gists_url": "h...
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "c5def5", "default": fals...
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Thanks a lot, @aleSuglia for the very detailed and introductive feature request.\r\nIt seems like we could build something pretty useful here indeed.\r\n\r\nOne of the questions here is that Arrow doesn't have built-in support for generic \"tensors\" in records but...
2020-06-11T13:42:18Z
2022-02-10T13:26:35Z
2022-02-10T13:26:35Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
# Background In recent years many researchers have advocated that learning meanings from text-based only datasets is just like asking a human to "learn to speak by listening to the radio" [[E. Bender and A. Koller,2020](https://openreview.net/forum?id=GKTvAcb12b), [Y. Bisk et. al, 2020](https://arxiv.org/abs/2004.10...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",...
{ "+1": 18, "-1": 0, "confused": 0, "eyes": 4, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 23, "url": "https://api.github.com/repos/huggingface/datasets/issues/263/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/263/timeline
null
completed
null
null
2020-06-11T13:42:18Z
2022-02-10T13:26:35Z
608 days, 23:44:17
false
https://api.github.com/repos/huggingface/datasets/issues/265
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/265/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/265/comments
https://api.github.com/repos/huggingface/datasets/issues/265/events
https://github.com/huggingface/datasets/pull/265
637,139,220
MDExOlB1bGxSZXF1ZXN0NDMzMTgxNDMz
265
Add pyarrow warning colab
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[]
2020-06-11T15:57:51Z
2020-08-02T18:14:36Z
2020-06-12T08:14:16Z
MEMBER
null
null
null
When a user installs `nlp` on google colab, then google colab doesn't update pyarrow, and the runtime needs to be restarted to use the updated version of pyarrow. This is an issue because `nlp` requires the updated version to work correctly. In this PR I added en error that is shown to the user in google colab if...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/265/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/265/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/265.diff", "html_url": "https://github.com/huggingface/datasets/pull/265", "merged_at": "2020-06-12T08:14:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/265.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/265...
2020-06-11T15:57:51Z
2020-06-12T08:14:16Z
16:16:25
true
https://api.github.com/repos/huggingface/datasets/issues/264
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/264/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/264/comments
https://api.github.com/repos/huggingface/datasets/issues/264/events
https://github.com/huggingface/datasets/pull/264
637,106,170
MDExOlB1bGxSZXF1ZXN0NDMzMTU0ODQ4
264
Fix small issues creating dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[]
2020-06-11T15:20:16Z
2020-06-12T08:15:57Z
2020-06-12T08:15:56Z
MEMBER
null
null
null
Fix many small issues mentioned in #249: - don't force to install apache beam for commands - fix None cache dir when using `dl_manager.download_custom` - added new extras in `setup.py` named `dev` that contains tests and quality dependencies - mock dataset sizes when running tests with dummy data - add a note abou...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/264/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/264/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/264.diff", "html_url": "https://github.com/huggingface/datasets/pull/264", "merged_at": "2020-06-12T08:15:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/264.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/264...
2020-06-11T15:20:16Z
2020-06-12T08:15:56Z
16:55:40
true
https://api.github.com/repos/huggingface/datasets/issues/262
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/262/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/262/comments
https://api.github.com/repos/huggingface/datasets/issues/262/events
https://github.com/huggingface/datasets/pull/262
636,702,849
MDExOlB1bGxSZXF1ZXN0NDMyODI3Mzcz
262
Add new dataset ANLI Round 1
{ "avatar_url": "https://avatars.githubusercontent.com/u/11016329?v=4", "events_url": "https://api.github.com/users/easonnie/events{/privacy}", "followers_url": "https://api.github.com/users/easonnie/followers", "following_url": "https://api.github.com/users/easonnie/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Hello ! Thanks for adding this one :)\r\n\r\nThis looks great, you just have to do the last steps to make the CI pass.\r\nI can see that two things are missing:\r\n1. the dummy data that is used to test that the script is working as expected\r\n2. the json file wit...
2020-06-11T04:14:57Z
2020-06-12T22:03:03Z
2020-06-12T22:03:03Z
CONTRIBUTOR
null
null
null
Adding new dataset [ANLI](https://github.com/facebookresearch/anli/). I'm not familiar with how to add new dataset. Let me know if there is any issue. I only include round 1 data here. There will be round 2, round 3 and more in the future with potentially different format. I think it will be better to separate them.
{ "avatar_url": "https://avatars.githubusercontent.com/u/11016329?v=4", "events_url": "https://api.github.com/users/easonnie/events{/privacy}", "followers_url": "https://api.github.com/users/easonnie/followers", "following_url": "https://api.github.com/users/easonnie/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/262/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/262/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/262.diff", "html_url": "https://github.com/huggingface/datasets/pull/262", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/262.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/262" }
2020-06-11T04:14:57Z
2020-06-12T22:03:03Z
1 day, 17:48:06
true
https://api.github.com/repos/huggingface/datasets/issues/261
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/261/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/261/comments
https://api.github.com/repos/huggingface/datasets/issues/261/events
https://github.com/huggingface/datasets/issues/261
636,372,380
MDU6SXNzdWU2MzYzNzIzODA=
261
Downloading dataset error with pyarrow.lib.RecordBatch
{ "avatar_url": "https://avatars.githubusercontent.com/u/5248968?v=4", "events_url": "https://api.github.com/users/cuent/events{/privacy}", "followers_url": "https://api.github.com/users/cuent/followers", "following_url": "https://api.github.com/users/cuent/following{/other_user}", "gists_url": "https://api.g...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "When you install `nlp` for the first time on a Colab runtime, it updates the `pyarrow` library that was already on colab. This update shows this message on colab:\r\n```\r\nWARNING: The following packages were previously imported in this runtime:\r\n [pyarrow]\r\n...
2020-06-10T16:04:19Z
2020-06-11T14:35:12Z
2020-06-11T14:35:12Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I am trying to download `sentiment140` and I have the following error ``` /usr/local/lib/python3.6/dist-packages/nlp/load.py in load_dataset(path, name, version, data_dir, data_files, split, cache_dir, download_config, download_mode, ignore_verifications, save_infos, **config_kwargs) 518 download_mode=...
{ "avatar_url": "https://avatars.githubusercontent.com/u/5248968?v=4", "events_url": "https://api.github.com/users/cuent/events{/privacy}", "followers_url": "https://api.github.com/users/cuent/followers", "following_url": "https://api.github.com/users/cuent/following{/other_user}", "gists_url": "https://api.g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/261/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/261/timeline
null
completed
null
null
2020-06-10T16:04:19Z
2020-06-11T14:35:12Z
22:30:53
false
https://api.github.com/repos/huggingface/datasets/issues/259
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/259/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/259/comments
https://api.github.com/repos/huggingface/datasets/issues/259/events
https://github.com/huggingface/datasets/issues/259
636,239,529
MDU6SXNzdWU2MzYyMzk1Mjk=
259
documentation missing how to split a dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/2873355?v=4", "events_url": "https://api.github.com/users/fotisj/events{/privacy}", "followers_url": "https://api.github.com/users/fotisj/followers", "following_url": "https://api.github.com/users/fotisj/following{/other_user}", "gists_url": "https://ap...
[]
closed
false
null
[]
null
[ { "author_association": "NONE", "body": "this seems to work for my specific problem:\r\n\r\n`self.train_ds, self.test_ds, self.val_ds = map(_prepare_ds, ('train', 'test[:25%]+test[50%:75%]', 'test[75%:]'))`", "created_at": "2020-06-10T22:17:12Z", "html_url": "https://github.com/huggingface/datasets/...
2020-06-10T13:18:13Z
2023-03-14T13:56:07Z
2020-06-18T22:20:24Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I am trying to understand how to split a dataset ( as arrow_dataset). I know I can do something like this to access a split which is already in the original dataset : `ds_test = nlp.load_dataset('imdb, split='test') ` But how can I split ds_test into a test and a validation set (without reading the data into m...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/259/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/259/timeline
null
completed
null
null
2020-06-10T13:18:13Z
2020-06-18T22:20:24Z
8 days, 9:02:11
false
https://api.github.com/repos/huggingface/datasets/issues/260
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/260/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/260/comments
https://api.github.com/repos/huggingface/datasets/issues/260/events
https://github.com/huggingface/datasets/pull/260
636,261,118
MDExOlB1bGxSZXF1ZXN0NDMyNDY3NDM5
260
Consistency fixes
{ "avatar_url": "https://avatars.githubusercontent.com/u/326577?v=4", "events_url": "https://api.github.com/users/julien-c/events{/privacy}", "followers_url": "https://api.github.com/users/julien-c/followers", "following_url": "https://api.github.com/users/julien-c/following{/other_user}", "gists_url": "https...
[]
closed
false
null
[]
null
[]
2020-06-10T13:44:42Z
2020-06-11T10:34:37Z
2020-06-11T10:34:36Z
MEMBER
null
null
null
A few bugs I've found while hacking
{ "avatar_url": "https://avatars.githubusercontent.com/u/326577?v=4", "events_url": "https://api.github.com/users/julien-c/events{/privacy}", "followers_url": "https://api.github.com/users/julien-c/followers", "following_url": "https://api.github.com/users/julien-c/following{/other_user}", "gists_url": "https...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/260/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/260/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/260.diff", "html_url": "https://github.com/huggingface/datasets/pull/260", "merged_at": "2020-06-11T10:34:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/260.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/260...
2020-06-10T13:44:42Z
2020-06-11T10:34:36Z
20:49:54
true
https://api.github.com/repos/huggingface/datasets/issues/258
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/258/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/258/comments
https://api.github.com/repos/huggingface/datasets/issues/258/events
https://github.com/huggingface/datasets/issues/258
635,859,525
MDU6SXNzdWU2MzU4NTk1MjU=
258
Why is dataset after tokenization far more larger than the orginal one ?
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Hi ! This is because `.map` added the new column `input_ids` to the dataset, and so all the other columns were kept. Therefore the dataset size increased a lot.\r\n If you want to only keep the `input_ids` column, you can stash the other ones by specifying `remove_...
2020-06-10T01:27:07Z
2020-06-10T12:46:34Z
2020-06-10T12:46:34Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I tokenize wiki dataset by `map` and cache the results. ``` def tokenize_tfm(example): example['input_ids'] = hf_fast_tokenizer.convert_tokens_to_ids(hf_fast_tokenizer.tokenize(example['text'])) return example wiki = nlp.load_dataset('wikipedia', '20200501.en', cache_dir=cache_dir)['train'] wiki.map(token...
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/258/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/258/timeline
null
completed
null
null
2020-06-10T01:27:07Z
2020-06-10T12:46:34Z
11:19:27
false
https://api.github.com/repos/huggingface/datasets/issues/257
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/257/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/257/comments
https://api.github.com/repos/huggingface/datasets/issues/257/events
https://github.com/huggingface/datasets/issues/257
635,620,979
MDU6SXNzdWU2MzU2MjA5Nzk=
257
Tokenizer pickling issue fix not landed in `nlp` yet?
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Yes, the new release of tokenizers solves this and should be out soon.\r\nIn the meantime, you can install it with `pip install tokenizers==0.8.0-dev2`", "created_at": "2020-06-09T17:26:23Z", "html_url": "https://github.com/huggingface/datasets/issues/257#i...
2020-06-09T17:12:34Z
2020-06-10T21:45:32Z
2020-06-09T17:26:53Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Unless I recreate an arrow_dataset from my loaded nlp dataset myself (which I think does not use the cache by default), I get the following error when applying the map function: ``` dataset = nlp.load_dataset('cos_e') tokenizer = GPT2TokenizerFast.from_pretrained('gpt2', cache_dir=cache_dir) for split in datase...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/257/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/257/timeline
null
completed
null
null
2020-06-09T17:12:34Z
2020-06-09T17:26:53Z
0:14:19
false
https://api.github.com/repos/huggingface/datasets/issues/254
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/254/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/254/comments
https://api.github.com/repos/huggingface/datasets/issues/254/events
https://github.com/huggingface/datasets/issues/254
635,057,568
MDU6SXNzdWU2MzUwNTc1Njg=
254
[Feature request] Be able to remove a specific sample of the dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Oh yes you can now do that with the `dataset.filter()` method that was added in #214 ", "created_at": "2020-06-09T08:22:34Z", "html_url": "https://github.com/huggingface/datasets/issues/254#issuecomment-641118305", "id": 641118305, "issue_url": "htt...
2020-06-09T02:22:13Z
2020-06-09T08:41:38Z
2020-06-09T08:41:38Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
As mentioned in #117, it's currently not possible to remove a sample of the dataset. But it is a important use case : After applying some preprocessing, some samples might be empty for example. We should be able to remove these samples from the dataset, or at least mark them as `removed` so when iterating the datase...
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/254/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/254/timeline
null
completed
null
null
2020-06-09T02:22:13Z
2020-06-09T08:41:38Z
6:19:25
false
https://api.github.com/repos/huggingface/datasets/issues/255
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/255/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/255/comments
https://api.github.com/repos/huggingface/datasets/issues/255/events
https://github.com/huggingface/datasets/pull/255
635,300,822
MDExOlB1bGxSZXF1ZXN0NDMxNjg3MDM0
255
Add dataset/piaf
{ "avatar_url": "https://avatars.githubusercontent.com/u/36986299?v=4", "events_url": "https://api.github.com/users/RachelKer/events{/privacy}", "followers_url": "https://api.github.com/users/RachelKer/followers", "following_url": "https://api.github.com/users/RachelKer/following{/other_user}", "gists_url": "...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Very nice !", "created_at": "2020-06-11T08:07:08Z", "html_url": "https://github.com/huggingface/datasets/pull/255#issuecomment-642485075", "id": 642485075, "issue_url": "https://api.github.com/repos/huggingface/datasets/issues/255", "node_id": "...
2020-06-09T10:16:01Z
2020-06-12T08:31:27Z
2020-06-12T08:31:27Z
CONTRIBUTOR
null
null
null
Small SQuAD-like French QA dataset [PIAF](https://www.aclweb.org/anthology/2020.lrec-1.673.pdf)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/255/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/255/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/255.diff", "html_url": "https://github.com/huggingface/datasets/pull/255", "merged_at": "2020-06-12T08:31:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/255.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/255...
2020-06-09T10:16:01Z
2020-06-12T08:31:27Z
2 days, 22:15:26
true
https://api.github.com/repos/huggingface/datasets/issues/253
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/253/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/253/comments
https://api.github.com/repos/huggingface/datasets/issues/253/events
https://github.com/huggingface/datasets/pull/253
634,791,939
MDExOlB1bGxSZXF1ZXN0NDMxMjgwOTYz
253
add flue dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "The dummy data file was wrong. I only fixed it for the book config. Even though the tests are all green here, this should also be fixed for all other configs. Could you take a look there @mariamabarham ? ", "created_at": "2020-06-08T18:00:18Z", "html_u...
2020-06-08T17:11:09Z
2023-09-24T09:46:03Z
2020-07-16T07:50:59Z
CONTRIBUTOR
null
null
null
This PR add the Flue dataset as requested in this issue #223 . @lbourdois made a detailed description in that issue.
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/253/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/253/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/253.diff", "html_url": "https://github.com/huggingface/datasets/pull/253", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/253.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/253" }
2020-06-08T17:11:09Z
2020-07-16T07:50:59Z
37 days, 14:39:50
true
https://api.github.com/repos/huggingface/datasets/issues/256
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/256/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/256/comments
https://api.github.com/repos/huggingface/datasets/issues/256/events
https://github.com/huggingface/datasets/issues/256
635,596,295
MDU6SXNzdWU2MzU1OTYyOTU=
256
[Feature request] Add a feature to dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Do you have an example of what you would like to do? (you can just add a field in the output of the unction you give to map and this will add this field in the output table)", "created_at": "2020-06-09T16:40:21Z", "html_url": "https://github.com/huggingface...
2020-06-09T16:38:12Z
2020-06-09T16:51:42Z
2020-06-09T16:51:42Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Is there a straightforward way to add a field to the arrow_dataset, prior to performing map?
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/256/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/256/timeline
null
completed
null
null
2020-06-09T16:38:12Z
2020-06-09T16:51:42Z
0:13:30
false
https://api.github.com/repos/huggingface/datasets/issues/252
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/252/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/252/comments
https://api.github.com/repos/huggingface/datasets/issues/252/events
https://github.com/huggingface/datasets/issues/252
634,563,239
MDU6SXNzdWU2MzQ1NjMyMzk=
252
NonMatchingSplitsSizesError error when reading the IMDB dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/17463361?v=4", "events_url": "https://api.github.com/users/antmarakis/events{/privacy}", "followers_url": "https://api.github.com/users/antmarakis/followers", "following_url": "https://api.github.com/users/antmarakis/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "I just tried on my side and I didn't encounter your problem.\r\nApparently the script doesn't generate all the examples on your side.\r\n\r\nCan you provide the version of `nlp` you're using ?\r\nCan you try to clear your cache and re-run the code ?", "created_...
2020-06-08T12:26:24Z
2021-08-27T15:20:58Z
2020-06-08T14:01:26Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi! I am trying to load the `imdb` dataset with this line: `dataset = nlp.load_dataset('imdb', data_dir='/A/PATH', cache_dir='/A/PATH')` but I am getting the following error: ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/mounts/Users/cisintern/antmarakis/anaconda3/...
{ "avatar_url": "https://avatars.githubusercontent.com/u/17463361?v=4", "events_url": "https://api.github.com/users/antmarakis/events{/privacy}", "followers_url": "https://api.github.com/users/antmarakis/followers", "following_url": "https://api.github.com/users/antmarakis/following{/other_user}", "gists_url"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/252/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/252/timeline
null
completed
null
null
2020-06-08T12:26:24Z
2020-06-08T14:01:26Z
1:35:02
false
https://api.github.com/repos/huggingface/datasets/issues/251
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/251/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/251/comments
https://api.github.com/repos/huggingface/datasets/issues/251/events
https://github.com/huggingface/datasets/pull/251
634,544,977
MDExOlB1bGxSZXF1ZXN0NDMxMDgwMDkw
251
Better access to all dataset information
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
null
[]
null
[]
2020-06-08T11:56:50Z
2020-06-12T08:13:00Z
2020-06-12T08:12:58Z
MEMBER
null
null
null
Moves all the dataset info down one level from `dataset.info.XXX` to `dataset.XXX` This way it's easier to access `dataset.feature['label']` for instance Also, add the original split instructions used to create the dataset in `dataset.split` Ex: ``` from nlp import load_dataset stsb = load_dataset('glue', name=...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/251/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/251/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/251.diff", "html_url": "https://github.com/huggingface/datasets/pull/251", "merged_at": "2020-06-12T08:12:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/251.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/251...
2020-06-08T11:56:50Z
2020-06-12T08:12:58Z
3 days, 20:16:08
true
https://api.github.com/repos/huggingface/datasets/issues/250
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/250/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/250/comments
https://api.github.com/repos/huggingface/datasets/issues/250/events
https://github.com/huggingface/datasets/pull/250
634,416,751
MDExOlB1bGxSZXF1ZXN0NDMwOTcyMzg4
250
Remove checksum download in c4
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[ { "author_association": "NONE", "body": "Commenting again in case [previous thread](https://github.com/huggingface/nlp/pull/233) was inactive.\r\n\r\n@lhoestq I am facing `IsADirectoryError` while downloading with this command.\r\nCan you pls look into it & help me.\r\nI'm using version 0.4.0 of `nlp`.\r\n\...
2020-06-08T09:13:00Z
2020-08-25T07:04:56Z
2020-06-08T09:16:59Z
MEMBER
null
null
null
There was a line from the original tfds script that was still there and causing issues when loading the c4 script. This one should fix #233 and allow anyone to load the c4 script to generate the dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/250/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/250/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/250.diff", "html_url": "https://github.com/huggingface/datasets/pull/250", "merged_at": "2020-06-08T09:16:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/250.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/250...
2020-06-08T09:13:00Z
2020-06-08T09:16:59Z
0:03:59
true
https://api.github.com/repos/huggingface/datasets/issues/249
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/249/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/249/comments
https://api.github.com/repos/huggingface/datasets/issues/249/events
https://github.com/huggingface/datasets/issues/249
633,393,443
MDU6SXNzdWU2MzMzOTM0NDM=
249
[Dataset created] some critical small issues when I was creating a dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
[ { "author_association": "MEMBER", "body": "Thanks for noticing all these :) They should be easy to fix indeed", "created_at": "2020-06-08T08:29:20Z", "html_url": "https://github.com/huggingface/datasets/issues/249#issuecomment-640452194", "id": 640452194, "issue_url": "https://api.github.com...
2020-06-07T12:58:54Z
2020-06-12T08:28:51Z
2020-06-12T08:28:51Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi, I successfully created a dataset and has made a pr #248. But I have encountered several problems when I was creating it, and those should be easy to fix. 1. Not found dataset_info.json should be fixed by #241 , eager to wait it be merged. 2. Forced to install `apach_beam` If we should install it, then it m...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/249/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/249/timeline
null
completed
null
null
2020-06-07T12:58:54Z
2020-06-12T08:28:51Z
4 days, 19:29:57
false
https://api.github.com/repos/huggingface/datasets/issues/248
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/248/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/248/comments
https://api.github.com/repos/huggingface/datasets/issues/248/events
https://github.com/huggingface/datasets/pull/248
633,390,427
MDExOlB1bGxSZXF1ZXN0NDMwMDQ0MzU0
248
add Toronto BooksCorpus
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Thanks for adding this one !\r\n\r\nAbout the three points you mentioned:\r\n1. I think the `toronto_books_corpus` branch can be removed @mariamabarham ? \r\n2. You can use the download manager to download from google drive. For you case you can just do something l...
2020-06-07T12:54:56Z
2020-06-12T08:45:03Z
2020-06-12T08:45:02Z
CONTRIBUTOR
null
null
null
1. I knew there is a branch `toronto_books_corpus` - After I downloaded it, I found it is all non-english, and only have one row. - It seems that it cites the wrong paper - according to papar using it, it is called `BooksCorpus` but not `TornotoBooksCorpus` 2. It use a text mirror in google drive - `bookscorpu...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/248/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/248/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/248.diff", "html_url": "https://github.com/huggingface/datasets/pull/248", "merged_at": "2020-06-12T08:45:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/248.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/248...
2020-06-07T12:54:56Z
2020-06-12T08:45:02Z
4 days, 19:50:06
true
https://api.github.com/repos/huggingface/datasets/issues/247
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/247/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/247/comments
https://api.github.com/repos/huggingface/datasets/issues/247/events
https://github.com/huggingface/datasets/pull/247
632,380,078
MDExOlB1bGxSZXF1ZXN0NDI5MTMwMzQ2
247
Make all dataset downloads deterministic by applying `sorted` to glob and os.listdir
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "That's great!\r\n\r\nI think it would be nice to test \"deterministic-ness\" of datasets in CI if we can do it (should be left for future PR of course)\r\n\r\nHere is a possibility (maybe there are other ways to do it): given that we may soon have efficient and lar...
2020-06-06T11:02:10Z
2020-06-08T09:18:16Z
2020-06-08T09:18:14Z
CONTRIBUTOR
null
null
null
This PR makes all datasets loading deterministic by applying `sorted()` to all `glob.glob` and `os.listdir` statements. Are there other "non-deterministic" functions apart from `glob.glob()` and `os.listdir()` that you can think of @thomwolf @lhoestq @mariamabarham @jplu ? **Important** It does break backward c...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/247/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/247/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/247.diff", "html_url": "https://github.com/huggingface/datasets/pull/247", "merged_at": "2020-06-08T09:18:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/247.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/247...
2020-06-06T11:02:10Z
2020-06-08T09:18:14Z
1 day, 22:16:04
true
https://api.github.com/repos/huggingface/datasets/issues/246
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/246/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/246/comments
https://api.github.com/repos/huggingface/datasets/issues/246/events
https://github.com/huggingface/datasets/issues/246
632,380,054
MDU6SXNzdWU2MzIzODAwNTQ=
246
What is the best way to cache a dataset?
{ "avatar_url": "https://avatars.githubusercontent.com/u/112599?v=4", "events_url": "https://api.github.com/users/Mistobaan/events{/privacy}", "followers_url": "https://api.github.com/users/Mistobaan/followers", "following_url": "https://api.github.com/users/Mistobaan/following{/other_user}", "gists_url": "ht...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Everything is already cached by default in 🤗nlp (in particular dataset\nloading and all the “map()” operations) so I don’t think you need to do any\nspecific caching in streamlit.\n\nTell us if you feel like it’s not the case.\n\nOn Sat, 6 Jun 2020 at 13:02, Fabri...
2020-06-06T11:02:07Z
2020-07-09T09:15:07Z
2020-07-09T09:15:07Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
For example if I want to use streamlit with a nlp dataset: ``` @st.cache def load_data(): return nlp.load_dataset('squad') ``` This code raises the error "uncachable object" Right now I just fixed with a constant for my specific case: ``` @st.cache(hash_funcs={pyarrow.lib.Buffer: lambda b: 0}) ```...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/246/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/246/timeline
null
completed
null
null
2020-06-06T11:02:07Z
2020-07-09T09:15:07Z
32 days, 22:13:00
false
https://api.github.com/repos/huggingface/datasets/issues/245
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/245/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/245/comments
https://api.github.com/repos/huggingface/datasets/issues/245/events
https://github.com/huggingface/datasets/issues/245
631,985,108
MDU6SXNzdWU2MzE5ODUxMDg=
245
SST-2 test labels are all -1
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "this also happened to me with `nlp.load_dataset('glue', 'mnli')`", "created_at": "2020-06-06T16:16:27Z", "html_url": "https://github.com/huggingface/datasets/issues/245#issuecomment-640083980", "id": 640083980, "issue_url": "https://api.github....
2020-06-05T21:41:42Z
2021-12-08T00:47:32Z
2020-06-06T16:56:41Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
I'm trying to test a model on the SST-2 task, but all the labels I see in the test set are -1. ``` >>> import nlp >>> glue = nlp.load_dataset('glue', 'sst2') >>> glue {'train': Dataset(schema: {'sentence': 'string', 'label': 'int64', 'idx': 'int32'}, num_rows: 67349), 'validation': Dataset(schema: {'sentence': 'st...
{ "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "events_url": "https://api.github.com/users/jxmorris12/events{/privacy}", "followers_url": "https://api.github.com/users/jxmorris12/followers", "following_url": "https://api.github.com/users/jxmorris12/following{/other_user}", "gists_url"...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/245/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/245/timeline
null
completed
null
null
2020-06-05T21:41:42Z
2020-06-06T16:56:41Z
19:14:59
false
https://api.github.com/repos/huggingface/datasets/issues/244
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/244/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/244/comments
https://api.github.com/repos/huggingface/datasets/issues/244/events
https://github.com/huggingface/datasets/pull/244
631,869,155
MDExOlB1bGxSZXF1ZXN0NDI4NjgxMTcx
244
Add Allociné Dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/37028092?v=4", "events_url": "https://api.github.com/users/TheophileBlard/events{/privacy}", "followers_url": "https://api.github.com/users/TheophileBlard/followers", "following_url": "https://api.github.com/users/TheophileBlard/following{/other_user}", ...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "great work @TheophileBlard ", "created_at": "2020-06-05T19:29:29Z", "html_url": "https://github.com/huggingface/datasets/pull/244#issuecomment-639744174", "id": 639744174, "issue_url": "https://api.github.com/repos/huggingface/datasets/issues/2...
2020-06-05T19:19:26Z
2020-06-11T07:47:26Z
2020-06-11T07:47:26Z
CONTRIBUTOR
null
null
null
This is a french binary sentiment classification dataset, which was used to train this model: https://huggingface.co/tblard/tf-allocine. Basically, it's a french "IMDB" dataset, with more reviews. More info on [this repo](https://github.com/TheophileBlard/french-sentiment-analysis-with-bert).
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/244/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/244/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/244.diff", "html_url": "https://github.com/huggingface/datasets/pull/244", "merged_at": "2020-06-11T07:47:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/244.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/244...
2020-06-05T19:19:26Z
2020-06-11T07:47:26Z
5 days, 12:28:00
true
https://api.github.com/repos/huggingface/datasets/issues/243
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/243/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/243/comments
https://api.github.com/repos/huggingface/datasets/issues/243/events
https://github.com/huggingface/datasets/pull/243
631,735,848
MDExOlB1bGxSZXF1ZXN0NDI4NTY2MTEy
243
Specify utf-8 encoding for GLUE
{ "avatar_url": "https://avatars.githubusercontent.com/u/15801338?v=4", "events_url": "https://api.github.com/users/patpizio/events{/privacy}", "followers_url": "https://api.github.com/users/patpizio/followers", "following_url": "https://api.github.com/users/patpizio/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Thanks for fixing the encoding :)", "created_at": "2020-06-08T08:41:40Z", "html_url": "https://github.com/huggingface/datasets/pull/243#issuecomment-640458977", "id": 640458977, "issue_url": "https://api.github.com/repos/huggingface/datasets/issues/...
2020-06-05T16:33:00Z
2020-06-17T21:16:06Z
2020-06-08T08:42:01Z
CONTRIBUTOR
null
null
null
#242 This makes the GLUE-MNLI dataset readable on my machine, not sure if it's a Windows-only bug.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/243/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/243/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/243.diff", "html_url": "https://github.com/huggingface/datasets/pull/243", "merged_at": "2020-06-08T08:42:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/243.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/243...
2020-06-05T16:33:00Z
2020-06-08T08:42:01Z
2 days, 16:09:01
true
https://api.github.com/repos/huggingface/datasets/issues/242
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/242/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/242/comments
https://api.github.com/repos/huggingface/datasets/issues/242/events
https://github.com/huggingface/datasets/issues/242
631,733,683
MDU6SXNzdWU2MzE3MzM2ODM=
242
UnicodeDecodeError when downloading GLUE-MNLI
{ "avatar_url": "https://avatars.githubusercontent.com/u/15801338?v=4", "events_url": "https://api.github.com/users/patpizio/events{/privacy}", "followers_url": "https://api.github.com/users/patpizio/followers", "following_url": "https://api.github.com/users/patpizio/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "It should be good now, thanks for noticing and fixing it ! I would say that it was because you are on windows but not 100% sure", "created_at": "2020-06-08T08:45:03Z", "html_url": "https://github.com/huggingface/datasets/issues/242#issuecomment-640460943", ...
2020-06-05T16:30:01Z
2020-06-09T16:06:47Z
2020-06-08T08:45:03Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
When I run ```python dataset = nlp.load_dataset('glue', 'mnli') ``` I get an encoding error (could it be because I'm using Windows?) : ```python # Lots of error log lines later... ~\Miniconda3\envs\nlp\lib\site-packages\tqdm\std.py in __iter__(self) 1128 try: -> 1129 for obj in iterable:...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/242/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/242/timeline
null
completed
null
null
2020-06-05T16:30:01Z
2020-06-08T08:45:03Z
2 days, 16:15:02
false
https://api.github.com/repos/huggingface/datasets/issues/241
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/241/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/241/comments
https://api.github.com/repos/huggingface/datasets/issues/241/events
https://github.com/huggingface/datasets/pull/241
631,703,079
MDExOlB1bGxSZXF1ZXN0NDI4NTQwMDM0
241
Fix empty cache dir
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "Looks great! Will this change force all cached datasets to be redownloaded? But even if it does, it shoud not be a big problem, I think", "created_at": "2020-06-05T19:04:32Z", "html_url": "https://github.com/huggingface/datasets/pull/241#issuecomment-6...
2020-06-05T15:45:22Z
2020-06-08T08:35:33Z
2020-06-08T08:35:31Z
MEMBER
null
null
null
If the cache dir of a dataset is empty, the dataset fails to load and throws a FileNotFounfError. We could end up with empty cache dir because there was a line in the code that created the cache dir without using a temp dir. Using a temp dir is useful as it gets renamed to the real cache dir only if the full process is...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/241/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/241/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/241.diff", "html_url": "https://github.com/huggingface/datasets/pull/241", "merged_at": "2020-06-08T08:35:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/241.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/241...
2020-06-05T15:45:22Z
2020-06-08T08:35:31Z
2 days, 16:50:09
true
https://api.github.com/repos/huggingface/datasets/issues/240
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/240/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/240/comments
https://api.github.com/repos/huggingface/datasets/issues/240/events
https://github.com/huggingface/datasets/issues/240
631,434,677
MDU6SXNzdWU2MzE0MzQ2Nzc=
240
Deterministic dataset loading
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "Yes good point !", "created_at": "2020-06-05T09:06:32Z", "html_url": "https://github.com/huggingface/datasets/issues/240#issuecomment-639356179", "id": 639356179, "issue_url": "https://api.github.com/repos/huggingface/datasets/issues/240", "node...
2020-06-05T09:03:26Z
2020-06-08T09:18:14Z
2020-06-08T09:18:14Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
When calling: ```python import nlp dataset = nlp.load_dataset("trivia_qa", split="validation[:1%]") ``` the resulting dataset is not deterministic over different google colabs. After talking to @thomwolf, I suspect the reason to be the use of `glob.glob` in line: https://github.com/huggingface/nlp/blob/2e0...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/240/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/240/timeline
null
completed
null
null
2020-06-05T09:03:26Z
2020-06-08T09:18:14Z
3 days, 0:14:48
false
https://api.github.com/repos/huggingface/datasets/issues/239
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/239/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/239/comments
https://api.github.com/repos/huggingface/datasets/issues/239/events
https://github.com/huggingface/datasets/issues/239
631,340,440
MDU6SXNzdWU2MzEzNDA0NDA=
239
[Creating new dataset] Not found dataset_info.json
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
[ { "author_association": "MEMBER", "body": "I think you can just `rm` this directory and it should be good :)", "created_at": "2020-06-05T07:31:50Z", "html_url": "https://github.com/huggingface/datasets/issues/239#issuecomment-639310157", "id": 639310157, "issue_url": "https://api.github.com/...
2020-06-05T06:15:04Z
2020-06-07T13:01:04Z
2020-06-07T13:01:04Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
Hi, I am trying to create Toronto Book Corpus. #131 I ran `~/nlp % python nlp-cli test datasets/bookcorpus --save_infos --all_configs` but this doesn't create `dataset_info.json` and try to use it ``` INFO:nlp.load:Checking datasets/bookcorpus/bookcorpus.py for additional imports. INFO:filelock:Lock 1397953257...
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/239/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/239/timeline
null
completed
null
null
2020-06-05T06:15:04Z
2020-06-07T13:01:04Z
2 days, 6:46:00
false
https://api.github.com/repos/huggingface/datasets/issues/238
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/238/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/238/comments
https://api.github.com/repos/huggingface/datasets/issues/238/events
https://github.com/huggingface/datasets/issues/238
631,260,143
MDU6SXNzdWU2MzEyNjAxNDM=
238
[Metric] Bertscore : Warning : Empty candidate sentence; Setting recall to be 0.
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[ { "color": "25b21e", "default": false, "description": "A bug in a metric script", "id": 2067393914, "name": "metric bug", "node_id": "MDU6TGFiZWwyMDY3MzkzOTE0", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug" } ]
closed
false
null
[]
null
[ { "author_association": "MEMBER", "body": "This print statement comes from the official implementation of bert_score (see [here](https://github.com/Tiiiger/bert_score/blob/master/bert_score/utils.py#L343)). The warning shows up only if the attention mask outputs no candidate.\r\nRight now we want to only us...
2020-06-05T02:14:47Z
2020-06-29T17:10:19Z
2020-06-29T17:10:19Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
When running BERT-Score, I'm meeting this warning : > Warning: Empty candidate sentence; Setting recall to be 0. Code : ``` import nlp metric = nlp.load_metric("bertscore") scores = metric.compute(["swag", "swags"], ["swags", "totally something different"], lang="en", device=0) ``` --- **What am I do...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/238/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/238/timeline
null
completed
null
null
2020-06-05T02:14:47Z
2020-06-29T17:10:19Z
24 days, 14:55:32
false
https://api.github.com/repos/huggingface/datasets/issues/237
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/237/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/237/comments
https://api.github.com/repos/huggingface/datasets/issues/237/events
https://github.com/huggingface/datasets/issues/237
631,199,940
MDU6SXNzdWU2MzExOTk5NDA=
237
Can't download MultiNLI
{ "avatar_url": "https://avatars.githubusercontent.com/u/15801338?v=4", "events_url": "https://api.github.com/users/patpizio/events{/privacy}", "followers_url": "https://api.github.com/users/patpizio/followers", "following_url": "https://api.github.com/users/patpizio/following{/other_user}", "gists_url": "htt...
[]
closed
false
null
[]
null
[ { "author_association": "CONTRIBUTOR", "body": "You should use `load_dataset('glue', 'mnli')`", "created_at": "2020-06-05T01:46:09Z", "html_url": "https://github.com/huggingface/datasets/issues/237#issuecomment-639209008", "id": 639209008, "issue_url": "https://api.github.com/repos/huggingfa...
2020-06-04T23:05:21Z
2020-06-06T10:51:34Z
2020-06-06T10:51:34Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
When I try to download MultiNLI with ```python dataset = load_dataset('multi_nli') ``` I get this long error: ```python --------------------------------------------------------------------------- OSError Traceback (most recent call last) <ipython-input-13-3b11f6be4cb9> in <m...
{ "avatar_url": "https://avatars.githubusercontent.com/u/15801338?v=4", "events_url": "https://api.github.com/users/patpizio/events{/privacy}", "followers_url": "https://api.github.com/users/patpizio/followers", "following_url": "https://api.github.com/users/patpizio/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/237/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/237/timeline
null
completed
null
null
2020-06-04T23:05:21Z
2020-06-06T10:51:34Z
1 day, 11:46:13
false