comments_url
stringlengths
70
70
timeline_url
stringlengths
70
70
closed_at
stringlengths
20
20
performed_via_github_app
null
state_reason
stringclasses
3 values
node_id
stringlengths
18
32
state
stringclasses
2 values
assignees
listlengths
0
4
draft
bool
2 classes
number
int64
1.61k
6.73k
user
dict
title
stringlengths
1
290
events_url
stringlengths
68
68
milestone
dict
labels_url
stringlengths
75
75
created_at
stringlengths
20
20
active_lock_reason
null
locked
bool
1 class
assignee
dict
pull_request
dict
id
int64
771M
2.18B
labels
listlengths
0
4
url
stringlengths
61
61
comments
listlengths
0
30
repository_url
stringclasses
1 value
author_association
stringclasses
3 values
body
stringlengths
0
228k
updated_at
stringlengths
20
20
html_url
stringlengths
49
51
reactions
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/3041/comments
https://api.github.com/repos/huggingface/datasets/issues/3041/timeline
2021-10-12T15:25:46Z
null
null
PR_kwDODunzps4s1ZAc
closed
[]
false
3,041
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Load private data files + use glob on ZIP archives for json/csv/etc. module inference
https://api.github.com/repos/huggingface/datasets/issues/3041/events
null
https://api.github.com/repos/huggingface/datasets/issues/3041/labels{/name}
2021-10-06T18:16:36Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3041.diff", "html_url": "https://github.com/huggingface/datasets/pull/3041", "merged_at": "2021-10-12T15:25:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/3041.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3041" }
1,018,911,385
[]
https://api.github.com/repos/huggingface/datasets/issues/3041
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
As mentioned in https://github.com/huggingface/datasets/issues/3032 loading data files from private repository isn't working correctly because of the data files resolved. #2986 did a refactor of the data files resolver. I added authentication to it. I also improved it to glob inside ZIP archives to look for json/csv/etc. files and infer which dataset builder (json/csv/etc.) to use. Fix https://github.com/huggingface/datasets/issues/3032 Note that #2986 needs to get merged first
2021-10-12T15:25:48Z
https://github.com/huggingface/datasets/pull/3041
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/3041/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3040/comments
https://api.github.com/repos/huggingface/datasets/issues/3040/timeline
2021-11-02T15:41:08Z
null
completed
I_kwDODunzps48uWML
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
3,040
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
[save_to_disk] Using `select()` followed by `save_to_disk` saves complete dataset making it hard to create dummy dataset
https://api.github.com/repos/huggingface/datasets/issues/3040/events
null
https://api.github.com/repos/huggingface/datasets/issues/3040/labels{/name}
2021-10-06T17:08:47Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,018,782,475
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/3040
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug When only keeping a dummy size of a dataset (say the first 100 samples), and then saving it to disk to upload it in the following to the hub for easy demo/use - not just the small dataset is saved but the whole dataset with an indices file. The problem with this is that the dataset is still very big. ## Steps to reproduce the bug E.g. run the following: ```python from datasets import load_dataset, save_to_disk nlp = load_dataset("glue", "mnli", split="train") nlp.save_to_disk("full") nlp = nlp.select(range(100)) nlp.save_to_disk("dummy") ``` Now one can see that both `"dummy"` and `"full"` have the same size. This shouldn't be the case IMO. ## Expected results IMO `"dummy"` should be much smaller so that one can easily play around with the dataset on the hub. ## Actual results Specify the actual results or traceback. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.2.dev0 - Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.10 - Python version: 3.8.5 - PyArrow version: 5.0.0
2021-11-02T15:41:08Z
https://github.com/huggingface/datasets/issues/3040
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3040/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3039/comments
https://api.github.com/repos/huggingface/datasets/issues/3039/timeline
2021-10-13T10:16:04Z
null
null
PR_kwDODunzps4sy_J-
closed
[]
false
3,039
{ "avatar_url": "https://avatars.githubusercontent.com/u/13781234?v=4", "events_url": "https://api.github.com/users/Alenush/events{/privacy}", "followers_url": "https://api.github.com/users/Alenush/followers", "following_url": "https://api.github.com/users/Alenush/following{/other_user}", "gists_url": "https://api.github.com/users/Alenush/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Alenush", "id": 13781234, "login": "Alenush", "node_id": "MDQ6VXNlcjEzNzgxMjM0", "organizations_url": "https://api.github.com/users/Alenush/orgs", "received_events_url": "https://api.github.com/users/Alenush/received_events", "repos_url": "https://api.github.com/users/Alenush/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Alenush/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Alenush/subscriptions", "type": "User", "url": "https://api.github.com/users/Alenush" }
Add sberquad dataset
https://api.github.com/repos/huggingface/datasets/issues/3039/events
null
https://api.github.com/repos/huggingface/datasets/issues/3039/labels{/name}
2021-10-06T12:32:02Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3039.diff", "html_url": "https://github.com/huggingface/datasets/pull/3039", "merged_at": "2021-10-13T10:16:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/3039.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3039" }
1,018,219,800
[]
https://api.github.com/repos/huggingface/datasets/issues/3039
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2021-10-13T10:19:11Z
https://github.com/huggingface/datasets/pull/3039
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3039/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3038/comments
https://api.github.com/repos/huggingface/datasets/issues/3038/timeline
2021-10-06T11:58:01Z
null
null
PR_kwDODunzps4syno_
closed
[]
false
3,038
{ "avatar_url": "https://avatars.githubusercontent.com/u/13781234?v=4", "events_url": "https://api.github.com/users/Alenush/events{/privacy}", "followers_url": "https://api.github.com/users/Alenush/followers", "following_url": "https://api.github.com/users/Alenush/following{/other_user}", "gists_url": "https://api.github.com/users/Alenush/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Alenush", "id": 13781234, "login": "Alenush", "node_id": "MDQ6VXNlcjEzNzgxMjM0", "organizations_url": "https://api.github.com/users/Alenush/orgs", "received_events_url": "https://api.github.com/users/Alenush/received_events", "repos_url": "https://api.github.com/users/Alenush/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Alenush/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Alenush/subscriptions", "type": "User", "url": "https://api.github.com/users/Alenush" }
add sberquad dataset
https://api.github.com/repos/huggingface/datasets/issues/3038/events
null
https://api.github.com/repos/huggingface/datasets/issues/3038/labels{/name}
2021-10-06T11:33:39Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3038.diff", "html_url": "https://github.com/huggingface/datasets/pull/3038", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/3038.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3038" }
1,018,113,499
[]
https://api.github.com/repos/huggingface/datasets/issues/3038
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2021-10-06T11:58:01Z
https://github.com/huggingface/datasets/pull/3038
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3038/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3037/comments
https://api.github.com/repos/huggingface/datasets/issues/3037/timeline
2021-10-06T11:33:08Z
null
null
PR_kwDODunzps4syi15
closed
[]
false
3,037
{ "avatar_url": "https://avatars.githubusercontent.com/u/13781234?v=4", "events_url": "https://api.github.com/users/Alenush/events{/privacy}", "followers_url": "https://api.github.com/users/Alenush/followers", "following_url": "https://api.github.com/users/Alenush/following{/other_user}", "gists_url": "https://api.github.com/users/Alenush/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Alenush", "id": 13781234, "login": "Alenush", "node_id": "MDQ6VXNlcjEzNzgxMjM0", "organizations_url": "https://api.github.com/users/Alenush/orgs", "received_events_url": "https://api.github.com/users/Alenush/received_events", "repos_url": "https://api.github.com/users/Alenush/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Alenush/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Alenush/subscriptions", "type": "User", "url": "https://api.github.com/users/Alenush" }
SberQuad
https://api.github.com/repos/huggingface/datasets/issues/3037/events
null
https://api.github.com/repos/huggingface/datasets/issues/3037/labels{/name}
2021-10-06T11:21:08Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3037.diff", "html_url": "https://github.com/huggingface/datasets/pull/3037", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/3037.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3037" }
1,018,091,919
[]
https://api.github.com/repos/huggingface/datasets/issues/3037
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2021-10-06T11:33:08Z
https://github.com/huggingface/datasets/pull/3037
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3037/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3036/comments
https://api.github.com/repos/huggingface/datasets/issues/3036/timeline
2021-10-07T06:49:52Z
null
completed
I_kwDODunzps48qK-I
closed
[]
null
3,036
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Protect master branch to force contributions via Pull Requests
https://api.github.com/repos/huggingface/datasets/issues/3036/events
null
https://api.github.com/repos/huggingface/datasets/issues/3036/labels{/name}
2021-10-06T07:34:17Z
null
false
null
null
1,017,687,944
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/3036
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
In order to have a clearer Git history in the master branch, I propose to protect it so that all contributions must be done through a Pull Request and no direct commits to master are allowed. - The Pull Request allows to give context, discuss any potential issues and improve the quality of the contribution - The Pull Request will eventually be squashed and merged into master with a single commit that links to the Pull Request page (with all the context/discussions) Note that we already implemented a protection in the master branch to avoid *merge* commits and ensure a linear history. This proposal goes one step further by avoiding all kind of direct commits and forcing contributions **only** through Pull Requests. Please note that we can temporarily deactivate this protection if we need to make a direct commit, e.g. at each new version release. The only way GitHub allows this kind or protection is by requiring a minimal number (at least one) of approvals of the Pull Request. The inconvenient is that the PR creator cannot approve their own PR: another person must approve it before it can be merged into master. To circumvent this, we could eventually disable this protection in the master branch when an urgent commit is needed (e.g. for a hotfix) and there is no other person available at that time to approve the PR.
2021-10-07T06:51:47Z
https://github.com/huggingface/datasets/issues/3036
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/3036/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3035/comments
https://api.github.com/repos/huggingface/datasets/issues/3035/timeline
null
null
null
I_kwDODunzps48mq4X
open
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
3,035
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
`load_dataset` does not work with uploaded arrow file
https://api.github.com/repos/huggingface/datasets/issues/3035/events
null
https://api.github.com/repos/huggingface/datasets/issues/3035/labels{/name}
2021-10-05T20:15:10Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,016,770,071
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/3035
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug I've preprocessed and uploaded a dataset here: https://huggingface.co/datasets/ami-wav2vec2/ami_headset_single_preprocessed . The dataset is in `.arrow` format. The dataset can correctly be loaded when doing: ```bash git lfs install git clone https://huggingface.co/datasets/ami-wav2vec2/ami_headset_single_preprocessed ``` followed by ```python from datasets import load_from_disk ds = load_from_disk("./ami_headset_single_preprocessed") ``` However when I try to directly download the dataset as follows: ```python from datasets import load_dataset ds = load_dataset("ami-wav2vec2/ami_headset_single_preprocessed") ``` the following error occurs: ```bash /usr/local/lib/python3.7/dist-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs) 1115 ignore_verifications=ignore_verifications, 1116 try_from_hf_gcs=try_from_hf_gcs, -> 1117 use_auth_token=use_auth_token, 1118 ) 1119 /usr/local/lib/python3.7/dist-packages/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs) 635 if not downloaded_from_gcs: 636 self._download_and_prepare( --> 637 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs 638 ) 639 # Sync info /usr/local/lib/python3.7/dist-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 724 try: 725 # Prepare split will record examples associated to the split --> 726 self._prepare_split(split_generator, **prepare_split_kwargs) 727 except OSError as e: 728 raise OSError( /usr/local/lib/python3.7/dist-packages/datasets/builder.py in _prepare_split(self, split_generator) 1186 generator, unit=" tables", leave=False, disable=bool(logging.get_verbosity() == logging.NOTSET) 1187 ): -> 1188 writer.write_table(table) 1189 num_examples, num_bytes = writer.finalize() 1190 /usr/local/lib/python3.7/dist-packages/datasets/arrow_writer.py in write_table(self, pa_table, writer_batch_size) 424 # reorder the arrays if necessary + cast to self._schema 425 # we can't simply use .cast here because we may need to change the order of the columns --> 426 pa_table = pa.Table.from_arrays([pa_table[name] for name in self._schema.names], schema=self._schema) 427 batches: List[pa.RecordBatch] = pa_table.to_batches(max_chunksize=writer_batch_size) 428 self._num_bytes += sum(batch.nbytes for batch in batches) /usr/local/lib/python3.7/dist-packages/pyarrow/table.pxi in pyarrow.lib.Table.from_arrays() /usr/local/lib/python3.7/dist-packages/pyarrow/table.pxi in pyarrow.lib._sanitize_arrays() /usr/local/lib/python3.7/dist-packages/pyarrow/array.pxi in pyarrow.lib.asarray() /usr/local/lib/python3.7/dist-packages/pyarrow/table.pxi in pyarrow.lib.ChunkedArray.cast() /usr/local/lib/python3.7/dist-packages/pyarrow/compute.py in cast(arr, target_type, safe) 279 else: 280 options = CastOptions.unsafe(target_type) --> 281 return call_function("cast", [arr], options) 282 283 /usr/local/lib/python3.7/dist-packages/pyarrow/_compute.pyx in pyarrow._compute.call_function() /usr/local/lib/python3.7/dist-packages/pyarrow/_compute.pyx in pyarrow._compute.Function.call() /usr/local/lib/python3.7/dist-packages/pyarrow/error.pxi in pyarrow.lib.pyarrow_internal_check_status() /usr/local/lib/python3.7/dist-packages/pyarrow/error.pxi in pyarrow.lib.check_status() ArrowNotImplementedError: Unsupported cast from struct<train: struct<name: string, num_bytes: int64, num_examples: int64, dataset_name: string>, validation: struct<name: string, num_bytes: int64, num_examples: int64, dataset_name: string>, test: struct<name: string, num_bytes: int64, num_examples: int64, dataset_name: string>> to list using function cast_list ``` ## Expected results The dataset should be correctly loaded with `load_dataset` IMO. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.2.dev0 - Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.10 - Python version: 3.8.5 - PyArrow version: 5.0.0
2021-10-06T17:01:37Z
https://github.com/huggingface/datasets/issues/3035
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3035/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3034/comments
https://api.github.com/repos/huggingface/datasets/issues/3034/timeline
null
null
null
I_kwDODunzps48moOi
open
[]
null
3,034
{ "avatar_url": "https://avatars.githubusercontent.com/u/74556552?v=4", "events_url": "https://api.github.com/users/dconatha/events{/privacy}", "followers_url": "https://api.github.com/users/dconatha/followers", "following_url": "https://api.github.com/users/dconatha/following{/other_user}", "gists_url": "https://api.github.com/users/dconatha/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dconatha", "id": 74556552, "login": "dconatha", "node_id": "MDQ6VXNlcjc0NTU2NTUy", "organizations_url": "https://api.github.com/users/dconatha/orgs", "received_events_url": "https://api.github.com/users/dconatha/received_events", "repos_url": "https://api.github.com/users/dconatha/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dconatha/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dconatha/subscriptions", "type": "User", "url": "https://api.github.com/users/dconatha" }
Errors loading dataset using fs = a gcsfs.GCSFileSystem
https://api.github.com/repos/huggingface/datasets/issues/3034/events
null
https://api.github.com/repos/huggingface/datasets/issues/3034/labels{/name}
2021-10-05T20:07:08Z
null
false
null
null
1,016,759,202
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/3034
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Cannot load dataset using a `gcsfs.GCSFileSystem`. I'm not sure if this should be a bug in `gcsfs` or here... Basically what seems to be happening is that since datasets saves datasets as folders and folders aren't "real objects" in gcs, gcsfs raises a 404 error. There are workarounds if you use gcsfs directly to download the file, but as is I can't get `load_from_disk` to work. ## Steps to reproduce the bug ```python from datasets import load_dataset # load some dataset dataset = load_dataset("squad", split="train") # save it to gcs import gcsfs fs = gcsfs.GCSFileSystem(project="my-gs-project") dataset.save_to_disk("gs://my-bucket/squad", fs=fs) # try to load it from gcs from datasets import load_from_disk dataset2 = load_from_disk("my-bucket/squad", fs=fs) ``` ## Expected results `dataset2` would be a copy of `dataset` but loaded from my bucket. ## Actual results Long traceback but essentially it's a 404 error from gcsfs saying the object `my-bucket/squad` doesn't exist when this is called: https://github.com/huggingface/datasets/blob/9c81b7d2e6d9feae69a084a3abda265a4ca07fb5/src/datasets/arrow_dataset.py#L977 This is because there is no actual object called `my-bucket/squad`, there are objects called `my-bucket/squad/dataset.arrow`, etc. Note that *this* works fine, since it's explicitly saying "download all the objects with this prefix": ```python fs.download(src_dataset_path + "/*", dataset_path.as_posix(), recursive=True) ``` For example, I can do a workaround this way: ```python import tempfile with tempfile.TemporaryDirectory() as temppath: fs.download("gs://my-bucket/squad/*", temppath) dataset2 = load_from_disk(temppath) ``` It's unclear to me if it's `gcsfs`'s responsibility to say "hey that's folder not a file, I should try to get objects inside of it not the object itself", or if that's `datasets`'s responsibility... I'm leaning towards the latter since you're never loading a dataset from one file using this function/method, only a dataset folder? Another minor thing that should maybe should be rolled into this bug... https://github.com/huggingface/datasets/blob/9c81b7d2e6d9feae69a084a3abda265a4ca07fb5/src/datasets/arrow_dataset.py#L968 These fail if you pass in a `gs://` path, e.g. ```python dataset2 = load_from_disk("gs://my-bucket/squad", fs=fs) ``` Because at this point, `dataset_info_path` is `gs:/my-bucket/squad/dataset_info.json`, gcsfs throws a: ``` Invalid bucket name: 'gs:' ``` error ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.1 - Platform: macOS Big Sur 11.6 - Python version: 3.7.12 - PyArrow version: 5.0.0
2021-10-05T20:26:39Z
https://github.com/huggingface/datasets/issues/3034
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3034/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3033/comments
https://api.github.com/repos/huggingface/datasets/issues/3033/timeline
2021-10-05T17:54:57Z
null
null
PR_kwDODunzps4std7u
closed
[]
false
3,033
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Actual "proper" install of ruamel.yaml in the windows CI
https://api.github.com/repos/huggingface/datasets/issues/3033/events
null
https://api.github.com/repos/huggingface/datasets/issues/3033/labels{/name}
2021-10-05T17:52:07Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3033.diff", "html_url": "https://github.com/huggingface/datasets/pull/3033", "merged_at": "2021-10-05T17:54:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/3033.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3033" }
1,016,619,572
[]
https://api.github.com/repos/huggingface/datasets/issues/3033
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
It was impossible to update the package directly with `pip`. Indeed it was installed with `distutils` which prevents `pip` or `conda` to uninstall it. I had to `rm` a directory from the `site-packages` python directory, and then do `pip install ruamel.yaml` It's not that "proper" but I couldn't find better solutions
2021-10-05T17:54:57Z
https://github.com/huggingface/datasets/pull/3033
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3033/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3032/comments
https://api.github.com/repos/huggingface/datasets/issues/3032/timeline
2021-10-12T15:25:46Z
null
completed
I_kwDODunzps48lmIb
closed
[]
null
3,032
{ "avatar_url": "https://avatars.githubusercontent.com/u/715491?v=4", "events_url": "https://api.github.com/users/borisdayma/events{/privacy}", "followers_url": "https://api.github.com/users/borisdayma/followers", "following_url": "https://api.github.com/users/borisdayma/following{/other_user}", "gists_url": "https://api.github.com/users/borisdayma/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/borisdayma", "id": 715491, "login": "borisdayma", "node_id": "MDQ6VXNlcjcxNTQ5MQ==", "organizations_url": "https://api.github.com/users/borisdayma/orgs", "received_events_url": "https://api.github.com/users/borisdayma/received_events", "repos_url": "https://api.github.com/users/borisdayma/repos", "site_admin": false, "starred_url": "https://api.github.com/users/borisdayma/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/borisdayma/subscriptions", "type": "User", "url": "https://api.github.com/users/borisdayma" }
Error when loading private dataset with "data_files" arg
https://api.github.com/repos/huggingface/datasets/issues/3032/events
null
https://api.github.com/repos/huggingface/datasets/issues/3032/labels{/name}
2021-10-05T15:46:27Z
null
false
null
null
1,016,488,475
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/3032
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug A clear and concise description of what the bug is. Private datasets with no loading script can't be loaded using `data_files` parameter. ## Steps to reproduce the bug ```python from datasets import load_dataset data_files = {"train": "**/train/*/*.jsonl", "valid": "**/valid/*/*.jsonl"} dataset = load_dataset('dalle-mini/encoded', data_files=data_files, use_auth_token=True, streaming=True) ``` Same error happens in non-streaming mode. ## Expected results Files should be loaded (whether in streaming or not). ## Actual results Error: ``` --------------------------------------------------------------------------- FileNotFoundError Traceback (most recent call last) /usr/local/lib/python3.7/dist-packages/datasets/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, dynamic_modules_path, return_resolved_file_path, return_associated_base_path, data_files, **download_kwargs) 539 try: --> 540 local_path = cached_path(file_path, download_config=download_config) 541 except FileNotFoundError: 8 frames FileNotFoundError: Couldn't find file at https://huggingface.co/datasets/dalle-mini/encoded/resolve/main/encoded.py During handling of the above exception, another exception occurred: HTTPError Traceback (most recent call last) HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/api/datasets/dalle-mini/encoded?full=true During handling of the above exception, another exception occurred: FileNotFoundError Traceback (most recent call last) /usr/local/lib/python3.7/dist-packages/datasets/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, dynamic_modules_path, return_resolved_file_path, return_associated_base_path, data_files, **download_kwargs) 547 except Exception: 548 raise FileNotFoundError( --> 549 f"Couldn't find a directory or a {resource_type} named '{path}'. " 550 f"It doesn't exist locally at {expected_dir_for_combined_path_abs} or remotely on {hf_api.endpoint}/datasets" 551 ) FileNotFoundError: Couldn't find a directory or a dataset named 'dalle-mini/encoded'. It doesn't exist locally at /content/dalle-mini/encoded or remotely on https://huggingface.co/datasets ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.1 - Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.12 - PyArrow version: 3.0.0 @lhoestq
2021-10-12T15:26:22Z
https://github.com/huggingface/datasets/issues/3032
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3032/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3031/comments
https://api.github.com/repos/huggingface/datasets/issues/3031/timeline
2021-10-18T14:59:30Z
null
null
PR_kwDODunzps4ss9jn
closed
[]
false
3,031
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Align tqdm control with cache control
https://api.github.com/repos/huggingface/datasets/issues/3031/events
null
https://api.github.com/repos/huggingface/datasets/issues/3031/labels{/name}
2021-10-05T15:18:49Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3031.diff", "html_url": "https://github.com/huggingface/datasets/pull/3031", "merged_at": "2021-10-18T14:59:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/3031.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3031" }
1,016,458,496
[]
https://api.github.com/repos/huggingface/datasets/issues/3031
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Currently, once disabled with `disable_progress_bar`, progress bars cannot be re-enabled again. To overcome this limitation, this PR introduces the `set_progress_bar_enabled` function that accepts a boolean indicating whether to display progress bars. The goal is to provide a similar API to the existing cache control API. Following the Zen of Python (😄), there should be one and preferably only one obvious way to do it, so I'm also deprecating the aforementioned `disable_progress_bar` function. Additionally, I justify the deprecation with the fact that this function has never been in the docs. Moreover, similar API changes have recently been introduced to [`tfds`](https://github.com/tensorflow/datasets/blob/a1e8b98f45b0214082b546cc967c67c43fffda55/tensorflow_datasets/core/utils/tqdm_utils.py#L98-L112). Considering the popularity of the [comment](https://github.com/huggingface/datasets/issues/1627#issuecomment-751383559) I made a while ago, this API (`set_progress_bar_enabled` and `is_progress_bar_enabled`) should be mentioned in the docs, but I'm not sure where to put it exactly. Maybe we can replace the `logging_methods` page under `package_reference` with `utility_methods` and then introduce two subsections on that page: `Logging methods` and `tqdm control`. Additionally, this PR: * adds the `disable_tqdm` keyword arg of `Dataset._map_single` to the `ignore_kwargs` list to ignore it when computing the fingerprint (forgot to add it in #2696) * deletes the unused components in `tqdm_utils.py`, which seem to be inherited from `tfds` * disables the tqdm output in the test suite. As I see it, this output doesn't seem informative, but let me know if this is not a good idea
2021-10-18T15:00:21Z
https://github.com/huggingface/datasets/pull/3031
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3031/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3030/comments
https://api.github.com/repos/huggingface/datasets/issues/3030/timeline
2021-10-08T15:31:53Z
null
null
PR_kwDODunzps4ss41W
closed
[]
false
3,030
{ "avatar_url": "https://avatars.githubusercontent.com/u/31893406?v=4", "events_url": "https://api.github.com/users/cccntu/events{/privacy}", "followers_url": "https://api.github.com/users/cccntu/followers", "following_url": "https://api.github.com/users/cccntu/following{/other_user}", "gists_url": "https://api.github.com/users/cccntu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cccntu", "id": 31893406, "login": "cccntu", "node_id": "MDQ6VXNlcjMxODkzNDA2", "organizations_url": "https://api.github.com/users/cccntu/orgs", "received_events_url": "https://api.github.com/users/cccntu/received_events", "repos_url": "https://api.github.com/users/cccntu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cccntu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cccntu/subscriptions", "type": "User", "url": "https://api.github.com/users/cccntu" }
Add `remove_columns` to `IterableDataset`
https://api.github.com/repos/huggingface/datasets/issues/3030/events
null
https://api.github.com/repos/huggingface/datasets/issues/3030/labels{/name}
2021-10-05T14:58:33Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3030.diff", "html_url": "https://github.com/huggingface/datasets/pull/3030", "merged_at": "2021-10-08T15:31:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/3030.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3030" }
1,016,435,324
[]
https://api.github.com/repos/huggingface/datasets/issues/3030
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Fixes #2944 WIP * Not tested yet. * We might want to allow batched remove for efficiency. @lhoestq Do you think it should have `batched=` and `batch_size=`?
2021-10-08T15:33:15Z
https://github.com/huggingface/datasets/pull/3030
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3030/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3029/comments
https://api.github.com/repos/huggingface/datasets/issues/3029/timeline
2021-10-05T14:56:45Z
null
null
PR_kwDODunzps4ssvkr
closed
[]
false
3,029
{ "avatar_url": "https://avatars.githubusercontent.com/u/417568?v=4", "events_url": "https://api.github.com/users/craffel/events{/privacy}", "followers_url": "https://api.github.com/users/craffel/followers", "following_url": "https://api.github.com/users/craffel/following{/other_user}", "gists_url": "https://api.github.com/users/craffel/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/craffel", "id": 417568, "login": "craffel", "node_id": "MDQ6VXNlcjQxNzU2OA==", "organizations_url": "https://api.github.com/users/craffel/orgs", "received_events_url": "https://api.github.com/users/craffel/received_events", "repos_url": "https://api.github.com/users/craffel/repos", "site_admin": false, "starred_url": "https://api.github.com/users/craffel/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/craffel/subscriptions", "type": "User", "url": "https://api.github.com/users/craffel" }
Use standard open-domain validation split in nq_open
https://api.github.com/repos/huggingface/datasets/issues/3029/events
null
https://api.github.com/repos/huggingface/datasets/issues/3029/labels{/name}
2021-10-05T14:19:27Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3029.diff", "html_url": "https://github.com/huggingface/datasets/pull/3029", "merged_at": "2021-10-05T14:56:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/3029.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3029" }
1,016,389,901
[]
https://api.github.com/repos/huggingface/datasets/issues/3029
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
The nq_open dataset originally drew the validation set from this file: https://github.com/google-research-datasets/natural-questions/blob/master/nq_open/NQ-open.efficientqa.dev.1.1.sample.jsonl However, that's the dev set used specifically and only for the efficientqa competition, and it's not the same dev set as is used in every open-domain question answering paper (including the Lee et al paper that introduced the open-domain variant of NQ, cited at the top of the dataset file). This PR changes nq_open to use the standard validation split and bumps the version to 2.0.0 since this is a breaking change.
2021-10-05T14:56:46Z
https://github.com/huggingface/datasets/pull/3029
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3029/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3028/comments
https://api.github.com/repos/huggingface/datasets/issues/3028/timeline
2021-10-05T11:51:22Z
null
null
PR_kwDODunzps4ssO4s
closed
[]
false
3,028
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Properly install ruamel-yaml for windows CI
https://api.github.com/repos/huggingface/datasets/issues/3028/events
null
https://api.github.com/repos/huggingface/datasets/issues/3028/labels{/name}
2021-10-05T11:51:15Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3028.diff", "html_url": "https://github.com/huggingface/datasets/pull/3028", "merged_at": "2021-10-05T11:51:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/3028.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3028" }
1,016,230,272
[]
https://api.github.com/repos/huggingface/datasets/issues/3028
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
null
2021-10-05T14:02:12Z
https://github.com/huggingface/datasets/pull/3028
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3028/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3027/comments
https://api.github.com/repos/huggingface/datasets/issues/3027/timeline
2021-11-05T17:49:57Z
null
completed
I_kwDODunzps48kThl
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
3,027
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Resolve data_files by split name
https://api.github.com/repos/huggingface/datasets/issues/3027/events
null
https://api.github.com/repos/huggingface/datasets/issues/3027/labels{/name}
2021-10-05T10:24:36Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
1,016,150,117
[]
https://api.github.com/repos/huggingface/datasets/issues/3027
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This issue is about discussing the default behavior when someone loads a dataset that consists in data files. For example: ```python load_dataset("lhoestq/demo1") ``` should return two splits "train" and "test" since the dataset repostiory is like ``` data/ ├── train.csv └── test.csv ``` Currently it returns only one split "train" which contains the data of both files I started playing with this idea on this branch btw: `resolve-data_files-by-split-name` Basically the idea is that if you named you data files after split names then the default pattern is ```python { "train": ["*train*"], "test": ["*test*"], "validation": ["*dev*", "valid"], } ``` otherwise it's ```python { "train": ["*"] } ``` Let me know what you think ! cc @albertvillanova @LysandreJik @vblagoje
2021-11-05T17:49:58Z
https://github.com/huggingface/datasets/issues/3027
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/3027/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3026/comments
https://api.github.com/repos/huggingface/datasets/issues/3026/timeline
2021-10-08T16:01:24Z
null
null
PR_kwDODunzps4srtyc
closed
[]
false
3,026
{ "avatar_url": "https://avatars.githubusercontent.com/u/3775944?v=4", "events_url": "https://api.github.com/users/JoelNiklaus/events{/privacy}", "followers_url": "https://api.github.com/users/JoelNiklaus/followers", "following_url": "https://api.github.com/users/JoelNiklaus/following{/other_user}", "gists_url": "https://api.github.com/users/JoelNiklaus/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JoelNiklaus", "id": 3775944, "login": "JoelNiklaus", "node_id": "MDQ6VXNlcjM3NzU5NDQ=", "organizations_url": "https://api.github.com/users/JoelNiklaus/orgs", "received_events_url": "https://api.github.com/users/JoelNiklaus/received_events", "repos_url": "https://api.github.com/users/JoelNiklaus/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JoelNiklaus/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JoelNiklaus/subscriptions", "type": "User", "url": "https://api.github.com/users/JoelNiklaus" }
added arxiv paper inswiss_judgment_prediction dataset card
https://api.github.com/repos/huggingface/datasets/issues/3026/events
null
https://api.github.com/repos/huggingface/datasets/issues/3026/labels{/name}
2021-10-05T09:02:01Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3026.diff", "html_url": "https://github.com/huggingface/datasets/pull/3026", "merged_at": "2021-10-08T16:01:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/3026.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3026" }
1,016,067,794
[]
https://api.github.com/repos/huggingface/datasets/issues/3026
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2021-10-08T16:01:44Z
https://github.com/huggingface/datasets/pull/3026
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3026/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3025/comments
https://api.github.com/repos/huggingface/datasets/issues/3025/timeline
2021-10-05T09:58:27Z
null
null
PR_kwDODunzps4srsgG
closed
[]
false
3,025
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix Windows test suite
https://api.github.com/repos/huggingface/datasets/issues/3025/events
null
https://api.github.com/repos/huggingface/datasets/issues/3025/labels{/name}
2021-10-05T08:55:22Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3025.diff", "html_url": "https://github.com/huggingface/datasets/pull/3025", "merged_at": "2021-10-05T09:58:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/3025.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3025" }
1,016,061,222
[]
https://api.github.com/repos/huggingface/datasets/issues/3025
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Try a hotfix to restore Windows test suite. Fix #3024.
2021-10-05T09:58:28Z
https://github.com/huggingface/datasets/pull/3025
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3025/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3024/comments
https://api.github.com/repos/huggingface/datasets/issues/3024/timeline
2021-10-05T09:58:27Z
null
completed
I_kwDODunzps48j7yv
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
3,024
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Windows test suite fails
https://api.github.com/repos/huggingface/datasets/issues/3024/events
null
https://api.github.com/repos/huggingface/datasets/issues/3024/labels{/name}
2021-10-05T08:46:46Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,016,052,911
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/3024
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Describe the bug There is an error during installation of tests dependencies for Windows: https://app.circleci.com/pipelines/github/huggingface/datasets/7981/workflows/9b6a0114-2b8e-4069-94e5-e844dbbdba4e/jobs/49206 ``` ERROR: Cannot uninstall 'ruamel-yaml'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. ```
2021-10-05T09:58:27Z
https://github.com/huggingface/datasets/issues/3024
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3024/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3023/comments
https://api.github.com/repos/huggingface/datasets/issues/3023/timeline
2021-10-05T11:56:55Z
null
null
PR_kwDODunzps4srQ4i
closed
[]
false
3,023
{ "avatar_url": "https://avatars.githubusercontent.com/u/24835382?v=4", "events_url": "https://api.github.com/users/qqaatw/events{/privacy}", "followers_url": "https://api.github.com/users/qqaatw/followers", "following_url": "https://api.github.com/users/qqaatw/following{/other_user}", "gists_url": "https://api.github.com/users/qqaatw/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/qqaatw", "id": 24835382, "login": "qqaatw", "node_id": "MDQ6VXNlcjI0ODM1Mzgy", "organizations_url": "https://api.github.com/users/qqaatw/orgs", "received_events_url": "https://api.github.com/users/qqaatw/received_events", "repos_url": "https://api.github.com/users/qqaatw/repos", "site_admin": false, "starred_url": "https://api.github.com/users/qqaatw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/qqaatw/subscriptions", "type": "User", "url": "https://api.github.com/users/qqaatw" }
Fix typo
https://api.github.com/repos/huggingface/datasets/issues/3023/events
null
https://api.github.com/repos/huggingface/datasets/issues/3023/labels{/name}
2021-10-05T06:06:11Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3023.diff", "html_url": "https://github.com/huggingface/datasets/pull/3023", "merged_at": "2021-10-05T11:56:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/3023.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3023" }
1,015,923,031
[]
https://api.github.com/repos/huggingface/datasets/issues/3023
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2021-10-05T11:56:55Z
https://github.com/huggingface/datasets/pull/3023
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3023/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3022/comments
https://api.github.com/repos/huggingface/datasets/issues/3022/timeline
2021-10-13T09:03:09Z
null
null
PR_kwDODunzps4sqve6
closed
[]
false
3,022
{ "avatar_url": "https://avatars.githubusercontent.com/u/21180505?v=4", "events_url": "https://api.github.com/users/xhluca/events{/privacy}", "followers_url": "https://api.github.com/users/xhluca/followers", "following_url": "https://api.github.com/users/xhluca/following{/other_user}", "gists_url": "https://api.github.com/users/xhluca/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/xhluca", "id": 21180505, "login": "xhluca", "node_id": "MDQ6VXNlcjIxMTgwNTA1", "organizations_url": "https://api.github.com/users/xhluca/orgs", "received_events_url": "https://api.github.com/users/xhluca/received_events", "repos_url": "https://api.github.com/users/xhluca/repos", "site_admin": false, "starred_url": "https://api.github.com/users/xhluca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xhluca/subscriptions", "type": "User", "url": "https://api.github.com/users/xhluca" }
MeDAL dataset: Add further description and update download URL
https://api.github.com/repos/huggingface/datasets/issues/3022/events
null
https://api.github.com/repos/huggingface/datasets/issues/3022/labels{/name}
2021-10-05T00:13:28Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3022.diff", "html_url": "https://github.com/huggingface/datasets/pull/3022", "merged_at": "2021-10-13T09:03:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/3022.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3022" }
1,015,750,221
[]
https://api.github.com/repos/huggingface/datasets/issues/3022
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Added more details in the following sections: * Dataset Structure * Data Instances * Data Splits * Source Data * Annotations * Discussions of Biases * LIcensing Information
2021-10-13T09:03:09Z
https://github.com/huggingface/datasets/pull/3022
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3022/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3021/comments
https://api.github.com/repos/huggingface/datasets/issues/3021/timeline
2021-10-06T08:36:45Z
null
null
PR_kwDODunzps4spzJU
closed
[]
false
3,021
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Support loading dataset from multiple zipped CSV data files
https://api.github.com/repos/huggingface/datasets/issues/3021/events
null
https://api.github.com/repos/huggingface/datasets/issues/3021/labels{/name}
2021-10-04T17:33:57Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3021.diff", "html_url": "https://github.com/huggingface/datasets/pull/3021", "merged_at": "2021-10-06T08:36:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/3021.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3021" }
1,015,444,094
[]
https://api.github.com/repos/huggingface/datasets/issues/3021
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix partially #3018. CC: @lewtun
2021-10-06T08:36:46Z
https://github.com/huggingface/datasets/pull/3021
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/3021/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3020/comments
https://api.github.com/repos/huggingface/datasets/issues/3020/timeline
2021-10-22T10:29:31Z
null
null
PR_kwDODunzps4sprfa
closed
[]
false
3,020
{ "avatar_url": "https://avatars.githubusercontent.com/u/91226467?v=4", "events_url": "https://api.github.com/users/hacobe/events{/privacy}", "followers_url": "https://api.github.com/users/hacobe/followers", "following_url": "https://api.github.com/users/hacobe/following{/other_user}", "gists_url": "https://api.github.com/users/hacobe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hacobe", "id": 91226467, "login": "hacobe", "node_id": "MDQ6VXNlcjkxMjI2NDY3", "organizations_url": "https://api.github.com/users/hacobe/orgs", "received_events_url": "https://api.github.com/users/hacobe/received_events", "repos_url": "https://api.github.com/users/hacobe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hacobe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hacobe/subscriptions", "type": "User", "url": "https://api.github.com/users/hacobe" }
Add a metric for the MATH dataset (competition_math).
https://api.github.com/repos/huggingface/datasets/issues/3020/events
null
https://api.github.com/repos/huggingface/datasets/issues/3020/labels{/name}
2021-10-04T16:52:16Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3020.diff", "html_url": "https://github.com/huggingface/datasets/pull/3020", "merged_at": "2021-10-22T10:29:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/3020.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3020" }
1,015,406,105
[]
https://api.github.com/repos/huggingface/datasets/issues/3020
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This metric computes accuracy for the MATH dataset (https://arxiv.org/abs/2103.03874) after canonicalizing the prediction and the reference (e.g., converting "1/2" to "\\\\frac{1}{2}").
2021-10-22T10:29:31Z
https://github.com/huggingface/datasets/pull/3020
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3020/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3019/comments
https://api.github.com/repos/huggingface/datasets/issues/3019/timeline
2021-10-05T08:33:07Z
null
null
PR_kwDODunzps4speOB
closed
[]
false
3,019
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix filter leaking
https://api.github.com/repos/huggingface/datasets/issues/3019/events
null
https://api.github.com/repos/huggingface/datasets/issues/3019/labels{/name}
2021-10-04T15:42:58Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3019.diff", "html_url": "https://github.com/huggingface/datasets/pull/3019", "merged_at": "2021-10-05T08:33:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/3019.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3019" }
1,015,339,983
[]
https://api.github.com/repos/huggingface/datasets/issues/3019
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
If filter is called after using a first transform `shuffle`, `select`, `shard`, `train_test_split`, or `filter`, then it could not work as expected and return examples from before the first transform. This is because the indices mapping was not taken into account when saving the indices to keep when doing the filtering Affected versions: 1.12.0 and 1.12.1 This should fix #3010
2022-06-03T08:28:14Z
https://github.com/huggingface/datasets/pull/3019
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3019/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3018/comments
https://api.github.com/repos/huggingface/datasets/issues/3018/timeline
null
null
null
I_kwDODunzps48hG4F
open
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
3,018
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Support multiple zipped CSV data files
https://api.github.com/repos/huggingface/datasets/issues/3018/events
null
https://api.github.com/repos/huggingface/datasets/issues/3018/labels{/name}
2021-10-04T15:16:59Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,015,311,877
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/3018
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
As requested by @lewtun, support loading multiple zipped CSV data files. ```python from datasets import load_dataset url = "https://domain.org/filename.zip" data_files = {"train": "train_filename.csv", "test": "test_filename.csv"} dataset = load_dataset("csv", data_dir=url, data_files=data_files) ```
2021-10-05T14:32:57Z
https://github.com/huggingface/datasets/issues/3018
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/3018/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3017/comments
https://api.github.com/repos/huggingface/datasets/issues/3017/timeline
2021-10-05T11:37:00Z
null
null
PR_kwDODunzps4spE9m
closed
[]
false
3,017
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Remove unused parameter in xdirname
https://api.github.com/repos/huggingface/datasets/issues/3017/events
null
https://api.github.com/repos/huggingface/datasets/issues/3017/labels{/name}
2021-10-04T13:55:53Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3017.diff", "html_url": "https://github.com/huggingface/datasets/pull/3017", "merged_at": "2021-10-05T11:37:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/3017.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3017" }
1,015,215,528
[]
https://api.github.com/repos/huggingface/datasets/issues/3017
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Minor fix to remove unused args `*p` in `xdirname`.
2021-10-05T11:37:01Z
https://github.com/huggingface/datasets/pull/3017
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3017/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3016/comments
https://api.github.com/repos/huggingface/datasets/issues/3016/timeline
2021-10-04T15:23:04Z
null
null
PR_kwDODunzps4spDlX
closed
[]
false
3,016
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix Windows paths in LJ Speech dataset
https://api.github.com/repos/huggingface/datasets/issues/3016/events
null
https://api.github.com/repos/huggingface/datasets/issues/3016/labels{/name}
2021-10-04T13:49:37Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3016.diff", "html_url": "https://github.com/huggingface/datasets/pull/3016", "merged_at": "2021-10-04T15:23:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/3016.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3016" }
1,015,208,654
[]
https://api.github.com/repos/huggingface/datasets/issues/3016
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Minor fix in LJ Speech dataset for Windows pathname component separator. Related to #1878.
2021-10-04T15:23:05Z
https://github.com/huggingface/datasets/pull/3016
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3016/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3015/comments
https://api.github.com/repos/huggingface/datasets/issues/3015/timeline
2021-10-05T13:46:38Z
null
null
PR_kwDODunzps4so0GX
closed
[]
false
3,015
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Extend support for streaming datasets that use glob.glob
https://api.github.com/repos/huggingface/datasets/issues/3015/events
null
https://api.github.com/repos/huggingface/datasets/issues/3015/labels{/name}
2021-10-04T12:42:37Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3015.diff", "html_url": "https://github.com/huggingface/datasets/pull/3015", "merged_at": "2021-10-05T13:46:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/3015.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3015" }
1,015,130,845
[]
https://api.github.com/repos/huggingface/datasets/issues/3015
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This PR extends the support in streaming mode for datasets that use `glob`, by patching the function `glob.glob`. Related to #2880, #2876, #2874
2021-10-05T13:46:39Z
https://github.com/huggingface/datasets/pull/3015
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3015/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3014/comments
https://api.github.com/repos/huggingface/datasets/issues/3014/timeline
2021-10-04T12:46:44Z
null
null
PR_kwDODunzps4son8A
closed
[]
false
3,014
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix Windows path in MATH dataset
https://api.github.com/repos/huggingface/datasets/issues/3014/events
null
https://api.github.com/repos/huggingface/datasets/issues/3014/labels{/name}
2021-10-04T11:41:07Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3014.diff", "html_url": "https://github.com/huggingface/datasets/pull/3014", "merged_at": "2021-10-04T12:46:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/3014.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3014" }
1,015,070,751
[]
https://api.github.com/repos/huggingface/datasets/issues/3014
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Minor fix in MATH dataset for Windows pathname component separator. Related to #2982.
2021-10-04T12:46:44Z
https://github.com/huggingface/datasets/pull/3014
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3014/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3013/comments
https://api.github.com/repos/huggingface/datasets/issues/3013/timeline
2022-02-21T15:57:10Z
null
completed
I_kwDODunzps48fxEj
closed
[]
null
3,013
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
Improve `get_dataset_infos`?
https://api.github.com/repos/huggingface/datasets/issues/3013/events
null
https://api.github.com/repos/huggingface/datasets/issues/3013/labels{/name}
2021-10-04T09:47:04Z
null
false
null
null
1,014,960,419
[ { "color": "d876e3", "default": true, "description": "Further information is requested", "id": 1935892912, "name": "question", "node_id": "MDU6TGFiZWwxOTM1ODkyOTEy", "url": "https://api.github.com/repos/huggingface/datasets/labels/question" }, { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
https://api.github.com/repos/huggingface/datasets/issues/3013
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Using the dedicated function `get_dataset_infos` on a dataset that has no dataset-info.json file returns an empty info: ``` >>> from datasets import get_dataset_infos >>> get_dataset_infos('wit') {} ``` While it's totally possible to get it (regenerate it) with: ``` >>> from datasets import load_dataset_builder >>> builder = load_dataset_builder('wit') >>> builder.info DatasetInfo(description='Wikipedia-based Image Text (WIT) Dataset is a large multimodal multilingual dataset. WIT is composed of a curated set\n of 37.6 million entity rich image-text examples with 11.5 million unique images across 108 Wikipedia languages. Its\n size enables WIT to be used as a pretraining dataset for multimodal machine learning models.\n', citation='@article{srinivasan2021wit,\n title={WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning},\n author={Srinivasan, Krishna and Raman, Karthik and Chen, Jiecao and Bendersky, Michael and Najork, Marc},\n journal={arXiv preprint arXiv:2103.01913},\n year={2021}\n}\n', homepage='https://github.com/google-research-datasets/wit', license='', features={'b64_bytes': Value(dtype='string', id=None), 'embedding': Sequence(feature=Value(dtype='float64', id=None), length=-1, id=None), 'image_url': Value(dtype='string', id=None), 'metadata_url': Value(dtype='string', id=None), 'original_height': Value(dtype='int32', id=None), 'original_width': Value(dtype='int32', id=None), 'mime_type': Value(dtype='string', id=None), 'caption_attribution_description': Value(dtype='string', id=None), 'wit_features': Sequence(feature={'language': Value(dtype='string', id=None), 'page_url': Value(dtype='string', id=None), 'attribution_passes_lang_id': Value(dtype='string', id=None), 'caption_alt_text_description': Value(dtype='string', id=None), 'caption_reference_description': Value(dtype='string', id=None), 'caption_title_and_reference_description': Value(dtype='string', id=None), 'context_page_description': Value(dtype='string', id=None), 'context_section_description': Value(dtype='string', id=None), 'hierarchical_section_title': Value(dtype='string', id=None), 'is_main_image': Value(dtype='string', id=None), 'page_changed_recently': Value(dtype='string', id=None), 'page_title': Value(dtype='string', id=None), 'section_title': Value(dtype='string', id=None)}, length=-1, id=None)}, post_processed=None, supervised_keys=None, task_templates=None, builder_name='wit', config_name='default', version=0.0.0, splits=None, download_checksums=None, download_size=None, post_processing_size=None, dataset_size=None, size_in_bytes=None) ``` Should we test if info is empty, and in that case regenerate it? Or always generate it?
2022-02-21T15:57:10Z
https://github.com/huggingface/datasets/issues/3013
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3013/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3012/comments
https://api.github.com/repos/huggingface/datasets/issues/3012/timeline
2021-10-04T11:30:33Z
null
null
PR_kwDODunzps4soRTu
closed
[]
false
3,012
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Replace item with float in metrics
https://api.github.com/repos/huggingface/datasets/issues/3012/events
null
https://api.github.com/repos/huggingface/datasets/issues/3012/labels{/name}
2021-10-04T09:45:28Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3012.diff", "html_url": "https://github.com/huggingface/datasets/pull/3012", "merged_at": "2021-10-04T11:30:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/3012.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3012" }
1,014,958,931
[]
https://api.github.com/repos/huggingface/datasets/issues/3012
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
As pointed out by @mariosasko in #3001, calling `float()` instad of `.item()` is faster. Moreover, it might avoid potential issues if any of the third-party functions eventually returns a `float` instead of an `np.float64`. Related to #3001.
2021-10-04T11:30:34Z
https://github.com/huggingface/datasets/pull/3012
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3012/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3011/comments
https://api.github.com/repos/huggingface/datasets/issues/3011/timeline
null
null
null
I_kwDODunzps48frCh
open
[]
null
3,011
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
load_dataset_builder should error if "name" does not exist?
https://api.github.com/repos/huggingface/datasets/issues/3011/events
null
https://api.github.com/repos/huggingface/datasets/issues/3011/labels{/name}
2021-10-04T09:20:46Z
null
false
null
null
1,014,935,713
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
https://api.github.com/repos/huggingface/datasets/issues/3011
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
``` import datasets as ds builder = ds.load_dataset_builder('sent_comp', name="doesnotexist") builder.info.config_name ``` returns ``` 'doesnotexist' ``` Shouldn't it raise an error instead? For this dataset, the only valid values for `name` should be: `"default"` or `None` (ie. argument not passed)
2022-09-20T13:05:07Z
https://github.com/huggingface/datasets/issues/3011
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3011/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3010/comments
https://api.github.com/repos/huggingface/datasets/issues/3010/timeline
2022-06-01T17:36:44Z
null
completed
I_kwDODunzps48fm1G
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
3,010
{ "avatar_url": "https://avatars.githubusercontent.com/u/22641583?v=4", "events_url": "https://api.github.com/users/DrMatters/events{/privacy}", "followers_url": "https://api.github.com/users/DrMatters/followers", "following_url": "https://api.github.com/users/DrMatters/following{/other_user}", "gists_url": "https://api.github.com/users/DrMatters/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/DrMatters", "id": 22641583, "login": "DrMatters", "node_id": "MDQ6VXNlcjIyNjQxNTgz", "organizations_url": "https://api.github.com/users/DrMatters/orgs", "received_events_url": "https://api.github.com/users/DrMatters/received_events", "repos_url": "https://api.github.com/users/DrMatters/repos", "site_admin": false, "starred_url": "https://api.github.com/users/DrMatters/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DrMatters/subscriptions", "type": "User", "url": "https://api.github.com/users/DrMatters" }
Chain filtering is leaking
https://api.github.com/repos/huggingface/datasets/issues/3010/events
null
https://api.github.com/repos/huggingface/datasets/issues/3010/labels{/name}
2021-10-04T09:04:55Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
1,014,918,470
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/3010
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug As there's no support for lists within dataset fields, I convert my lists to json-string format. However, the bug described is occurring even when the data format is 'string'. These samples show that filtering behavior diverges from what's expected when chaining filterings. On sample 2 the second filtering leads to "leaking" of data that should've been filtered on the first filtering into the results. ## Steps to reproduce the bug Sample 1: ```python import datasets import json items = [[1, 2], [3], [4]] jsoned_items = map(json.dumps, [[1, 2], [3], [4]]) ds = datasets.Dataset.from_dict({'a': jsoned_items}) print(list(ds)) # > Prints: [{'a': '[1, 2]'}, {'a': '[3]'}, {'a': '[4]'}] as expected filtered = ds # get all lists that are shorter than 2 filtered = filtered.filter(lambda x: len(json.loads(x['a'])) < 2, load_from_cache_file=False) print(list(filtered)) # > Prints: [{'a': '[3]'}, {'a': '[4]'}] as expected # get all lists, which have a value bigger than 3 on its zero index filtered = filtered.filter(lambda x: json.loads(x['a'])[0] > 3, load_from_cache_file=False) print(list(filtered)) # > Should be: [{'a': [4]}] # > Prints: [{'a': [3]}] ``` Sample 2: ```python import datasets import json items = [[1, 2], [3], [4]] jsoned_items = map(json.dumps, [[1, 2], [3], [4]]) ds = datasets.Dataset.from_dict({'a': jsoned_items}) print(list(ds)) # > Prints: [{'a': '[1, 2]'}, {'a': '[3]'}, {'a': '[4]'}] filtered = ds # get all lists, which have a value bigger than 3 on its zero index filtered = filtered.filter(lambda x: json.loads(x['a'])[0] > 3, load_from_cache_file=False) print(list(filtered)) # > Prints: [{'a': '[4]'}] as expected # get all lists that are shorter than 2 filtered = filtered.filter(lambda x: len(json.loads(x['a'])) < 2, load_from_cache_file=False) print(list(filtered)) # > Prints: [{'a': '[1, 2]'}] # > Should be: [{'a': '[4]'}] (remain intact) ``` ## Expected results Expected and actual results are attached to the code snippets. ## Actual results Expected and actual results are attached to the code snippets. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.1 - Platform: Windows-10-10.0.19042-SP0 - Python version: 3.9.7 - PyArrow version: 5.0.0
2022-06-01T17:36:44Z
https://github.com/huggingface/datasets/issues/3010
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3010/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3009/comments
https://api.github.com/repos/huggingface/datasets/issues/3009/timeline
2021-10-04T13:43:25Z
null
null
PR_kwDODunzps4sn_YG
closed
[]
false
3,009
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix Windows paths in SUPERB benchmark datasets
https://api.github.com/repos/huggingface/datasets/issues/3009/events
null
https://api.github.com/repos/huggingface/datasets/issues/3009/labels{/name}
2021-10-04T08:13:49Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3009.diff", "html_url": "https://github.com/huggingface/datasets/pull/3009", "merged_at": "2021-10-04T13:43:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/3009.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3009" }
1,014,868,235
[]
https://api.github.com/repos/huggingface/datasets/issues/3009
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Minor fix in SUPERB benchmark datasets for Windows pathname component separator. Related to #2884, #2783 and #2619.
2021-10-04T13:43:25Z
https://github.com/huggingface/datasets/pull/3009
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3009/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3008/comments
https://api.github.com/repos/huggingface/datasets/issues/3008/timeline
2021-10-04T09:29:36Z
null
null
PR_kwDODunzps4sn7iU
closed
[]
false
3,008
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix precision/recall metrics with None average
https://api.github.com/repos/huggingface/datasets/issues/3008/events
null
https://api.github.com/repos/huggingface/datasets/issues/3008/labels{/name}
2021-10-04T07:54:15Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3008.diff", "html_url": "https://github.com/huggingface/datasets/pull/3008", "merged_at": "2021-10-04T09:29:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/3008.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3008" }
1,014,849,163
[]
https://api.github.com/repos/huggingface/datasets/issues/3008
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Related to issue #2979 and PR #2992.
2021-10-04T09:29:37Z
https://github.com/huggingface/datasets/pull/3008
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3008/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3007/comments
https://api.github.com/repos/huggingface/datasets/issues/3007/timeline
2021-10-04T09:27:57Z
null
null
PR_kwDODunzps4sns-n
closed
[]
false
3,007
{ "avatar_url": "https://avatars.githubusercontent.com/u/35955430?v=4", "events_url": "https://api.github.com/users/Yann21/events{/privacy}", "followers_url": "https://api.github.com/users/Yann21/followers", "following_url": "https://api.github.com/users/Yann21/following{/other_user}", "gists_url": "https://api.github.com/users/Yann21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Yann21", "id": 35955430, "login": "Yann21", "node_id": "MDQ6VXNlcjM1OTU1NDMw", "organizations_url": "https://api.github.com/users/Yann21/orgs", "received_events_url": "https://api.github.com/users/Yann21/received_events", "repos_url": "https://api.github.com/users/Yann21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Yann21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Yann21/subscriptions", "type": "User", "url": "https://api.github.com/users/Yann21" }
Correct a typo
https://api.github.com/repos/huggingface/datasets/issues/3007/events
null
https://api.github.com/repos/huggingface/datasets/issues/3007/labels{/name}
2021-10-04T06:15:47Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3007.diff", "html_url": "https://github.com/huggingface/datasets/pull/3007", "merged_at": "2021-10-04T09:27:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/3007.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3007" }
1,014,775,450
[]
https://api.github.com/repos/huggingface/datasets/issues/3007
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2021-10-04T09:27:57Z
https://github.com/huggingface/datasets/pull/3007
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3007/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3006/comments
https://api.github.com/repos/huggingface/datasets/issues/3006/timeline
2021-10-04T09:07:58Z
null
null
PR_kwDODunzps4snsBm
closed
[]
false
3,006
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix Windows paths in CommonLanguage dataset
https://api.github.com/repos/huggingface/datasets/issues/3006/events
null
https://api.github.com/repos/huggingface/datasets/issues/3006/labels{/name}
2021-10-04T06:08:58Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3006.diff", "html_url": "https://github.com/huggingface/datasets/pull/3006", "merged_at": "2021-10-04T09:07:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/3006.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3006" }
1,014,770,821
[]
https://api.github.com/repos/huggingface/datasets/issues/3006
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Minor fix in CommonLanguage dataset for Windows pathname component separator. Related to #2989.
2021-10-04T09:07:58Z
https://github.com/huggingface/datasets/pull/3006
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3006/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3005/comments
https://api.github.com/repos/huggingface/datasets/issues/3005/timeline
2021-10-04T08:46:13Z
null
completed
I_kwDODunzps48ec18
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
3,005
{ "avatar_url": "https://avatars.githubusercontent.com/u/22641583?v=4", "events_url": "https://api.github.com/users/DrMatters/events{/privacy}", "followers_url": "https://api.github.com/users/DrMatters/followers", "following_url": "https://api.github.com/users/DrMatters/following{/other_user}", "gists_url": "https://api.github.com/users/DrMatters/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/DrMatters", "id": 22641583, "login": "DrMatters", "node_id": "MDQ6VXNlcjIyNjQxNTgz", "organizations_url": "https://api.github.com/users/DrMatters/orgs", "received_events_url": "https://api.github.com/users/DrMatters/received_events", "repos_url": "https://api.github.com/users/DrMatters/repos", "site_admin": false, "starred_url": "https://api.github.com/users/DrMatters/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DrMatters/subscriptions", "type": "User", "url": "https://api.github.com/users/DrMatters" }
DatasetDict.filter and Dataset.filter crashes with any "fn_kwargs" argument
https://api.github.com/repos/huggingface/datasets/issues/3005/events
null
https://api.github.com/repos/huggingface/datasets/issues/3005/labels{/name}
2021-10-04T00:49:29Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,014,615,420
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/3005
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug The ".filter" method of DatasetDict or Dataset objects fails when passing any "fn_kwargs" argument ## Steps to reproduce the bug ```python import datasets example_dataset = datasets.Dataset.from_dict({"a": {1, 2, 3, 4}}) def filter_value(example, value): return example['a'] == value filtered = example_dataset.filter(filter_value, fn_kwargs={'value': 3}) ``` ## Expected results `filtered` is a dataset containing {"a": {3}} ## Actual results > Traceback (most recent call last): > File "C:\Users\qsemi\Documents\git\nlp_experiments\gpt_celebrity\src\test_faulty_filter.py", line 8, in <module> > filtered = example_dataset.filter(filter_value, fn_kwargs={'value': 3}) > File "C:\Users\qsemi\miniconda3\envs\main\lib\site-packages\datasets\arrow_dataset.py", line 185, in wrapper > out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) > File "C:\Users\qsemi\miniconda3\envs\main\lib\site-packages\datasets\fingerprint.py", line 398, in wrapper > out = func(self, *args, **kwargs) > File "C:\Users\qsemi\miniconda3\envs\main\lib\site-packages\datasets\arrow_dataset.py", line 2169, in filter > indices = self.map( > File "C:\Users\qsemi\miniconda3\envs\main\lib\site-packages\datasets\arrow_dataset.py", line 1686, in map > return self._map_single( > File "C:\Users\qsemi\miniconda3\envs\main\lib\site-packages\datasets\arrow_dataset.py", line 185, in wrapper > out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) > File "C:\Users\qsemi\miniconda3\envs\main\lib\site-packages\datasets\fingerprint.py", line 398, in wrapper > out = func(self, *args, **kwargs) > File "C:\Users\qsemi\miniconda3\envs\main\lib\site-packages\datasets\arrow_dataset.py", line 2048, in _map_single > batch = apply_function_on_filtered_inputs( > File "C:\Users\qsemi\miniconda3\envs\main\lib\site-packages\datasets\arrow_dataset.py", line 1939, in apply_function_on_filtered_inputs > function(*fn_args, effective_indices, **fn_kwargs) if with_indices else function(*fn_args, **fn_kwargs) > TypeError: get_indices_from_mask_function() got an unexpected keyword argument 'value' ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.1 - Platform: Windows-10-10.0.19042-SP0 - Python version: 3.9.7 - PyArrow version: 5.0.0
2021-10-11T10:18:01Z
https://github.com/huggingface/datasets/issues/3005
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3005/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/3004/comments
https://api.github.com/repos/huggingface/datasets/issues/3004/timeline
2021-10-13T13:37:01Z
null
null
PR_kwDODunzps4smfPF
closed
[]
false
3,004
{ "avatar_url": "https://avatars.githubusercontent.com/u/1626984?v=4", "events_url": "https://api.github.com/users/iliaschalkidis/events{/privacy}", "followers_url": "https://api.github.com/users/iliaschalkidis/followers", "following_url": "https://api.github.com/users/iliaschalkidis/following{/other_user}", "gists_url": "https://api.github.com/users/iliaschalkidis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/iliaschalkidis", "id": 1626984, "login": "iliaschalkidis", "node_id": "MDQ6VXNlcjE2MjY5ODQ=", "organizations_url": "https://api.github.com/users/iliaschalkidis/orgs", "received_events_url": "https://api.github.com/users/iliaschalkidis/received_events", "repos_url": "https://api.github.com/users/iliaschalkidis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/iliaschalkidis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iliaschalkidis/subscriptions", "type": "User", "url": "https://api.github.com/users/iliaschalkidis" }
LexGLUE: A Benchmark Dataset for Legal Language Understanding in English.
https://api.github.com/repos/huggingface/datasets/issues/3004/events
null
https://api.github.com/repos/huggingface/datasets/issues/3004/labels{/name}
2021-10-03T10:03:25Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3004.diff", "html_url": "https://github.com/huggingface/datasets/pull/3004", "merged_at": "2021-10-13T13:37:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/3004.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3004" }
1,014,336,617
[]
https://api.github.com/repos/huggingface/datasets/issues/3004
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Inspired by the recent widespread use of the GLUE multi-task benchmark NLP dataset (Wang et al., 2018), the subsequent more difficult SuperGLUE (Wang et al., 2019), other previous multi-task NLP benchmarks (Conneau and Kiela, 2018; McCann et al., 2018), and similar initiatives in other domains (Peng et al., 2019), we introduce the Legal General Language Understanding Evaluation (LexGLUE) benchmark, a benchmark dataset to evaluate the performance of NLP methods in legal tasks. LexGLUE is based on seven existing legal NLP datasets, selected using criteria largely from SuperGLUE. As in GLUE and SuperGLUE (Wang et al., 2019b,a), one of our goals is to push towards generic (or ‘foundation’) models that can cope with multiple NLP tasks, in our case legal NLP tasks possibly with limited task-specific fine-tuning. Another goal is to provide a convenient and informative entry point for NLP researchers and practitioners wishing to explore or develop methods for legalNLP. Having these goals in mind, the datasets we include in LexGLUE and the tasks they address have been simplified in several ways to make it easier for newcomers and generic models to address all tasks. LexGLUE benchmark is accompanied by experimental infrastructure that relies on Hugging Face Transformers library and resides at: https://github.com/coastalcph/lex-glue.
2021-10-13T13:37:02Z
https://github.com/huggingface/datasets/pull/3004
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/3004/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3003/comments
https://api.github.com/repos/huggingface/datasets/issues/3003/timeline
2021-10-04T09:27:01Z
null
null
PR_kwDODunzps4smExP
closed
[]
false
3,003
{ "avatar_url": "https://avatars.githubusercontent.com/u/227350?v=4", "events_url": "https://api.github.com/users/jimregan/events{/privacy}", "followers_url": "https://api.github.com/users/jimregan/followers", "following_url": "https://api.github.com/users/jimregan/following{/other_user}", "gists_url": "https://api.github.com/users/jimregan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jimregan", "id": 227350, "login": "jimregan", "node_id": "MDQ6VXNlcjIyNzM1MA==", "organizations_url": "https://api.github.com/users/jimregan/orgs", "received_events_url": "https://api.github.com/users/jimregan/received_events", "repos_url": "https://api.github.com/users/jimregan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jimregan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jimregan/subscriptions", "type": "User", "url": "https://api.github.com/users/jimregan" }
common_language: Fix license in README.md
https://api.github.com/repos/huggingface/datasets/issues/3003/events
null
https://api.github.com/repos/huggingface/datasets/issues/3003/labels{/name}
2021-10-02T18:47:37Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3003.diff", "html_url": "https://github.com/huggingface/datasets/pull/3003", "merged_at": "2021-10-04T09:27:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/3003.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3003" }
1,014,137,933
[]
https://api.github.com/repos/huggingface/datasets/issues/3003
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
...it's correct elsewhere
2021-10-04T09:27:01Z
https://github.com/huggingface/datasets/pull/3003
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3003/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3002/comments
https://api.github.com/repos/huggingface/datasets/issues/3002/timeline
2021-10-13T09:03:23Z
null
null
PR_kwDODunzps4smCNO
closed
[]
false
3,002
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Remove a reference to the open Arrow file when deleting a TF dataset created with to_tf_dataset
https://api.github.com/repos/huggingface/datasets/issues/3002/events
null
https://api.github.com/repos/huggingface/datasets/issues/3002/labels{/name}
2021-10-02T17:44:09Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3002.diff", "html_url": "https://github.com/huggingface/datasets/pull/3002", "merged_at": "2021-10-13T09:03:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/3002.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3002" }
1,014,120,524
[]
https://api.github.com/repos/huggingface/datasets/issues/3002
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This [comment](https://github.com/huggingface/datasets/issues/2934#issuecomment-922970919) explains the issue. This PR fixes that with a `weakref` callback, and additionally: * renames `TensorflowDatasetMixIn` to `TensorflowDatasetMixin` for consistency * correctly indents `TensorflowDatasetMixin`'s docstring * replaces `tf.data.AUTOTUNE` with `tf.data.experimental.AUTOTUNE` (we support TF>=2.2 according to the [setup.py](https://github.com/huggingface/datasets/blob/fc46bba66ba4f432cc10501c16a677112e13984c/setup.py#L188) and `AUTOTUNE` has been moved to the experimental part of `tf.data` in 1.X if I'm not mistaken) Fixes #2934
2021-10-13T11:48:00Z
https://github.com/huggingface/datasets/pull/3002
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3002/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3001/comments
https://api.github.com/repos/huggingface/datasets/issues/3001/timeline
2021-10-04T09:26:12Z
null
null
PR_kwDODunzps4sl0BY
closed
[]
false
3,001
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Fix cast to Python scalar in Matthews Correlation metric
https://api.github.com/repos/huggingface/datasets/issues/3001/events
null
https://api.github.com/repos/huggingface/datasets/issues/3001/labels{/name}
2021-10-02T11:44:59Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3001.diff", "html_url": "https://github.com/huggingface/datasets/pull/3001", "merged_at": "2021-10-04T09:26:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/3001.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3001" }
1,014,024,982
[]
https://api.github.com/repos/huggingface/datasets/issues/3001
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This PR is motivated by issue #2964. The Matthews Correlation metric relies on sklearn's `matthews_corrcoef` function to compute the result. This function returns either `float` or `np.float64` (see the [source](https://github.com/scikit-learn/scikit-learn/blob/844b4be24d20fc42cc13b957374c718956a0db39/sklearn/metrics/_classification.py#L906-L909)). Obviously, calling `.item()` on the float value will fail, so I'm fixing this with the built-in `float()` function, which covers both cases. Surprisingly, on my machine, casting `np.float64` to a Python scalar with `float()` is even faster than with the `.item()` method.
2021-10-04T09:54:04Z
https://github.com/huggingface/datasets/pull/3001
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3001/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/3000/comments
https://api.github.com/repos/huggingface/datasets/issues/3000/timeline
2021-10-01T17:54:23Z
null
null
PR_kwDODunzps4skusL
closed
[]
false
3,000
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix json loader when conversion not implemented
https://api.github.com/repos/huggingface/datasets/issues/3000/events
null
https://api.github.com/repos/huggingface/datasets/issues/3000/labels{/name}
2021-10-01T17:47:22Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/3000.diff", "html_url": "https://github.com/huggingface/datasets/pull/3000", "merged_at": "2021-10-01T17:54:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/3000.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3000" }
1,013,613,219
[]
https://api.github.com/repos/huggingface/datasets/issues/3000
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Sometimes the arrow json parser fails if the `block_size` is too small and returns an `ArrowNotImplementedError: JSON conversion to struct...` error. By increasing the block size it makes it work again. Hopefully it should help with https://github.com/huggingface/datasets/issues/2799 I tried with the file mentioned in the issue and it worked for me cc @lewtun can you try again from this branch ?
2021-10-01T18:05:00Z
https://github.com/huggingface/datasets/pull/3000
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/3000/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2999/comments
https://api.github.com/repos/huggingface/datasets/issues/2999/timeline
2021-10-01T16:34:55Z
null
null
PR_kwDODunzps4skgCm
closed
[]
false
2,999
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Set trivia_qa writer batch size
https://api.github.com/repos/huggingface/datasets/issues/2999/events
null
https://api.github.com/repos/huggingface/datasets/issues/2999/labels{/name}
2021-10-01T16:23:26Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2999.diff", "html_url": "https://github.com/huggingface/datasets/pull/2999", "merged_at": "2021-10-01T16:34:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/2999.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2999" }
1,013,536,933
[]
https://api.github.com/repos/huggingface/datasets/issues/2999
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Save some RAM when generating trivia_qa
2021-10-01T16:34:55Z
https://github.com/huggingface/datasets/pull/2999
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2999/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2998/comments
https://api.github.com/repos/huggingface/datasets/issues/2998/timeline
null
null
null
I_kwDODunzps48ZtfH
open
[]
null
2,998
{ "avatar_url": "https://avatars.githubusercontent.com/u/54274249?v=4", "events_url": "https://api.github.com/users/pya25/events{/privacy}", "followers_url": "https://api.github.com/users/pya25/followers", "following_url": "https://api.github.com/users/pya25/following{/other_user}", "gists_url": "https://api.github.com/users/pya25/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/pya25", "id": 54274249, "login": "pya25", "node_id": "MDQ6VXNlcjU0Mjc0MjQ5", "organizations_url": "https://api.github.com/users/pya25/orgs", "received_events_url": "https://api.github.com/users/pya25/received_events", "repos_url": "https://api.github.com/users/pya25/repos", "site_admin": false, "starred_url": "https://api.github.com/users/pya25/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pya25/subscriptions", "type": "User", "url": "https://api.github.com/users/pya25" }
cannot shuffle dataset loaded from disk
https://api.github.com/repos/huggingface/datasets/issues/2998/events
null
https://api.github.com/repos/huggingface/datasets/issues/2998/labels{/name}
2021-10-01T13:49:52Z
null
false
null
null
1,013,372,871
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2998
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug dataset loaded from disk cannot be shuffled. ## Steps to reproduce the bug ``` my_dataset = load_from_disk('s3://my_file/validate', fs=s3) sample = my_dataset.select(range(100)).shuffle(seed=1234) ``` ## Actual results ``` sample = my_dataset .select(range(100)).shuffle(seed=1234) File "/home/ubuntu/anaconda3/envs/pytorch_p37/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 185, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/ubuntu/anaconda3/envs/pytorch_p37/lib/python3.7/site-packages/datasets/fingerprint.py", line 398, in wrapper out = func(self, *args, **kwargs) File "/home/ubuntu/anaconda3/envs/pytorch_p37/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 2494, in shuffle new_fingerprint=new_fingerprint, File "/home/ubuntu/anaconda3/envs/pytorch_p37/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 185, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/ubuntu/anaconda3/envs/pytorch_p37/lib/python3.7/site-packages/datasets/fingerprint.py", line 398, in wrapper out = func(self, *args, **kwargs) File "/home/ubuntu/anaconda3/envs/pytorch_p37/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 2303, in select tmp_file = tempfile.NamedTemporaryFile("wb", dir=os.path.dirname(indices_cache_file_name), delete=False) File "/home/ubuntu/anaconda3/envs/pytorch_p37/lib/python3.7/tempfile.py", line 547, in NamedTemporaryFile (fd, name) = _mkstemp_inner(dir, prefix, suffix, flags, output_type) File "/home/ubuntu/anaconda3/envs/pytorch_p37/lib/python3.7/tempfile.py", line 258, in _mkstemp_inner fd = _os.open(file, flags, 0o600) FileNotFoundError: [Errno 2] No such file or directory: '/tmp/tmpnnu5uhnx/my_file/validate/tmpy76d70g4' ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.1 - Python version: 3.7 - PyArrow version: 5.0.0
2021-10-01T13:49:52Z
https://github.com/huggingface/datasets/issues/2998
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2998/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2997/comments
https://api.github.com/repos/huggingface/datasets/issues/2997/timeline
2021-10-01T13:54:34Z
null
completed
I_kwDODunzps48ZUY1
closed
[]
null
2,997
{ "avatar_url": "https://avatars.githubusercontent.com/u/63367770?v=4", "events_url": "https://api.github.com/users/marshmellow77/events{/privacy}", "followers_url": "https://api.github.com/users/marshmellow77/followers", "following_url": "https://api.github.com/users/marshmellow77/following{/other_user}", "gists_url": "https://api.github.com/users/marshmellow77/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/marshmellow77", "id": 63367770, "login": "marshmellow77", "node_id": "MDQ6VXNlcjYzMzY3Nzcw", "organizations_url": "https://api.github.com/users/marshmellow77/orgs", "received_events_url": "https://api.github.com/users/marshmellow77/received_events", "repos_url": "https://api.github.com/users/marshmellow77/repos", "site_admin": false, "starred_url": "https://api.github.com/users/marshmellow77/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/marshmellow77/subscriptions", "type": "User", "url": "https://api.github.com/users/marshmellow77" }
Dataset has incorrect labels
https://api.github.com/repos/huggingface/datasets/issues/2997/events
null
https://api.github.com/repos/huggingface/datasets/issues/2997/labels{/name}
2021-10-01T12:09:06Z
null
false
null
null
1,013,270,069
[]
https://api.github.com/repos/huggingface/datasets/issues/2997
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
The dataset https://huggingface.co/datasets/turkish_product_reviews has incorrect labels - all reviews are labelled with "1" (positive sentiment). None of the reviews is labelled with "0". See screenshot attached: ![Capture](https://user-images.githubusercontent.com/63367770/135617428-14ce0b27-5208-4e66-a3ee-71542e3257b4.PNG)
2021-10-01T15:32:00Z
https://github.com/huggingface/datasets/issues/2997
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2997/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2996/comments
https://api.github.com/repos/huggingface/datasets/issues/2996/timeline
2021-10-04T08:48:13Z
null
null
PR_kwDODunzps4sjrP6
closed
[]
false
2,996
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Remove all query parameters when extracting protocol
https://api.github.com/repos/huggingface/datasets/issues/2996/events
null
https://api.github.com/repos/huggingface/datasets/issues/2996/labels{/name}
2021-10-01T12:05:34Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2996.diff", "html_url": "https://github.com/huggingface/datasets/pull/2996", "merged_at": "2021-10-04T08:48:13Z", "patch_url": "https://github.com/huggingface/datasets/pull/2996.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2996" }
1,013,266,373
[]
https://api.github.com/repos/huggingface/datasets/issues/2996
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix `_get_extraction_protocol` to remove all query parameters, like `?raw=true`, `?dl=1`,...
2021-10-04T08:48:13Z
https://github.com/huggingface/datasets/pull/2996
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2996/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2995/comments
https://api.github.com/repos/huggingface/datasets/issues/2995/timeline
2021-10-01T10:04:10Z
null
null
PR_kwDODunzps4sjThd
closed
[]
false
2,995
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix trivia_qa unfiltered
https://api.github.com/repos/huggingface/datasets/issues/2995/events
null
https://api.github.com/repos/huggingface/datasets/issues/2995/labels{/name}
2021-10-01T09:53:43Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2995.diff", "html_url": "https://github.com/huggingface/datasets/pull/2995", "merged_at": "2021-10-01T10:04:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/2995.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2995" }
1,013,143,868
[]
https://api.github.com/repos/huggingface/datasets/issues/2995
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix https://github.com/huggingface/datasets/issues/2993
2021-10-01T10:04:11Z
https://github.com/huggingface/datasets/pull/2995
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2995/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2994/comments
https://api.github.com/repos/huggingface/datasets/issues/2994/timeline
2021-10-01T15:53:16Z
null
null
PR_kwDODunzps4si4I2
closed
[]
false
2,994
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix loading compressed CSV without streaming
https://api.github.com/repos/huggingface/datasets/issues/2994/events
null
https://api.github.com/repos/huggingface/datasets/issues/2994/labels{/name}
2021-10-01T07:28:59Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2994.diff", "html_url": "https://github.com/huggingface/datasets/pull/2994", "merged_at": "2021-10-01T15:53:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/2994.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2994" }
1,013,000,475
[]
https://api.github.com/repos/huggingface/datasets/issues/2994
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
When implementing support to stream CSV files (https://github.com/huggingface/datasets/commit/ad489d4597381fc2d12c77841642cbeaecf7a2e0#diff-6f60f8d0552b75be8b3bfd09994480fd60dcd4e7eb08d02f721218c3acdd2782), a regression was introduced preventing loading compressed CSV files in non-streaming mode. This PR fixes it, allowing loading compressed/uncompressed CSV files in streaming/non-streaming mode. Fix #2977.
2021-10-01T15:53:16Z
https://github.com/huggingface/datasets/pull/2994
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2994/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2993/comments
https://api.github.com/repos/huggingface/datasets/issues/2993/timeline
2021-10-01T19:07:22Z
null
completed
I_kwDODunzps48XJ3J
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
2,993
{ "avatar_url": "https://avatars.githubusercontent.com/u/16107619?v=4", "events_url": "https://api.github.com/users/VictorSanh/events{/privacy}", "followers_url": "https://api.github.com/users/VictorSanh/followers", "following_url": "https://api.github.com/users/VictorSanh/following{/other_user}", "gists_url": "https://api.github.com/users/VictorSanh/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/VictorSanh", "id": 16107619, "login": "VictorSanh", "node_id": "MDQ6VXNlcjE2MTA3NjE5", "organizations_url": "https://api.github.com/users/VictorSanh/orgs", "received_events_url": "https://api.github.com/users/VictorSanh/received_events", "repos_url": "https://api.github.com/users/VictorSanh/repos", "site_admin": false, "starred_url": "https://api.github.com/users/VictorSanh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/VictorSanh/subscriptions", "type": "User", "url": "https://api.github.com/users/VictorSanh" }
Can't download `trivia_qa/unfiltered`
https://api.github.com/repos/huggingface/datasets/issues/2993/events
null
https://api.github.com/repos/huggingface/datasets/issues/2993/labels{/name}
2021-09-30T23:00:18Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
1,012,702,665
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2993
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Describe the bug For some reason, I can't download `trivia_qa/unfilted`. A file seems to be missing... I am able to see it fine though the viewer tough... ## Steps to reproduce the bug ```python >>> from datasets import load_dataset >>> load_dataset("trivia_qa", "unfiltered") Downloading and preparing dataset trivia_qa/unfiltered (download: 3.07 GiB, generated: 27.23 GiB, post-processed: Unknown size, total: 30.30 GiB) to /gpfsscratch/rech/six/commun/datasets/trivia_qa/unfiltered/1.1.0/9977a5d6f72acfd92f587de052403e8138b43bb0d1ce595016c3baf7e14deba6... Traceback (most recent call last): File "/gpfswork/rech/six/commun/modules/datasets_modules/datasets/trivia_qa/9977a5d6f72acfd92f587de052403e8138b43bb0d1ce595016c3baf7e14deba6/trivia_qa.py", line 251, in _add_context with open(os.path.join(file_dir, fname), encoding="utf-8") as f: FileNotFoundError: [Errno 2] No such file or directory: '/gpfsscratch/rech/six/commun/datasets/downloads/extracted/9fcb7eddc6afd46fd074af3c5128931dfe4b548f933c925a23847faf4c1995ad/evidence/wikipedia/Peanuts.txt' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/gpfswork/rech/six/commun/conda/victor/lib/python3.7/site-packages/datasets/load.py", line 852, in load_dataset use_auth_token=use_auth_token, File "/gpfswork/rech/six/commun/conda/victor/lib/python3.7/site-packages/datasets/builder.py", line 616, in download_and_prepare dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "/gpfswork/rech/six/commun/conda/victor/lib/python3.7/site-packages/datasets/builder.py", line 693, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/gpfswork/rech/six/commun/conda/victor/lib/python3.7/site-packages/datasets/builder.py", line 1107, in _prepare_split disable=bool(logging.get_verbosity() == logging.NOTSET), File "/gpfswork/rech/six/commun/conda/victor/lib/python3.7/site-packages/tqdm/std.py", line 1133, in __iter__ for obj in iterable: File "/gpfswork/rech/six/commun/modules/datasets_modules/datasets/trivia_qa/9977a5d6f72acfd92f587de052403e8138b43bb0d1ce595016c3baf7e14deba6/trivia_qa.py", line 303, in _generate_examples example = parse_example(article) File "/gpfswork/rech/six/commun/modules/datasets_modules/datasets/trivia_qa/9977a5d6f72acfd92f587de052403e8138b43bb0d1ce595016c3baf7e14deba6/trivia_qa.py", line 274, in parse_example _add_context(article.get("EntityPages", []), "WikiContext", wiki_dir), File "/gpfswork/rech/six/commun/modules/datasets_modules/datasets/trivia_qa/9977a5d6f72acfd92f587de052403e8138b43bb0d1ce595016c3baf7e14deba6/trivia_qa.py", line 253, in _add_context except (IOError, datasets.Value("errors").NotFoundError): File "<string>", line 5, in __init__ File "/gpfswork/rech/six/commun/conda/victor/lib/python3.7/site-packages/datasets/features.py", line 265, in __post_init__ self.pa_type = string_to_arrow(self.dtype) File "/gpfswork/rech/six/commun/conda/victor/lib/python3.7/site-packages/datasets/features.py", line 134, in string_to_arrow f"Neither {datasets_dtype} nor {datasets_dtype + '_'} seems to be a pyarrow data type. " ValueError: Neither errors nor errors_ seems to be a pyarrow data type. Please make sure to use a correct data type, see: https://arrow.apache.org/docs/python/api/datatypes.html#factory-functions ``` ## Expected results I am able to load another subset (`rc`), but unable to load. I am not sure why the try/except doesn't catch it... https://github.com/huggingface/datasets/blob/9675a5a1e7b99a86f9c250f6ea5fa5d1e6d5cc7d/datasets/trivia_qa/trivia_qa.py#L253 ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.11.0 - Platform: Linux-4.18.0-147.51.2.el8_1.x86_64-x86_64-with-redhat-8.1-Ootpa - Python version: 3.7.10 - PyArrow version: 3.0.0
2021-10-01T19:07:23Z
https://github.com/huggingface/datasets/issues/2993
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2993/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2992/comments
https://api.github.com/repos/huggingface/datasets/issues/2992/timeline
2021-10-01T14:17:38Z
null
null
PR_kwDODunzps4sg4ZP
closed
[]
false
2,992
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix f1 metric with None average
https://api.github.com/repos/huggingface/datasets/issues/2992/events
null
https://api.github.com/repos/huggingface/datasets/issues/2992/labels{/name}
2021-09-30T15:31:57Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2992.diff", "html_url": "https://github.com/huggingface/datasets/pull/2992", "merged_at": "2021-10-01T14:17:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/2992.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2992" }
1,012,325,594
[]
https://api.github.com/repos/huggingface/datasets/issues/2992
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix #2979.
2021-10-01T14:17:39Z
https://github.com/huggingface/datasets/pull/2992
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2992/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2991/comments
https://api.github.com/repos/huggingface/datasets/issues/2991/timeline
null
null
null
I_kwDODunzps48VI_n
open
[]
null
2,991
{ "avatar_url": "https://avatars.githubusercontent.com/u/55560583?v=4", "events_url": "https://api.github.com/users/SaulLu/events{/privacy}", "followers_url": "https://api.github.com/users/SaulLu/followers", "following_url": "https://api.github.com/users/SaulLu/following{/other_user}", "gists_url": "https://api.github.com/users/SaulLu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SaulLu", "id": 55560583, "login": "SaulLu", "node_id": "MDQ6VXNlcjU1NTYwNTgz", "organizations_url": "https://api.github.com/users/SaulLu/orgs", "received_events_url": "https://api.github.com/users/SaulLu/received_events", "repos_url": "https://api.github.com/users/SaulLu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SaulLu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SaulLu/subscriptions", "type": "User", "url": "https://api.github.com/users/SaulLu" }
add docmentation for the `Unix style pattern` matching feature that can be leverage for `data_files` into `load_dataset`
https://api.github.com/repos/huggingface/datasets/issues/2991/events
null
https://api.github.com/repos/huggingface/datasets/issues/2991/labels{/name}
2021-09-30T13:22:01Z
null
false
null
null
1,012,174,823
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2991
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Unless I'm mistaken, it seems that in the new documentation it is no longer mentioned that you can use Unix style pattern matching in the `data_files` argument of the `load_dataset` method. This feature was mentioned [here](https://huggingface.co/docs/datasets/loading_datasets.html#from-a-community-dataset-on-the-hugging-face-hub) in the previous documentation. I'd love to hear your opinion @lhoestq , @albertvillanova and @stevhliu
2021-09-30T13:22:01Z
https://github.com/huggingface/datasets/issues/2991
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2991/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2990/comments
https://api.github.com/repos/huggingface/datasets/issues/2990/timeline
2021-10-01T13:57:46Z
null
null
PR_kwDODunzps4sgLt5
closed
[]
false
2,990
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Make Dataset.map accept list of np.array
https://api.github.com/repos/huggingface/datasets/issues/2990/events
null
https://api.github.com/repos/huggingface/datasets/issues/2990/labels{/name}
2021-09-30T12:08:54Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2990.diff", "html_url": "https://github.com/huggingface/datasets/pull/2990", "merged_at": "2021-10-01T13:57:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/2990.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2990" }
1,012,097,418
[]
https://api.github.com/repos/huggingface/datasets/issues/2990
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix #2987.
2021-10-01T13:57:46Z
https://github.com/huggingface/datasets/pull/2990
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2990/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2989/comments
https://api.github.com/repos/huggingface/datasets/issues/2989/timeline
2021-10-01T17:00:03Z
null
null
PR_kwDODunzps4sdlt1
closed
[]
false
2,989
{ "avatar_url": "https://avatars.githubusercontent.com/u/26864830?v=4", "events_url": "https://api.github.com/users/anton-l/events{/privacy}", "followers_url": "https://api.github.com/users/anton-l/followers", "following_url": "https://api.github.com/users/anton-l/following{/other_user}", "gists_url": "https://api.github.com/users/anton-l/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/anton-l", "id": 26864830, "login": "anton-l", "node_id": "MDQ6VXNlcjI2ODY0ODMw", "organizations_url": "https://api.github.com/users/anton-l/orgs", "received_events_url": "https://api.github.com/users/anton-l/received_events", "repos_url": "https://api.github.com/users/anton-l/repos", "site_admin": false, "starred_url": "https://api.github.com/users/anton-l/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/anton-l/subscriptions", "type": "User", "url": "https://api.github.com/users/anton-l" }
Add CommonLanguage
https://api.github.com/repos/huggingface/datasets/issues/2989/events
null
https://api.github.com/repos/huggingface/datasets/issues/2989/labels{/name}
2021-09-29T17:21:30Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2989.diff", "html_url": "https://github.com/huggingface/datasets/pull/2989", "merged_at": "2021-10-01T17:00:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/2989.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2989" }
1,011,220,375
[]
https://api.github.com/repos/huggingface/datasets/issues/2989
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This PR adds the Common Language dataset (https://zenodo.org/record/5036977) The dataset is intended for language-identification speech classifiers and is already used by models on the Hub: * https://huggingface.co/speechbrain/lang-id-commonlanguage_ecapa * https://huggingface.co/anton-l/wav2vec2-base-langid cc @patrickvonplaten
2021-10-01T17:36:39Z
https://github.com/huggingface/datasets/pull/2989
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 1, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2989/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2988/comments
https://api.github.com/repos/huggingface/datasets/issues/2988/timeline
2022-04-10T14:49:49Z
null
completed
I_kwDODunzps48ROTx
closed
[]
null
2,988
{ "avatar_url": "https://avatars.githubusercontent.com/u/79165106?v=4", "events_url": "https://api.github.com/users/dorost1234/events{/privacy}", "followers_url": "https://api.github.com/users/dorost1234/followers", "following_url": "https://api.github.com/users/dorost1234/following{/other_user}", "gists_url": "https://api.github.com/users/dorost1234/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dorost1234", "id": 79165106, "login": "dorost1234", "node_id": "MDQ6VXNlcjc5MTY1MTA2", "organizations_url": "https://api.github.com/users/dorost1234/orgs", "received_events_url": "https://api.github.com/users/dorost1234/received_events", "repos_url": "https://api.github.com/users/dorost1234/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dorost1234/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dorost1234/subscriptions", "type": "User", "url": "https://api.github.com/users/dorost1234" }
IndexError: Invalid key: 14 is out of bounds for size 0
https://api.github.com/repos/huggingface/datasets/issues/2988/events
null
https://api.github.com/repos/huggingface/datasets/issues/2988/labels{/name}
2021-09-29T16:04:24Z
null
false
null
null
1,011,148,017
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2988
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug A clear and concise description of what the bug is. Hi. I am trying to implement stochastic weighted averaging optimizer with transformer library as described here https://pytorch.org/blog/pytorch-1.6-now-includes-stochastic-weight-averaging/ , for this I am using a run_clm.py codes which is working fine before adding SWA optimizer, the moment I modify the model with `swa_model = AveragedModel(model)` in this script, I am getting the below error, since I am NOT touching the dataloader part, I am confused why this is occurring, I very much appreciate your opinion on this @lhoestq ## Steps to reproduce the bug ``` Traceback (most recent call last): File "run_clm.py", line 723, in <module> main() File "run_clm.py", line 669, in main train_result = trainer.train(resume_from_checkpoint=checkpoint) File "/user/dara/libs/anaconda3/envs/success/lib/python3.7/site-packages/transformers/trainer.py", line 1258, in train for step, inputs in enumerate(epoch_iterator): File "/user/dara/libs/anaconda3/envs/success/lib/python3.7/site-packages/torch/utils/data/dataloader.py", line 435, in __next__ data = self._next_data() File "/user/dara/libs/anaconda3/envs/success/lib/python3.7/site-packages/torch/utils/data/dataloader.py", line 475, in _next_data data = self._dataset_fetcher.fetch(index) # may raise StopIteration File "/user/dara/libs/anaconda3/envs/success/lib/python3.7/site-packages/torch/utils/data/_utils/fetch.py", line 44, in fetch data = [self.dataset[idx] for idx in possibly_batched_index] File "/user/dara/libs/anaconda3/envs/success/lib/python3.7/site-packages/torch/utils/data/_utils/fetch.py", line 44, in <listcomp> data = [self.dataset[idx] for idx in possibly_batched_index] File "/user/dara/libs/anaconda3/envs/success/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 1530, in __getitem__ format_kwargs=self._format_kwargs, File "/user/dara/libs/anaconda3/envs/success/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 1517, in _getitem pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None) File "/user/dara/libs/anaconda3/envs/success/lib/python3.7/site-packages/datasets/formatting/formatting.py", line 368, in query_table _check_valid_index_key(key, size) File "/user/dara/libs/anaconda3/envs/success/lib/python3.7/site-packages/datasets/formatting/formatting.py", line 311, in _check_valid_index_key raise IndexError(f"Invalid key: {key} is out of bounds for size {size}") IndexError: Invalid key: 14 is out of bounds for size 0 ``` ## Expected results not getting the index error ## Actual results Please see the above ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: datasets 1.12.1 - Platform: linux - Python version: 3.7.11 - PyArrow version: 5.0.0
2022-04-10T14:49:49Z
https://github.com/huggingface/datasets/issues/2988
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2988/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2987/comments
https://api.github.com/repos/huggingface/datasets/issues/2987/timeline
2021-10-01T13:57:45Z
null
completed
I_kwDODunzps48Qwjd
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,987
{ "avatar_url": "https://avatars.githubusercontent.com/u/48327001?v=4", "events_url": "https://api.github.com/users/NielsRogge/events{/privacy}", "followers_url": "https://api.github.com/users/NielsRogge/followers", "following_url": "https://api.github.com/users/NielsRogge/following{/other_user}", "gists_url": "https://api.github.com/users/NielsRogge/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NielsRogge", "id": 48327001, "login": "NielsRogge", "node_id": "MDQ6VXNlcjQ4MzI3MDAx", "organizations_url": "https://api.github.com/users/NielsRogge/orgs", "received_events_url": "https://api.github.com/users/NielsRogge/received_events", "repos_url": "https://api.github.com/users/NielsRogge/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NielsRogge/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NielsRogge/subscriptions", "type": "User", "url": "https://api.github.com/users/NielsRogge" }
ArrowInvalid: Can only convert 1-dimensional array values
https://api.github.com/repos/huggingface/datasets/issues/2987/events
null
https://api.github.com/repos/huggingface/datasets/issues/2987/labels{/name}
2021-09-29T14:18:52Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,011,026,141
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2987
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug For the ViT and LayoutLMv2 demo notebooks in my [Transformers-Tutorials repo](https://github.com/NielsRogge/Transformers-Tutorials), people reported an ArrowInvalid issue after applying the following function to a Dataset: ``` def preprocess_data(examples): images = [Image.open(path).convert("RGB") for path in examples['image_path']] words = examples['words'] boxes = examples['bboxes'] word_labels = examples['ner_tags'] encoded_inputs = processor(images, words, boxes=boxes, word_labels=word_labels, padding="max_length", truncation=True) return encoded_inputs ``` ``` Full trace: --------------------------------------------------------------------------- ArrowInvalid Traceback (most recent call last) <ipython-input-8-0fc3efc6f0c2> in <module>() 27 28 train_dataset = datasets['train'].map(preprocess_data, batched=True, remove_columns=datasets['train'].column_names, ---> 29 features=features) 30 test_dataset = datasets['test'].map(preprocess_data, batched=True, remove_columns=datasets['test'].column_names, 31 features=features) 13 frames /usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc) 1701 new_fingerprint=new_fingerprint, 1702 disable_tqdm=disable_tqdm, -> 1703 desc=desc, 1704 ) 1705 else: /usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py in wrapper(*args, **kwargs) 183 } 184 # apply actual function --> 185 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 186 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 187 # re-apply format to the output /usr/local/lib/python3.7/dist-packages/datasets/fingerprint.py in wrapper(*args, **kwargs) 396 # Call actual function 397 --> 398 out = func(self, *args, **kwargs) 399 400 # Update fingerprint of in-place transforms + update in-place history of transforms /usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, disable_tqdm, desc, cache_only) 2063 writer.write_table(batch) 2064 else: -> 2065 writer.write_batch(batch) 2066 if update_data and writer is not None: 2067 writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file /usr/local/lib/python3.7/dist-packages/datasets/arrow_writer.py in write_batch(self, batch_examples, writer_batch_size) 409 typed_sequence = OptimizedTypedSequence(batch_examples[col], type=col_type, try_type=col_try_type, col=col) 410 typed_sequence_examples[col] = typed_sequence --> 411 pa_table = pa.Table.from_pydict(typed_sequence_examples) 412 self.write_table(pa_table, writer_batch_size) 413 /usr/local/lib/python3.7/dist-packages/pyarrow/table.pxi in pyarrow.lib.Table.from_pydict() /usr/local/lib/python3.7/dist-packages/pyarrow/array.pxi in pyarrow.lib.asarray() /usr/local/lib/python3.7/dist-packages/pyarrow/array.pxi in pyarrow.lib.array() /usr/local/lib/python3.7/dist-packages/pyarrow/array.pxi in pyarrow.lib._handle_arrow_array_protocol() /usr/local/lib/python3.7/dist-packages/datasets/arrow_writer.py in __arrow_array__(self, type) 106 storage = numpy_to_pyarrow_listarray(self.data, type=type.value_type) 107 else: --> 108 storage = pa.array(self.data, type.storage_dtype) 109 out = pa.ExtensionArray.from_storage(type, storage) 110 elif isinstance(self.data, np.ndarray): /usr/local/lib/python3.7/dist-packages/pyarrow/array.pxi in pyarrow.lib.array() /usr/local/lib/python3.7/dist-packages/pyarrow/array.pxi in pyarrow.lib._sequence_to_array() /usr/local/lib/python3.7/dist-packages/pyarrow/error.pxi in pyarrow.lib.pyarrow_internal_check_status() /usr/local/lib/python3.7/dist-packages/pyarrow/error.pxi in pyarrow.lib.check_status() ArrowInvalid: Can only convert 1-dimensional array values ``` It can be fixed by adding the following line: ```diff def preprocess_data(examples): images = [Image.open(path).convert("RGB") for path in examples['image_path']] words = examples['words'] boxes = examples['bboxes'] word_labels = examples['ner_tags'] encoded_inputs = processor(images, words, boxes=boxes, word_labels=word_labels, padding="max_length", truncation=True) + encoded_inputs["image"] = np.array(encoded_inputs["image"]) return encoded_inputs ``` However, would be great if this can be fixed within Datasets itself.
2021-10-01T13:57:45Z
https://github.com/huggingface/datasets/issues/2987
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2987/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2986/comments
https://api.github.com/repos/huggingface/datasets/issues/2986/timeline
2021-10-11T11:05:52Z
null
null
PR_kwDODunzps4scSHR
closed
[]
false
2,986
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Refac module factory + avoid etag requests for hub datasets
https://api.github.com/repos/huggingface/datasets/issues/2986/events
null
https://api.github.com/repos/huggingface/datasets/issues/2986/labels{/name}
2021-09-29T10:42:00Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2986.diff", "html_url": "https://github.com/huggingface/datasets/pull/2986", "merged_at": "2021-10-11T11:05:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/2986.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2986" }
1,010,792,783
[]
https://api.github.com/repos/huggingface/datasets/issues/2986
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Refactor the module factory When trying to extend the `data_files` logic to avoid doing unnecessary ETag requests, I noticed that the module preparation mechanism needed a refactor: - the function was 600 lines long - it was not readable - it contained many different cases that made it complex to maintain - it was hard to properly test it - it was hard to extend without breaking anything The module preparation mechanism is in charge of taking the name of a dataset or a metric given by the user (ex: "squad", "accuracy", "lhoestq/test", "path/to/my/script.py", "path/to/my/data/directory", "json", "csv") and return a module (possibly downloaded from the Hub) that contains the dataset builder or the metric class to use. ### Implementation details I decided to separate all these use cases into different dataset/metric module factories. First, the metric module factories: - **CanonicalMetricModuleFactory**: "accuracy", "rouge", ... - **LocalMetricModuleFactory**: "path/to/my/metric.py" Then, the dataset module factories: - **CanonicalDatasetModuleFactory**: "squad", "glue", ... - **CommunityDatasetModuleFactoryWithScript**: "lhoestq/test" - **CommunityDatasetModuleFactoryWithoutScript**: "lhoestq/demo1" - **PackagedDatasetModuleFactory**: "json", "csv", ... - **LocalDatasetModuleFactoryWithScript**: "path/to/my/script.py" - **LocalDatasetModuleFactoryWithoutScript**: "path/to/my/data/directory" And finally, additional factories when users have no internet: - **CachedDatasetModuleFactory** - **CachedMetricModuleFactory** ### Breaking changes One thing is that I still don't know at what extent we want to keep backward compatibility for `prepare_module`. For now I just kept it (except I removed two parameters) just in case, but it's not used anywhere anymore. ## Avoid etag requests for hub datasets To do this I added a class `DataFilesDict` that can be hashed to define the cache directory of the dataset. It contains the usual data files formatted as `{"train": ["train.txt"]}` for example. But each list of file is a `DataFilesList` that also has a `origin_metadata` attribute that contains metadata about the origin of each file: - for URLs: it stores the ETags of the files - for local files: it stores the last modification data - for files from a Hugging Face repository on the Hub: it stores the pattern (`*`, `*.csv`, "train.txt", etc.) and the commit sha of the repository (so there're no ETag requests !) This way if any file changes, the hash of the `DataFilesDict` changes too ! You can instantiate a `DataFilesDict` by using patterns for local/remote files or files in a HF repository: - for local/remote files: `DataFilesDict.from_local_or_remote(patterns)` - for files in a HF repository: `DataFilesDict.from_hf_repo(patterns, dataset_info)` Fix #2859 ## TODO Fix the latest test: - [x] fix the call to dataset_info in offline mode (related to https://github.com/huggingface/huggingface_hub/issues/372) Add some more tests: - [x] test all the factories - [x] test the new data files logic Other: - [x] docstrings - [x] comments
2021-10-11T11:05:53Z
https://github.com/huggingface/datasets/pull/2986
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 2, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/2986/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2985/comments
https://api.github.com/repos/huggingface/datasets/issues/2985/timeline
2021-10-01T16:55:19Z
null
null
PR_kwDODunzps4sbbbo
closed
[]
false
2,985
{ "avatar_url": "https://avatars.githubusercontent.com/u/46108405?v=4", "events_url": "https://api.github.com/users/adeepH/events{/privacy}", "followers_url": "https://api.github.com/users/adeepH/followers", "following_url": "https://api.github.com/users/adeepH/following{/other_user}", "gists_url": "https://api.github.com/users/adeepH/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/adeepH", "id": 46108405, "login": "adeepH", "node_id": "MDQ6VXNlcjQ2MTA4NDA1", "organizations_url": "https://api.github.com/users/adeepH/orgs", "received_events_url": "https://api.github.com/users/adeepH/received_events", "repos_url": "https://api.github.com/users/adeepH/repos", "site_admin": false, "starred_url": "https://api.github.com/users/adeepH/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/adeepH/subscriptions", "type": "User", "url": "https://api.github.com/users/adeepH" }
add new dataset kan_hope
https://api.github.com/repos/huggingface/datasets/issues/2985/events
null
https://api.github.com/repos/huggingface/datasets/issues/2985/labels{/name}
2021-09-29T05:20:28Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2985.diff", "html_url": "https://github.com/huggingface/datasets/pull/2985", "merged_at": "2021-10-01T16:55:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/2985.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2985" }
1,010,500,433
[]
https://api.github.com/repos/huggingface/datasets/issues/2985
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Adding a Dataset - **Name:** *KanHope* - **Description:** *A code-mixed English-Kannada dataset for Hope speech detection* - **Task:** *Binary Text Classification* - **Paper:** *https://arxiv.org/abs/2108.04616* - **Data:** *https://github.com/adeepH/kan_hope/tree/main/dataset* - **Motivation:** *The dataset is amongst the very few resources available for code-mixed low-resourced Dravidian languages of India*
2021-10-01T16:55:19Z
https://github.com/huggingface/datasets/pull/2985
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2985/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2984/comments
https://api.github.com/repos/huggingface/datasets/issues/2984/timeline
2021-10-12T06:05:42Z
null
completed
I_kwDODunzps48OsRm
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,984
{ "avatar_url": "https://avatars.githubusercontent.com/u/25057983?v=4", "events_url": "https://api.github.com/users/zijwang/events{/privacy}", "followers_url": "https://api.github.com/users/zijwang/followers", "following_url": "https://api.github.com/users/zijwang/following{/other_user}", "gists_url": "https://api.github.com/users/zijwang/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zijwang", "id": 25057983, "login": "zijwang", "node_id": "MDQ6VXNlcjI1MDU3OTgz", "organizations_url": "https://api.github.com/users/zijwang/orgs", "received_events_url": "https://api.github.com/users/zijwang/received_events", "repos_url": "https://api.github.com/users/zijwang/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zijwang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zijwang/subscriptions", "type": "User", "url": "https://api.github.com/users/zijwang" }
Exceeded maximum rows when reading large files
https://api.github.com/repos/huggingface/datasets/issues/2984/events
null
https://api.github.com/repos/huggingface/datasets/issues/2984/labels{/name}
2021-09-29T04:49:22Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,010,484,326
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2984
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug A clear and concise description of what the bug is. When using `load_dataset` with json files, if the files are too large, there will be "Exceeded maximum rows" error. ## Steps to reproduce the bug ```python dataset = load_dataset('json', data_files=data_files) # data files have 3M rows in a single file ``` ## Expected results No error ## Actual results ``` ~/anaconda3/envs/python/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py in _generate_tables(self, files) 134 with open(file, encoding="utf-8") as f: --> 135 dataset = json.load(f) 136 except json.JSONDecodeError: ~/anaconda3/envs/python/lib/python3.9/json/__init__.py in load(fp, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw) 292 """ --> 293 return loads(fp.read(), 294 cls=cls, object_hook=object_hook, ~/anaconda3/envs/python/lib/python3.9/json/__init__.py in loads(s, cls, object_hook, parse_float, parse_int, parse_constant, object_pairs_hook, **kw) 345 parse_constant is None and object_pairs_hook is None and not kw): --> 346 return _default_decoder.decode(s) 347 if cls is None: ~/anaconda3/envs/python/lib/python3.9/json/decoder.py in decode(self, s, _w) 339 if end != len(s): --> 340 raise JSONDecodeError("Extra data", s, end) 341 return obj JSONDecodeError: Extra data: line 2 column 1 (char 20321) During handling of the above exception, another exception occurred: ArrowInvalid Traceback (most recent call last) <ipython-input-20-ab3718a6482f> in <module> ----> 1 dataset = load_dataset('json', data_files=data_files) ~/anaconda3/envs/python/lib/python3.9/site-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs) 841 842 # Download and prepare data --> 843 builder_instance.download_and_prepare( 844 download_config=download_config, 845 download_mode=download_mode, ~/anaconda3/envs/python/lib/python3.9/site-packages/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs) 606 logger.warning("HF google storage unreachable. Downloading and preparing it from source") 607 if not downloaded_from_gcs: --> 608 self._download_and_prepare( 609 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs 610 ) ~/anaconda3/envs/python/lib/python3.9/site-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 684 try: 685 # Prepare split will record examples associated to the split --> 686 self._prepare_split(split_generator, **prepare_split_kwargs) 687 except OSError as e: 688 raise OSError( ~/anaconda3/envs/python/lib/python3.9/site-packages/datasets/builder.py in _prepare_split(self, split_generator) 1153 generator = self._generate_tables(**split_generator.gen_kwargs) 1154 with ArrowWriter(features=self.info.features, path=fpath) as writer: -> 1155 for key, table in utils.tqdm( 1156 generator, unit=" tables", leave=False, disable=bool(logging.get_verbosity() == logging.NOTSET) 1157 ): ~/anaconda3/envs/python/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py in _generate_tables(self, files) 135 dataset = json.load(f) 136 except json.JSONDecodeError: --> 137 raise e 138 raise ValueError( 139 f"Not able to read records in the JSON file at {file}. " ~/anaconda3/envs/python/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py in _generate_tables(self, files) 114 while True: 115 try: --> 116 pa_table = paj.read_json( 117 BytesIO(batch), read_options=paj.ReadOptions(block_size=block_size) 118 ) ~/anaconda3/envs/python/lib/python3.9/site-packages/pyarrow/_json.pyx in pyarrow._json.read_json() ~/anaconda3/envs/python/lib/python3.9/site-packages/pyarrow/error.pxi in pyarrow.lib.pyarrow_internal_check_status() ~/anaconda3/envs/python/lib/python3.9/site-packages/pyarrow/error.pxi in pyarrow.lib.check_status() ArrowInvalid: Exceeded maximum rows ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - Platform: Linux - Python version: 3.9 - PyArrow version: 4.0.1
2021-10-12T06:05:42Z
https://github.com/huggingface/datasets/issues/2984
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2984/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2983/comments
https://api.github.com/repos/huggingface/datasets/issues/2983/timeline
2021-10-01T16:03:05Z
null
null
PR_kwDODunzps4saw_v
closed
[]
false
2,983
{ "avatar_url": "https://avatars.githubusercontent.com/u/3775944?v=4", "events_url": "https://api.github.com/users/JoelNiklaus/events{/privacy}", "followers_url": "https://api.github.com/users/JoelNiklaus/followers", "following_url": "https://api.github.com/users/JoelNiklaus/following{/other_user}", "gists_url": "https://api.github.com/users/JoelNiklaus/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JoelNiklaus", "id": 3775944, "login": "JoelNiklaus", "node_id": "MDQ6VXNlcjM3NzU5NDQ=", "organizations_url": "https://api.github.com/users/JoelNiklaus/orgs", "received_events_url": "https://api.github.com/users/JoelNiklaus/received_events", "repos_url": "https://api.github.com/users/JoelNiklaus/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JoelNiklaus/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JoelNiklaus/subscriptions", "type": "User", "url": "https://api.github.com/users/JoelNiklaus" }
added SwissJudgmentPrediction dataset
https://api.github.com/repos/huggingface/datasets/issues/2983/events
null
https://api.github.com/repos/huggingface/datasets/issues/2983/labels{/name}
2021-09-28T22:17:56Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2983.diff", "html_url": "https://github.com/huggingface/datasets/pull/2983", "merged_at": "2021-10-01T16:03:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/2983.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2983" }
1,010,263,058
[]
https://api.github.com/repos/huggingface/datasets/issues/2983
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2021-10-01T16:03:05Z
https://github.com/huggingface/datasets/pull/2983
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2983/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2982/comments
https://api.github.com/repos/huggingface/datasets/issues/2982/timeline
2021-10-01T12:21:00Z
null
null
PR_kwDODunzps4saVLh
closed
[]
false
2,982
{ "avatar_url": "https://avatars.githubusercontent.com/u/91226467?v=4", "events_url": "https://api.github.com/users/hacobe/events{/privacy}", "followers_url": "https://api.github.com/users/hacobe/followers", "following_url": "https://api.github.com/users/hacobe/following{/other_user}", "gists_url": "https://api.github.com/users/hacobe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hacobe", "id": 91226467, "login": "hacobe", "node_id": "MDQ6VXNlcjkxMjI2NDY3", "organizations_url": "https://api.github.com/users/hacobe/orgs", "received_events_url": "https://api.github.com/users/hacobe/received_events", "repos_url": "https://api.github.com/users/hacobe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hacobe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hacobe/subscriptions", "type": "User", "url": "https://api.github.com/users/hacobe" }
Add the Math Aptitude Test of Heuristics dataset.
https://api.github.com/repos/huggingface/datasets/issues/2982/events
null
https://api.github.com/repos/huggingface/datasets/issues/2982/labels{/name}
2021-09-28T19:18:37Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2982.diff", "html_url": "https://github.com/huggingface/datasets/pull/2982", "merged_at": "2021-10-01T12:21:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/2982.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2982" }
1,010,118,418
[]
https://api.github.com/repos/huggingface/datasets/issues/2982
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2021-10-01T19:51:23Z
https://github.com/huggingface/datasets/pull/2982
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2982/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2981/comments
https://api.github.com/repos/huggingface/datasets/issues/2981/timeline
2022-05-05T14:26:41Z
null
null
PR_kwDODunzps4sZ4ke
closed
[]
false
2,981
{ "avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4", "events_url": "https://api.github.com/users/nateraw/events{/privacy}", "followers_url": "https://api.github.com/users/nateraw/followers", "following_url": "https://api.github.com/users/nateraw/following{/other_user}", "gists_url": "https://api.github.com/users/nateraw/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nateraw", "id": 32437151, "login": "nateraw", "node_id": "MDQ6VXNlcjMyNDM3MTUx", "organizations_url": "https://api.github.com/users/nateraw/orgs", "received_events_url": "https://api.github.com/users/nateraw/received_events", "repos_url": "https://api.github.com/users/nateraw/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nateraw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nateraw/subscriptions", "type": "User", "url": "https://api.github.com/users/nateraw" }
add wit dataset
https://api.github.com/repos/huggingface/datasets/issues/2981/events
null
https://api.github.com/repos/huggingface/datasets/issues/2981/labels{/name}
2021-09-28T16:34:49Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2981.diff", "html_url": "https://github.com/huggingface/datasets/pull/2981", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2981.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2981" }
1,009,969,310
[]
https://api.github.com/repos/huggingface/datasets/issues/2981
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Resolves #2902 based on conversation there - would also close #2810. Open to suggestions/help 😀 CC @hassiahk @lhoestq @yjernite
2022-05-05T14:26:41Z
https://github.com/huggingface/datasets/pull/2981
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 2, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2981/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2980/comments
https://api.github.com/repos/huggingface/datasets/issues/2980/timeline
null
null
null
I_kwDODunzps48MXJK
open
[]
null
2,980
{ "avatar_url": "https://avatars.githubusercontent.com/u/4109253?v=4", "events_url": "https://api.github.com/users/cdleong/events{/privacy}", "followers_url": "https://api.github.com/users/cdleong/followers", "following_url": "https://api.github.com/users/cdleong/following{/other_user}", "gists_url": "https://api.github.com/users/cdleong/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cdleong", "id": 4109253, "login": "cdleong", "node_id": "MDQ6VXNlcjQxMDkyNTM=", "organizations_url": "https://api.github.com/users/cdleong/orgs", "received_events_url": "https://api.github.com/users/cdleong/received_events", "repos_url": "https://api.github.com/users/cdleong/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cdleong/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cdleong/subscriptions", "type": "User", "url": "https://api.github.com/users/cdleong" }
OpenSLR 25: ASR data for Amharic, Swahili and Wolof
https://api.github.com/repos/huggingface/datasets/issues/2980/events
null
https://api.github.com/repos/huggingface/datasets/issues/2980/labels{/name}
2021-09-28T15:04:36Z
null
false
null
null
1,009,873,482
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
https://api.github.com/repos/huggingface/datasets/issues/2980
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Adding a Dataset - **Name:** *SLR25* - **Description:** *Subset 25 from OpenSLR. Other subsets have been added to https://huggingface.co/datasets/openslr, 25 covers Amharic, Swahili and Wolof data* - **Paper:** *https://www.openslr.org/25/ has citations for each of the three subsubsets. * - **Data:** *Currently the three links to the .tar.bz2 files can be found a thttps://www.openslr.org/25/* - **Motivation:** *Increase ASR data for underrepresented African languages. Also, other subsets of OpenSLR speech recognition have been uploaded, so this would be easy.* https://github.com/huggingface/datasets/blob/master/datasets/openslr/openslr.py already has been created for various other OpenSLR subsets, this should be relatively straightforward to do.
2021-09-29T17:25:14Z
https://github.com/huggingface/datasets/issues/2980
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2980/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2979/comments
https://api.github.com/repos/huggingface/datasets/issues/2979/timeline
2021-10-01T14:17:38Z
null
completed
I_kwDODunzps48Lctj
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,979
{ "avatar_url": "https://avatars.githubusercontent.com/u/74454835?v=4", "events_url": "https://api.github.com/users/asofiaoliveira/events{/privacy}", "followers_url": "https://api.github.com/users/asofiaoliveira/followers", "following_url": "https://api.github.com/users/asofiaoliveira/following{/other_user}", "gists_url": "https://api.github.com/users/asofiaoliveira/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/asofiaoliveira", "id": 74454835, "login": "asofiaoliveira", "node_id": "MDQ6VXNlcjc0NDU0ODM1", "organizations_url": "https://api.github.com/users/asofiaoliveira/orgs", "received_events_url": "https://api.github.com/users/asofiaoliveira/received_events", "repos_url": "https://api.github.com/users/asofiaoliveira/repos", "site_admin": false, "starred_url": "https://api.github.com/users/asofiaoliveira/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/asofiaoliveira/subscriptions", "type": "User", "url": "https://api.github.com/users/asofiaoliveira" }
ValueError when computing f1 metric with average None
https://api.github.com/repos/huggingface/datasets/issues/2979/events
null
https://api.github.com/repos/huggingface/datasets/issues/2979/labels{/name}
2021-09-28T11:34:53Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,009,634,147
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2979
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug When I try to compute the f1 score for each class in a multiclass classification problem, I get a ValueError. The same happens with recall and precision. I traced the error to the `.item()` in these scripts, which is probably there for the other averages. E.g. from f1.py: ```python return { "f1": f1_score( references, predictions, labels=labels, pos_label=pos_label, average=average, sample_weight=sample_weight, ).item(), } ``` Since the result is an array with more than one item, the `.item()` throws the error. I didn't submit a PR because this might be needed for the other averages, I'm not very familiar with the library ## Steps to reproduce the bug ```python from datasets import load_metric metric = load_metric("f1") metric.add_batch(predictions=[2,34,1,34,1,2,3], references=[23,52,1,3,523,5,8]) metric.compute(average=None) ``` ## Expected results `array([0.66666667, 0. , 0. , 0. , 0. , 0. , 0. , 0. , 0. ])` ## Actual results ValueError: can only convert an array of size 1 to a Python scalar ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.1 - Platform: Windows-10-10.0.19041-SP0 - Python version: 3.9.5 - PyArrow version: 5.0.0
2021-10-01T14:17:38Z
https://github.com/huggingface/datasets/issues/2979
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2979/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2978/comments
https://api.github.com/repos/huggingface/datasets/issues/2978/timeline
null
null
null
I_kwDODunzps48LBML
open
[]
null
2,978
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Run CI tests against non-production server
https://api.github.com/repos/huggingface/datasets/issues/2978/events
null
https://api.github.com/repos/huggingface/datasets/issues/2978/labels{/name}
2021-09-28T09:41:26Z
null
false
null
null
1,009,521,419
[]
https://api.github.com/repos/huggingface/datasets/issues/2978
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Currently, the CI test suite performs requests to the HF production server. As discussed with @elishowk, we should refactor our tests to use the HF staging server instead, like `huggingface_hub` and `transformers`.
2021-09-28T15:23:50Z
https://github.com/huggingface/datasets/issues/2978
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2978/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2977/comments
https://api.github.com/repos/huggingface/datasets/issues/2977/timeline
2021-10-01T15:53:15Z
null
completed
I_kwDODunzps48KeWE
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,977
{ "avatar_url": "https://avatars.githubusercontent.com/u/19476123?v=4", "events_url": "https://api.github.com/users/Valahaar/events{/privacy}", "followers_url": "https://api.github.com/users/Valahaar/followers", "following_url": "https://api.github.com/users/Valahaar/following{/other_user}", "gists_url": "https://api.github.com/users/Valahaar/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Valahaar", "id": 19476123, "login": "Valahaar", "node_id": "MDQ6VXNlcjE5NDc2MTIz", "organizations_url": "https://api.github.com/users/Valahaar/orgs", "received_events_url": "https://api.github.com/users/Valahaar/received_events", "repos_url": "https://api.github.com/users/Valahaar/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Valahaar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Valahaar/subscriptions", "type": "User", "url": "https://api.github.com/users/Valahaar" }
Impossible to load compressed csv
https://api.github.com/repos/huggingface/datasets/issues/2977/events
null
https://api.github.com/repos/huggingface/datasets/issues/2977/labels{/name}
2021-09-28T07:18:54Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,009,378,692
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2977
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug It is not possible to load from a compressed csv anymore. ## Steps to reproduce the bug ```python load_dataset('csv', data_files=['/path/to/csv.bz2']) ``` ## Problem and possible solution This used to work, but the commit that broke it is [this one](https://github.com/huggingface/datasets/commit/ad489d4597381fc2d12c77841642cbeaecf7a2e0#diff-6f60f8d0552b75be8b3bfd09994480fd60dcd4e7eb08d02f721218c3acdd2782). `pandas` usually gets the compression information from the filename itself (which was previously directly passed). Now, since it gets a file descriptor, it might be good to auto-infer the compression or let the user pass the `compression` kwarg to `load_dataset` (or maybe warn the user if the file ends with a commonly known compression scheme?). ## Environment info - `datasets` version: 1.10.0 (and over) - Platform: Linux-5.8.0-45-generic-x86_64-with-glibc2.17 - Python version: 3.8.10 - PyArrow version: 3.0.0
2021-10-01T15:53:16Z
https://github.com/huggingface/datasets/issues/2977
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2977/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2976/comments
https://api.github.com/repos/huggingface/datasets/issues/2976/timeline
2021-09-28T06:53:01Z
null
completed
I_kwDODunzps48Hr7R
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,976
{ "avatar_url": "https://avatars.githubusercontent.com/u/77006774?v=4", "events_url": "https://api.github.com/users/mskovalova/events{/privacy}", "followers_url": "https://api.github.com/users/mskovalova/followers", "following_url": "https://api.github.com/users/mskovalova/following{/other_user}", "gists_url": "https://api.github.com/users/mskovalova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mskovalova", "id": 77006774, "login": "mskovalova", "node_id": "MDQ6VXNlcjc3MDA2Nzc0", "organizations_url": "https://api.github.com/users/mskovalova/orgs", "received_events_url": "https://api.github.com/users/mskovalova/received_events", "repos_url": "https://api.github.com/users/mskovalova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mskovalova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mskovalova/subscriptions", "type": "User", "url": "https://api.github.com/users/mskovalova" }
Can't load dataset
https://api.github.com/repos/huggingface/datasets/issues/2976/events
null
https://api.github.com/repos/huggingface/datasets/issues/2976/labels{/name}
2021-09-27T21:38:14Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,008,647,889
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2976
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
I'm trying to load a wikitext dataset ``` from datasets import load_dataset raw_datasets = load_dataset("wikitext") ``` ValueError: Config name is missing. Please pick one among the available configs: ['wikitext-103-raw-v1', 'wikitext-2-raw-v1', 'wikitext-103-v1', 'wikitext-2-v1'] Example of usage: `load_dataset('wikitext', 'wikitext-103-raw-v1')`. If I try ``` from datasets import load_dataset raw_datasets = load_dataset("wikitext-2-v1") ``` FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/1.12.1/datasets/wikitext-2-v1/wikitext-2-v1.py #### Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.1 - Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic (colab) - Python version: 3.7.12 - PyArrow version: 3.0.0
2024-03-07T13:57:36Z
https://github.com/huggingface/datasets/issues/2976
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2976/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2975/comments
https://api.github.com/repos/huggingface/datasets/issues/2975/timeline
2021-09-29T09:05:38Z
null
null
PR_kwDODunzps4sVAOt
closed
[]
false
2,975
{ "avatar_url": "https://avatars.githubusercontent.com/u/46553104?v=4", "events_url": "https://api.github.com/users/Ishan-Kumar2/events{/privacy}", "followers_url": "https://api.github.com/users/Ishan-Kumar2/followers", "following_url": "https://api.github.com/users/Ishan-Kumar2/following{/other_user}", "gists_url": "https://api.github.com/users/Ishan-Kumar2/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Ishan-Kumar2", "id": 46553104, "login": "Ishan-Kumar2", "node_id": "MDQ6VXNlcjQ2NTUzMTA0", "organizations_url": "https://api.github.com/users/Ishan-Kumar2/orgs", "received_events_url": "https://api.github.com/users/Ishan-Kumar2/received_events", "repos_url": "https://api.github.com/users/Ishan-Kumar2/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Ishan-Kumar2/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Ishan-Kumar2/subscriptions", "type": "User", "url": "https://api.github.com/users/Ishan-Kumar2" }
ignore dummy folder and dataset_infos.json
https://api.github.com/repos/huggingface/datasets/issues/2975/events
null
https://api.github.com/repos/huggingface/datasets/issues/2975/labels{/name}
2021-09-27T18:09:03Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2975.diff", "html_url": "https://github.com/huggingface/datasets/pull/2975", "merged_at": "2021-09-29T09:05:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/2975.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2975" }
1,008,444,654
[]
https://api.github.com/repos/huggingface/datasets/issues/2975
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Fixes #2877 Added the `dataset_infos.json` to the ignored files list and also added check to ignore files which have parent directory as `dummy`. Let me know if it is correct. Thanks :)
2021-09-29T09:45:38Z
https://github.com/huggingface/datasets/pull/2975
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2975/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2974/comments
https://api.github.com/repos/huggingface/datasets/issues/2974/timeline
2021-09-29T09:04:41Z
null
null
PR_kwDODunzps4sUZCX
closed
[]
false
2,974
{ "avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4", "events_url": "https://api.github.com/users/Rocketknight1/events{/privacy}", "followers_url": "https://api.github.com/users/Rocketknight1/followers", "following_url": "https://api.github.com/users/Rocketknight1/following{/other_user}", "gists_url": "https://api.github.com/users/Rocketknight1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Rocketknight1", "id": 12866554, "login": "Rocketknight1", "node_id": "MDQ6VXNlcjEyODY2NTU0", "organizations_url": "https://api.github.com/users/Rocketknight1/orgs", "received_events_url": "https://api.github.com/users/Rocketknight1/received_events", "repos_url": "https://api.github.com/users/Rocketknight1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Rocketknight1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Rocketknight1/subscriptions", "type": "User", "url": "https://api.github.com/users/Rocketknight1" }
Actually disable dummy labels by default
https://api.github.com/repos/huggingface/datasets/issues/2974/events
null
https://api.github.com/repos/huggingface/datasets/issues/2974/labels{/name}
2021-09-27T14:50:20Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2974.diff", "html_url": "https://github.com/huggingface/datasets/pull/2974", "merged_at": "2021-09-29T09:04:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/2974.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2974" }
1,008,247,787
[]
https://api.github.com/repos/huggingface/datasets/issues/2974
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
So I might have just changed the docstring instead of the actual default argument value and not realized. @lhoestq I'm sorry >.>
2021-09-29T09:04:42Z
https://github.com/huggingface/datasets/pull/2974
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2974/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2973/comments
https://api.github.com/repos/huggingface/datasets/issues/2973/timeline
2021-09-27T12:59:59Z
null
null
PR_kwDODunzps4sTRvk
closed
[]
false
2,973
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix JSON metadata of masakhaner dataset
https://api.github.com/repos/huggingface/datasets/issues/2973/events
null
https://api.github.com/repos/huggingface/datasets/issues/2973/labels{/name}
2021-09-27T09:09:08Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2973.diff", "html_url": "https://github.com/huggingface/datasets/pull/2973", "merged_at": "2021-09-27T12:59:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/2973.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2973" }
1,007,894,592
[]
https://api.github.com/repos/huggingface/datasets/issues/2973
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix #2971.
2021-09-27T12:59:59Z
https://github.com/huggingface/datasets/pull/2973
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2973/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2972/comments
https://api.github.com/repos/huggingface/datasets/issues/2972/timeline
2021-09-28T06:43:15Z
null
completed
I_kwDODunzps48EfDK
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,972
{ "avatar_url": "https://avatars.githubusercontent.com/u/24835382?v=4", "events_url": "https://api.github.com/users/qqaatw/events{/privacy}", "followers_url": "https://api.github.com/users/qqaatw/followers", "following_url": "https://api.github.com/users/qqaatw/following{/other_user}", "gists_url": "https://api.github.com/users/qqaatw/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/qqaatw", "id": 24835382, "login": "qqaatw", "node_id": "MDQ6VXNlcjI0ODM1Mzgy", "organizations_url": "https://api.github.com/users/qqaatw/orgs", "received_events_url": "https://api.github.com/users/qqaatw/received_events", "repos_url": "https://api.github.com/users/qqaatw/repos", "site_admin": false, "starred_url": "https://api.github.com/users/qqaatw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/qqaatw/subscriptions", "type": "User", "url": "https://api.github.com/users/qqaatw" }
OSError: Not enough disk space.
https://api.github.com/repos/huggingface/datasets/issues/2972/events
null
https://api.github.com/repos/huggingface/datasets/issues/2972/labels{/name}
2021-09-27T07:41:22Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,007,808,714
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2972
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug I'm trying to download `natural_questions` dataset from the Internet, and I've specified the cache_dir which locates in a mounted disk and has enough disk space. However, even though the space is enough, the disk space checking function still reports the space of root `/` disk having no enough space. The file system structure is like below. The root `/` has `115G` disk space available, and the `sda1` is mounted to `/mnt`, which has `1.2T` disk space available: ``` / /mnt/sda1/path/to/args.dataset_cache_dir ``` ## Steps to reproduce the bug ```python dataset_config = DownloadConfig( cache_dir=os.path.abspath(args.dataset_cache_dir), resume_download=True, ) dataset = load_dataset("natural_questions", download_config=dataset_config) ``` ## Expected results Can download the dataset without an error. ## Actual results The following error raised: ``` OSError: Not enough disk space. Needed: 134.92 GiB (download: 41.97 GiB, generated: 92.95 GiB, post-processed: Unknown size) ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.9.0 - Platform: Ubuntu 18.04 - Python version: 3.8.10 - PyArrow version:
2022-08-29T23:21:36Z
https://github.com/huggingface/datasets/issues/2972
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2972/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2971/comments
https://api.github.com/repos/huggingface/datasets/issues/2971/timeline
2021-09-27T12:59:59Z
null
completed
I_kwDODunzps48EDqK
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,971
{ "avatar_url": "https://avatars.githubusercontent.com/u/8900094?v=4", "events_url": "https://api.github.com/users/huu4ontocord/events{/privacy}", "followers_url": "https://api.github.com/users/huu4ontocord/followers", "following_url": "https://api.github.com/users/huu4ontocord/following{/other_user}", "gists_url": "https://api.github.com/users/huu4ontocord/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/huu4ontocord", "id": 8900094, "login": "huu4ontocord", "node_id": "MDQ6VXNlcjg5MDAwOTQ=", "organizations_url": "https://api.github.com/users/huu4ontocord/orgs", "received_events_url": "https://api.github.com/users/huu4ontocord/received_events", "repos_url": "https://api.github.com/users/huu4ontocord/repos", "site_admin": false, "starred_url": "https://api.github.com/users/huu4ontocord/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/huu4ontocord/subscriptions", "type": "User", "url": "https://api.github.com/users/huu4ontocord" }
masakhaner dataset load problem
https://api.github.com/repos/huggingface/datasets/issues/2971/events
null
https://api.github.com/repos/huggingface/datasets/issues/2971/labels{/name}
2021-09-27T04:59:07Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,007,696,522
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2971
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug Masakhaner dataset is not loading ## Steps to reproduce the bug ``` from datasets import load_dataset dataset = load_dataset("masakhaner",'amh') ``` ## Expected results Expected the return of a dataset ## Actual results ``` NonMatchingSplitsSizesError Traceback (most recent call last) <ipython-input-3-a6abc1161d4c> in <module>() 1 from datasets import load_dataset 2 ----> 3 dataset = load_dataset("masakhaner",'amh') 3 frames /usr/local/lib/python3.7/dist-packages/datasets/utils/info_utils.py in verify_splits(expected_splits, recorded_splits) 72 ] 73 if len(bad_splits) > 0: ---> 74 raise NonMatchingSplitsSizesError(str(bad_splits)) 75 logger.info("All the splits matched successfully.") 76 NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=639927, num_examples=1751, dataset_name='masakhaner'), 'recorded': SplitInfo(name='train', num_bytes=639911, num_examples=1750, dataset_name='masakhaner')}, {'expected': SplitInfo(name='validation', num_bytes=92768, num_examples=251, dataset_name='masakhaner'), 'recorded': SplitInfo(name='validation', num_bytes=92753, num_examples=250, dataset_name='masakhaner')}, {'expected': SplitInfo(name='test', num_bytes=184286, num_examples=501, dataset_name='masakhaner'), 'recorded': SplitInfo(name='test', num_bytes=184271, num_examples=500, dataset_name='masakhaner')}] ``` ## Environment info Google Colab
2021-09-27T12:59:59Z
https://github.com/huggingface/datasets/issues/2971
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2971/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2970/comments
https://api.github.com/repos/huggingface/datasets/issues/2970/timeline
2021-09-26T10:38:59Z
null
completed
I_kwDODunzps48Cso5
closed
[]
null
2,970
{ "avatar_url": "https://avatars.githubusercontent.com/u/90449239?v=4", "events_url": "https://api.github.com/users/rcacho172/events{/privacy}", "followers_url": "https://api.github.com/users/rcacho172/followers", "following_url": "https://api.github.com/users/rcacho172/following{/other_user}", "gists_url": "https://api.github.com/users/rcacho172/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/rcacho172", "id": 90449239, "login": "rcacho172", "node_id": "MDQ6VXNlcjkwNDQ5MjM5", "organizations_url": "https://api.github.com/users/rcacho172/orgs", "received_events_url": "https://api.github.com/users/rcacho172/received_events", "repos_url": "https://api.github.com/users/rcacho172/repos", "site_admin": false, "starred_url": "https://api.github.com/users/rcacho172/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rcacho172/subscriptions", "type": "User", "url": "https://api.github.com/users/rcacho172" }
Magnet’s
https://api.github.com/repos/huggingface/datasets/issues/2970/events
null
https://api.github.com/repos/huggingface/datasets/issues/2970/labels{/name}
2021-09-26T09:50:29Z
null
false
null
null
1,007,340,089
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
https://api.github.com/repos/huggingface/datasets/issues/2970
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Adding a Dataset - **Name:** *name of the dataset* - **Description:** *short description of the dataset (or link to social media or blog post)* - **Paper:** *link to the dataset paper if available* - **Data:** *link to the Github repository or current dataset location* - **Motivation:** *what are some good reasons to have this dataset* Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
2021-09-26T10:38:59Z
https://github.com/huggingface/datasets/issues/2970
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2970/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2969/comments
https://api.github.com/repos/huggingface/datasets/issues/2969/timeline
2021-10-11T07:46:42Z
null
completed
I_kwDODunzps48COzL
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,969
{ "avatar_url": "https://avatars.githubusercontent.com/u/43877130?v=4", "events_url": "https://api.github.com/users/smeyerhot/events{/privacy}", "followers_url": "https://api.github.com/users/smeyerhot/followers", "following_url": "https://api.github.com/users/smeyerhot/following{/other_user}", "gists_url": "https://api.github.com/users/smeyerhot/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/smeyerhot", "id": 43877130, "login": "smeyerhot", "node_id": "MDQ6VXNlcjQzODc3MTMw", "organizations_url": "https://api.github.com/users/smeyerhot/orgs", "received_events_url": "https://api.github.com/users/smeyerhot/received_events", "repos_url": "https://api.github.com/users/smeyerhot/repos", "site_admin": false, "starred_url": "https://api.github.com/users/smeyerhot/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/smeyerhot/subscriptions", "type": "User", "url": "https://api.github.com/users/smeyerhot" }
medical-dialog error
https://api.github.com/repos/huggingface/datasets/issues/2969/events
null
https://api.github.com/repos/huggingface/datasets/issues/2969/labels{/name}
2021-09-25T23:08:44Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,007,217,867
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2969
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug A clear and concise description of what the bug is. When I attempt to download the huggingface datatset medical_dialog it errors out midway through ## Steps to reproduce the bug ```python raw_datasets = load_dataset("medical_dialog", "en", split="train", download_mode="force_redownload", data_dir="./Medical-Dialogue-Dataset-English") ``` ## Expected results A clear and concise description of the expected results. No error ## Actual results ``` 3 frames /usr/local/lib/python3.7/dist-packages/datasets/utils/info_utils.py in verify_splits(expected_splits, recorded_splits) 72 ] 73 if len(bad_splits) > 0: ---> 74 raise NonMatchingSplitsSizesError(str(bad_splits)) 75 logger.info("All the splits matched successfully.") 76 NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=0, num_examples=0, dataset_name='medical_dialog'), 'recorded': SplitInfo(name='train', num_bytes=295097913, num_examples=229674, dataset_name='medical_dialog')}] ``` Specify the actual results or traceback. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.21.1 - Platform: colab - Python version: colab 3.7 - PyArrow version: N/A
2024-01-08T09:55:12Z
https://github.com/huggingface/datasets/issues/2969
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2969/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2968/comments
https://api.github.com/repos/huggingface/datasets/issues/2968/timeline
2021-10-07T22:47:26Z
null
completed
I_kwDODunzps48CMwQ
closed
[]
null
2,968
{ "avatar_url": "https://avatars.githubusercontent.com/u/30755778?v=4", "events_url": "https://api.github.com/users/LysandreJik/events{/privacy}", "followers_url": "https://api.github.com/users/LysandreJik/followers", "following_url": "https://api.github.com/users/LysandreJik/following{/other_user}", "gists_url": "https://api.github.com/users/LysandreJik/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/LysandreJik", "id": 30755778, "login": "LysandreJik", "node_id": "MDQ6VXNlcjMwNzU1Nzc4", "organizations_url": "https://api.github.com/users/LysandreJik/orgs", "received_events_url": "https://api.github.com/users/LysandreJik/received_events", "repos_url": "https://api.github.com/users/LysandreJik/repos", "site_admin": false, "starred_url": "https://api.github.com/users/LysandreJik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LysandreJik/subscriptions", "type": "User", "url": "https://api.github.com/users/LysandreJik" }
`DatasetDict` cannot be exported to parquet if the splits have different features
https://api.github.com/repos/huggingface/datasets/issues/2968/events
null
https://api.github.com/repos/huggingface/datasets/issues/2968/labels{/name}
2021-09-25T22:18:39Z
null
false
null
null
1,007,209,488
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2968
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Describe the bug I'm trying to use parquet as a means of serialization for both `Dataset` and `DatasetDict` objects. Using `to_parquet` alongside `from_parquet` or `load_dataset` for a `Dataset` works perfectly. For `DatasetDict`, I use `to_parquet` on each split to save the parquet files in individual folders representing individual splits. This works too, as long as the splits have identical features. If a split has different features to neighboring splits, then loading the dataset will fail: a single schema is used to load both splits, resulting in a failure to load the second parquet file. ## Steps to reproduce the bug The following works as expected: ```python from datasets import load_dataset ds = load_dataset("lhoestq/custom_squad") ds['train'].to_parquet("./ds/train/split.parquet") ds['validation'].to_parquet("./ds/validation/split.parquet") brand_new_dataset = load_dataset("ds") ``` Modifying a single split to add a new feature ends up in a crash: ```python from datasets import load_dataset ds = load_dataset("lhoestq/custom_squad") def identical_answers(e): e['identical_answers'] = len(set(e['answers']['text'])) == 1 return e ds['validation'] = ds['validation'].map(identical_answers) ds['train'].to_parquet("./ds/train/split.parquet") ds['validation'].to_parquet("./ds/validation/split.parquet") brand_new_dataset = load_dataset("ds") ``` ``` File "/home/lysandre/.config/JetBrains/PyCharm2021.2/scratches/datasets/upload_dataset.py", line 26, in <module> brand_new_dataset = load_dataset("ds") File "/home/lysandre/Workspaces/Python/datasets/src/datasets/load.py", line 1151, in load_dataset builder_instance.download_and_prepare( File "/home/lysandre/Workspaces/Python/datasets/src/datasets/builder.py", line 642, in download_and_prepare self._download_and_prepare( File "/home/lysandre/Workspaces/Python/datasets/src/datasets/builder.py", line 732, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/builder.py", line 1194, in _prepare_split writer.write_table(table) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/arrow_writer.py", line 428, in write_table pa_table = pa.Table.from_arrays([pa_table[name] for name in self._schema.names], schema=self._schema) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/arrow_writer.py", line 428, in <listcomp> pa_table = pa.Table.from_arrays([pa_table[name] for name in self._schema.names], schema=self._schema) File "pyarrow/table.pxi", line 1257, in pyarrow.lib.Table.__getitem__ File "pyarrow/table.pxi", line 1833, in pyarrow.lib.Table.column File "pyarrow/table.pxi", line 1808, in pyarrow.lib.Table._ensure_integer_index KeyError: 'Field "identical_answers" does not exist in table schema' ``` It does work, however, to use the `save_to_disk` and `load_from_disk` methods: ```py from datasets import load_from_disk ds = load_dataset("lhoestq/custom_squad") def identical_answers(e): e['identical_answers'] = len(set(e['answers']['text'])) == 1 return e ds['validation'] = ds['validation'].map(identical_answers) ds.save_to_disk("local_path") brand_new_dataset = load_from_disk("local_path") ``` ## Expected results The saving works correctly - but the loading fails. I would expect either an error when saving or an error-less instantiation of the dataset through the parquet files. If it's helpful, I've traced a possible patch to the `write_table` method here: https://github.com/huggingface/datasets/blob/26ff41aa3a642e46489db9e95be1e9a8c4e64bea/src/datasets/arrow_writer.py#L424-L425 The writer is built only if the parquet writer is `None`, but I expect we would want to build a new writer as the table schema has changed. Furthermore, it relies on having the property `update_features` set to `True` in order to update the features: https://github.com/huggingface/datasets/blob/26ff41aa3a642e46489db9e95be1e9a8c4e64bea/src/datasets/arrow_writer.py#L254-L255 but the `ArrowWriter` is instantiated without that option in the `_prepare_split` method of the `ArrowBasedBuilder`: https://github.com/huggingface/datasets/blob/26ff41aa3a642e46489db9e95be1e9a8c4e64bea/src/datasets/builder.py#L1190 Updating these two parts to recreate a schema on each split results in an error that is, unfortunately, out of my expertise: ``` File "/home/lysandre/.config/JetBrains/PyCharm2021.2/scratches/datasets/upload_dataset.py", line 27, in <module> brand_new_dataset = load_dataset("ds") File "/home/lysandre/Workspaces/Python/datasets/src/datasets/load.py", line 1163, in load_dataset ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/builder.py", line 819, in as_dataset datasets = utils.map_nested( File "/home/lysandre/Workspaces/Python/datasets/src/datasets/utils/py_utils.py", line 207, in map_nested mapped = [ File "/home/lysandre/Workspaces/Python/datasets/src/datasets/utils/py_utils.py", line 208, in <listcomp> _single_map_nested((function, obj, types, None, True)) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/utils/py_utils.py", line 143, in _single_map_nested return function(data_struct) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/builder.py", line 850, in _build_single_dataset ds = self._as_dataset( File "/home/lysandre/Workspaces/Python/datasets/src/datasets/builder.py", line 920, in _as_dataset dataset_kwargs = ArrowReader(self._cache_dir, self.info).read( File "/home/lysandre/Workspaces/Python/datasets/src/datasets/arrow_reader.py", line 217, in read return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/arrow_reader.py", line 238, in read_files pa_table = self._read_files(files, in_memory=in_memory) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/arrow_reader.py", line 173, in _read_files pa_table: Table = self._get_table_from_filename(f_dict, in_memory=in_memory) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/arrow_reader.py", line 308, in _get_table_from_filename table = ArrowReader.read_table(filename, in_memory=in_memory) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/arrow_reader.py", line 327, in read_table return table_cls.from_file(filename) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/table.py", line 458, in from_file table = _memory_mapped_arrow_table_from_file(filename) File "/home/lysandre/Workspaces/Python/datasets/src/datasets/table.py", line 45, in _memory_mapped_arrow_table_from_file pa_table = opened_stream.read_all() File "pyarrow/ipc.pxi", line 563, in pyarrow.lib.RecordBatchReader.read_all File "pyarrow/error.pxi", line 114, in pyarrow.lib.check_status OSError: Header-type of flatbuffer-encoded Message is not RecordBatch. ``` ## Environment info - `datasets` version: 1.12.2.dev0 - Platform: Linux-5.14.7-arch1-1-x86_64-with-glibc2.33 - Python version: 3.9.7 - PyArrow version: 5.0.0
2021-10-07T22:47:42Z
https://github.com/huggingface/datasets/issues/2968
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2968/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2967/comments
https://api.github.com/repos/huggingface/datasets/issues/2967/timeline
2021-10-03T20:34:22Z
null
completed
I_kwDODunzps48CJLV
closed
[]
null
2,967
{ "avatar_url": "https://avatars.githubusercontent.com/u/42200725?v=4", "events_url": "https://api.github.com/users/WadeYin9712/events{/privacy}", "followers_url": "https://api.github.com/users/WadeYin9712/followers", "following_url": "https://api.github.com/users/WadeYin9712/following{/other_user}", "gists_url": "https://api.github.com/users/WadeYin9712/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/WadeYin9712", "id": 42200725, "login": "WadeYin9712", "node_id": "MDQ6VXNlcjQyMjAwNzI1", "organizations_url": "https://api.github.com/users/WadeYin9712/orgs", "received_events_url": "https://api.github.com/users/WadeYin9712/received_events", "repos_url": "https://api.github.com/users/WadeYin9712/repos", "site_admin": false, "starred_url": "https://api.github.com/users/WadeYin9712/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/WadeYin9712/subscriptions", "type": "User", "url": "https://api.github.com/users/WadeYin9712" }
Adding vision-and-language datasets (e.g., VQA, VCR) to Datasets
https://api.github.com/repos/huggingface/datasets/issues/2967/events
null
https://api.github.com/repos/huggingface/datasets/issues/2967/labels{/name}
2021-09-25T20:58:15Z
null
false
null
null
1,007,194,837
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2967
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
**Is your feature request related to a problem? Please describe.** Would you like to add any vision-and-language datasets (e.g., VQA, VCR) to Huggingface Datasets? **Describe the solution you'd like** N/A **Describe alternatives you've considered** N/A **Additional context** This is Da Yin at UCLA. Recently, we have published an EMNLP 2021 paper about geo-diverse visual commonsense reasoning (https://arxiv.org/abs/2109.06860). We propose a new dataset called GD-VCR, a vision-and-language dataset to evaluate how well V&L models perform on scenarios involving geo-location-specific commonsense. We hope to have our V&L dataset incorporated into Huggingface to further promote our project, but I haven't seen much V&L datasets in the current package. Is it possible to add V&L datasets, and if so, how should we prepare for the loading? Thank you very much!
2021-10-03T20:34:22Z
https://github.com/huggingface/datasets/issues/2967
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2967/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2966/comments
https://api.github.com/repos/huggingface/datasets/issues/2966/timeline
2021-10-13T13:37:30Z
null
null
PR_kwDODunzps4sRRMs
closed
[]
false
2,966
{ "avatar_url": "https://avatars.githubusercontent.com/u/9130406?v=4", "events_url": "https://api.github.com/users/christospi/events{/privacy}", "followers_url": "https://api.github.com/users/christospi/followers", "following_url": "https://api.github.com/users/christospi/following{/other_user}", "gists_url": "https://api.github.com/users/christospi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/christospi", "id": 9130406, "login": "christospi", "node_id": "MDQ6VXNlcjkxMzA0MDY=", "organizations_url": "https://api.github.com/users/christospi/orgs", "received_events_url": "https://api.github.com/users/christospi/received_events", "repos_url": "https://api.github.com/users/christospi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/christospi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/christospi/subscriptions", "type": "User", "url": "https://api.github.com/users/christospi" }
Upload greek-legal-code dataset
https://api.github.com/repos/huggingface/datasets/issues/2966/events
null
https://api.github.com/repos/huggingface/datasets/issues/2966/labels{/name}
2021-09-25T16:52:15Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2966.diff", "html_url": "https://github.com/huggingface/datasets/pull/2966", "merged_at": "2021-10-13T13:37:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/2966.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2966" }
1,007,142,233
[]
https://api.github.com/repos/huggingface/datasets/issues/2966
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2021-10-13T13:37:30Z
https://github.com/huggingface/datasets/pull/2966
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2966/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2965/comments
https://api.github.com/repos/huggingface/datasets/issues/2965/timeline
2022-08-31T06:47:10Z
null
completed
I_kwDODunzps48BuJ5
closed
[]
null
2,965
{ "avatar_url": "https://avatars.githubusercontent.com/u/3339950?v=4", "events_url": "https://api.github.com/users/Ririkoo/events{/privacy}", "followers_url": "https://api.github.com/users/Ririkoo/followers", "following_url": "https://api.github.com/users/Ririkoo/following{/other_user}", "gists_url": "https://api.github.com/users/Ririkoo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Ririkoo", "id": 3339950, "login": "Ririkoo", "node_id": "MDQ6VXNlcjMzMzk5NTA=", "organizations_url": "https://api.github.com/users/Ririkoo/orgs", "received_events_url": "https://api.github.com/users/Ririkoo/received_events", "repos_url": "https://api.github.com/users/Ririkoo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Ririkoo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Ririkoo/subscriptions", "type": "User", "url": "https://api.github.com/users/Ririkoo" }
Invalid download URL of WMT17 `zh-en` data
https://api.github.com/repos/huggingface/datasets/issues/2965/events
null
https://api.github.com/repos/huggingface/datasets/issues/2965/labels{/name}
2021-09-25T13:17:32Z
null
false
null
null
1,007,084,153
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2965
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Partial data (wmt17 zh-en) cannot be downloaded due to an invalid URL. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset('wmt17','zh-en') ``` ## Expected results ConnectionError: Couldn't reach ftp://cwmt-wmt:cwmt-wmt@datasets.nju.edu.cn/parallel/casia2015.zip
2022-08-31T06:47:11Z
https://github.com/huggingface/datasets/issues/2965
{ "+1": 0, "-1": 0, "confused": 1, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2965/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2964/comments
https://api.github.com/repos/huggingface/datasets/issues/2964/timeline
2021-09-25T08:06:07Z
null
completed
I_kwDODunzps47_5ZQ
closed
[]
null
2,964
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
Error when calculating Matthews Correlation Coefficient loaded with `load_metric`
https://api.github.com/repos/huggingface/datasets/issues/2964/events
null
https://api.github.com/repos/huggingface/datasets/issues/2964/labels{/name}
2021-09-24T15:55:21Z
null
false
null
null
1,006,605,904
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2964
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug After loading the metric named "[Matthews Correlation Coefficient](https://huggingface.co/metrics/matthews_correlation)" from `🤗datasets`, the `.compute` method fails with the following exception `AttributeError: 'float' object has no attribute 'item'` (complete stack trace can be provided if required). ## Steps to reproduce the bug ```python import torch predictions = torch.ones((10,)) references = torch.zeros((10,)) from datasets import load_metric METRIC = load_metric("matthews_correlation") result = METRIC.compute(predictions=predictions, references=references) ``` ## Expected results We should expect a Python `dict` as it follows: ``` { "matthews_correlation": float() } ``` as defined in https://github.com/huggingface/datasets/blob/master/metrics/matthews_correlation/matthews_correlation.py, so the fix will imply removing `.item()`, since the value returned by the `scikit-learn` function is not a `torch.Tensor` but a `float`, which means that the `.item()` will fail. ## Actual results ``` Traceback (most recent call last): File "/home/alvaro.bartolome/XXX/xxx/cli.py", line 59, in main app() File "/home/alvaro.bartolome/miniconda3/envs/xxx/lib/python3.9/site-packages/typer/main.py", line 214, in __call__ return get_command(self)(*args, **kwargs) File "/home/alvaro.bartolome/miniconda3/envs/xxx/lib/python3.9/site-packages/click/core.py", line 1137, in __call__ return self.main(*args, **kwargs) File "/home/alvaro.bartolome/miniconda3/envs/xxx/lib/python3.9/site-packages/click/core.py", line 1062, in main rv = self.invoke(ctx) File "/home/alvaro.bartolome/miniconda3/envs/xxx/lib/python3.9/site-packages/click/core.py", line 1668, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/home/alvaro.bartolome/miniconda3/envs/xxx/lib/python3.9/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/home/alvaro.bartolome/miniconda3/envs/xxx/lib/python3.9/site-packages/click/core.py", line 763, in invoke return __callback(*args, **kwargs) File "/home/alvaro.bartolome/miniconda3/envs/xxx/lib/python3.9/site-packages/typer/main.py", line 500, in wrapper return callback(**use_params) # type: ignore File "/home/alvaro.bartolome/XXX/xxx/cli.py", line 43, in train metrics = trainer.evaluate() File "/home/alvaro.bartolome/miniconda3/envs/xxx/lib/python3.9/site-packages/transformers/trainer.py", line 2051, in evaluate output = eval_loop( File "/home/alvaro.bartolome/miniconda3/envs/xxx/lib/python3.9/site-packages/transformers/trainer.py", line 2292, in evaluation_loop metrics = self.compute_metrics(EvalPrediction(predictions=all_preds, label_ids=all_labels)) File "/home/alvaro.bartolome/XXX/xxx/metrics.py", line 20, in compute_metrics res = METRIC.compute(predictions=predictions, references=eval_preds.label_ids) File "/home/alvaro.bartolome/miniconda3/envs/lang/lib/python3.9/site-packages/datasets/metric.py", line 402, in compute output = self._compute(predictions=predictions, references=references, **kwargs) File "/home/alvaro.bartolome/.cache/huggingface/modules/datasets_modules/metrics/matthews_correlation/0275f1e9a4d318e3ea8cdd87547ee0d58d894966616052e3d18444ac8ddd2357/matthews_correlation.py", line 88, in _compute "matthews_correlation": matthews_corrcoef(references, predictions, sample_weight=sample_weight).item(), AttributeError: 'float' object has no attribute 'item' ``` ## Environment info - `datasets` version: 1.12.1 - Platform: Linux-4.15.0-1113-azure-x86_64-with-glibc2.23 - Python version: 3.9.7 - PyArrow version: 5.0.0
2024-02-16T10:14:35Z
https://github.com/huggingface/datasets/issues/2964
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2964/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2963/comments
https://api.github.com/repos/huggingface/datasets/issues/2963/timeline
2021-09-24T15:38:24Z
null
completed
I_kwDODunzps47_1K9
closed
[]
null
2,963
{ "avatar_url": "https://avatars.githubusercontent.com/u/40454218?v=4", "events_url": "https://api.github.com/users/keloemma/events{/privacy}", "followers_url": "https://api.github.com/users/keloemma/followers", "following_url": "https://api.github.com/users/keloemma/following{/other_user}", "gists_url": "https://api.github.com/users/keloemma/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/keloemma", "id": 40454218, "login": "keloemma", "node_id": "MDQ6VXNlcjQwNDU0MjE4", "organizations_url": "https://api.github.com/users/keloemma/orgs", "received_events_url": "https://api.github.com/users/keloemma/received_events", "repos_url": "https://api.github.com/users/keloemma/repos", "site_admin": false, "starred_url": "https://api.github.com/users/keloemma/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/keloemma/subscriptions", "type": "User", "url": "https://api.github.com/users/keloemma" }
raise TypeError( TypeError: Provided `function` which is applied to all elements of table returns a variable of type <class 'list'>. Make sure provided `function` returns a variable of type `dict` to update the dataset or `None` if you are only interested in side effects.
https://api.github.com/repos/huggingface/datasets/issues/2963/events
null
https://api.github.com/repos/huggingface/datasets/issues/2963/labels{/name}
2021-09-24T15:35:11Z
null
false
null
null
1,006,588,605
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2963
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug A clear and concise description of what the bug is. I am trying to use Dataset to load my file in order to use Bert embeddings model baut when I finished loading using dataset and I want to pass to the tokenizer using the function map; I get the following error : raise TypeError( TypeError: Provided `function` which is applied to all elements of table returns a variable of type <class 'list'>. Make sure provided `function` returns a variable of type `dict` to update the dataset or `None` if you are only interested in side effects. I was able to load my file using dataset before but since this morning , I keep getting this erreor. ## Steps to reproduce the bug ```python # Xtrain, ytrain, filename, len_labels = read_file_2(fic) # Xtrain, lge_size = get_flaubert_layer(Xtrain, path_to_model_lge) data_preprocessed = make_new_traindata(Xtrain) my_dict = {"verbatim": data_preprocessed[1], "label": ytrain} # lemme avec conjonction dataset = Dataset.from_dict(my_dict) ``` ## Expected results A clear and concise description of the expected results. ## Actual results Specify the actual results or traceback. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - Platform: - Python version: - PyArrow version:
2021-09-24T15:38:24Z
https://github.com/huggingface/datasets/issues/2963
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2963/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2962/comments
https://api.github.com/repos/huggingface/datasets/issues/2962/timeline
null
null
null
I_kwDODunzps47_tni
open
[]
null
2,962
{ "avatar_url": "https://avatars.githubusercontent.com/u/53175384?v=4", "events_url": "https://api.github.com/users/merveenoyan/events{/privacy}", "followers_url": "https://api.github.com/users/merveenoyan/followers", "following_url": "https://api.github.com/users/merveenoyan/following{/other_user}", "gists_url": "https://api.github.com/users/merveenoyan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/merveenoyan", "id": 53175384, "login": "merveenoyan", "node_id": "MDQ6VXNlcjUzMTc1Mzg0", "organizations_url": "https://api.github.com/users/merveenoyan/orgs", "received_events_url": "https://api.github.com/users/merveenoyan/received_events", "repos_url": "https://api.github.com/users/merveenoyan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/merveenoyan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/merveenoyan/subscriptions", "type": "User", "url": "https://api.github.com/users/merveenoyan" }
Enable splits during streaming the dataset
https://api.github.com/repos/huggingface/datasets/issues/2962/events
null
https://api.github.com/repos/huggingface/datasets/issues/2962/labels{/name}
2021-09-24T15:01:29Z
null
false
null
null
1,006,557,666
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2962
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the Problem I'd like to stream only a specific percentage or part of the dataset. I want to do splitting when I'm streaming dataset as well. ## Solution Enabling splits when `streaming = True` as well. `e.g. dataset = load_dataset('dataset', split='train[:100]', streaming = True)` ## Alternatives Below is the alternative of doing it. `dataset = load_dataset("dataset", split='train', streaming = True).take(100)`
2021-09-24T15:01:29Z
https://github.com/huggingface/datasets/issues/2962
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2962/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2961/comments
https://api.github.com/repos/huggingface/datasets/issues/2961/timeline
2021-09-24T13:18:07Z
null
null
PR_kwDODunzps4sPTXV
closed
[]
false
2,961
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix CI doc build
https://api.github.com/repos/huggingface/datasets/issues/2961/events
null
https://api.github.com/repos/huggingface/datasets/issues/2961/labels{/name}
2021-09-24T13:13:28Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2961.diff", "html_url": "https://github.com/huggingface/datasets/pull/2961", "merged_at": "2021-09-24T13:18:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/2961.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2961" }
1,006,453,781
[]
https://api.github.com/repos/huggingface/datasets/issues/2961
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Pin `fsspec`. Before the issue: 'fsspec-2021.8.1', 's3fs-2021.8.1' Generating the issue: 'fsspec-2021.9.0', 's3fs-0.5.1'
2021-09-24T13:18:07Z
https://github.com/huggingface/datasets/pull/2961
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2961/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2960/comments
https://api.github.com/repos/huggingface/datasets/issues/2960/timeline
2021-09-24T11:22:30Z
null
null
PR_kwDODunzps4sOl0Y
closed
[]
false
2,960
{ "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "events_url": "https://api.github.com/users/SBrandeis/events{/privacy}", "followers_url": "https://api.github.com/users/SBrandeis/followers", "following_url": "https://api.github.com/users/SBrandeis/following{/other_user}", "gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SBrandeis", "id": 33657802, "login": "SBrandeis", "node_id": "MDQ6VXNlcjMzNjU3ODAy", "organizations_url": "https://api.github.com/users/SBrandeis/orgs", "received_events_url": "https://api.github.com/users/SBrandeis/received_events", "repos_url": "https://api.github.com/users/SBrandeis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions", "type": "User", "url": "https://api.github.com/users/SBrandeis" }
Support pandas 1.3 new `read_csv` parameters
https://api.github.com/repos/huggingface/datasets/issues/2960/events
null
https://api.github.com/repos/huggingface/datasets/issues/2960/labels{/name}
2021-09-24T08:37:24Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2960.diff", "html_url": "https://github.com/huggingface/datasets/pull/2960", "merged_at": "2021-09-24T11:22:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/2960.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2960" }
1,006,222,850
[]
https://api.github.com/repos/huggingface/datasets/issues/2960
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Support two new arguments introduced in pandas v1.3.0: - `encoding_errors` - `on_bad_lines` `read_csv` reference: https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.read_csv.html
2021-09-24T11:22:31Z
https://github.com/huggingface/datasets/pull/2960
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2960/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2959/comments
https://api.github.com/repos/huggingface/datasets/issues/2959/timeline
2022-03-01T17:41:51Z
null
null
PR_kwDODunzps4sMihl
closed
[]
false
2,959
{ "avatar_url": "https://avatars.githubusercontent.com/u/53175384?v=4", "events_url": "https://api.github.com/users/merveenoyan/events{/privacy}", "followers_url": "https://api.github.com/users/merveenoyan/followers", "following_url": "https://api.github.com/users/merveenoyan/following{/other_user}", "gists_url": "https://api.github.com/users/merveenoyan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/merveenoyan", "id": 53175384, "login": "merveenoyan", "node_id": "MDQ6VXNlcjUzMTc1Mzg0", "organizations_url": "https://api.github.com/users/merveenoyan/orgs", "received_events_url": "https://api.github.com/users/merveenoyan/received_events", "repos_url": "https://api.github.com/users/merveenoyan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/merveenoyan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/merveenoyan/subscriptions", "type": "User", "url": "https://api.github.com/users/merveenoyan" }
Added computer vision tasks
https://api.github.com/repos/huggingface/datasets/issues/2959/events
null
https://api.github.com/repos/huggingface/datasets/issues/2959/labels{/name}
2021-09-23T15:07:27Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2959.diff", "html_url": "https://github.com/huggingface/datasets/pull/2959", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2959.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2959" }
1,005,547,632
[]
https://api.github.com/repos/huggingface/datasets/issues/2959
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Added various image processing/computer vision tasks.
2022-03-01T17:41:51Z
https://github.com/huggingface/datasets/pull/2959
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2959/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2958/comments
https://api.github.com/repos/huggingface/datasets/issues/2958/timeline
2021-10-21T15:16:43Z
null
null
PR_kwDODunzps4sLTaB
closed
[]
false
2,958
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Add security policy to the project
https://api.github.com/repos/huggingface/datasets/issues/2958/events
null
https://api.github.com/repos/huggingface/datasets/issues/2958/labels{/name}
2021-09-23T08:20:55Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2958.diff", "html_url": "https://github.com/huggingface/datasets/pull/2958", "merged_at": "2021-10-21T15:16:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/2958.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2958" }
1,005,144,601
[]
https://api.github.com/repos/huggingface/datasets/issues/2958
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Add security policy to the project, as recommended by GitHub: https://docs.github.com/en/code-security/getting-started/adding-a-security-policy-to-your-repository Close #2953.
2021-10-21T15:16:44Z
https://github.com/huggingface/datasets/pull/2958
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2958/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2957/comments
https://api.github.com/repos/huggingface/datasets/issues/2957/timeline
2022-03-15T16:07:02Z
null
completed
I_kwDODunzps475RLx
closed
[]
null
2,957
{ "avatar_url": "https://avatars.githubusercontent.com/u/8754873?v=4", "events_url": "https://api.github.com/users/bradyneal/events{/privacy}", "followers_url": "https://api.github.com/users/bradyneal/followers", "following_url": "https://api.github.com/users/bradyneal/following{/other_user}", "gists_url": "https://api.github.com/users/bradyneal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bradyneal", "id": 8754873, "login": "bradyneal", "node_id": "MDQ6VXNlcjg3NTQ4NzM=", "organizations_url": "https://api.github.com/users/bradyneal/orgs", "received_events_url": "https://api.github.com/users/bradyneal/received_events", "repos_url": "https://api.github.com/users/bradyneal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bradyneal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bradyneal/subscriptions", "type": "User", "url": "https://api.github.com/users/bradyneal" }
MultiWOZ Dataset NonMatchingChecksumError
https://api.github.com/repos/huggingface/datasets/issues/2957/events
null
https://api.github.com/repos/huggingface/datasets/issues/2957/labels{/name}
2021-09-22T23:45:00Z
null
false
null
null
1,004,868,337
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2957
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug The checksums for the downloaded MultiWOZ dataset and source MultiWOZ dataset aren't matching. ## Steps to reproduce the bug Both of the below dataset versions yield the checksum error: ```python from datasets import load_dataset dataset = load_dataset('multi_woz_v22', 'v2.2') dataset = load_dataset('multi_woz_v22', 'v2.2_active_only') ``` ## Expected results For the above calls to `load_dataset` to work. ## Actual results NonMatchingChecksumError. Traceback: > Traceback (most recent call last): File "/Users/brady/anaconda3/envs/elysium/lib/python3.8/site-packages/IPython/core/interactiveshell.py", line 3441, in run_code exec(code_obj, self.user_global_ns, self.user_ns) File "<ipython-input-15-4e91280e112e>", line 1, in <module> dataset = load_dataset('multi_woz_v22', 'v2.2') File "/Users/brady/anaconda3/envs/elysium/lib/python3.8/site-packages/datasets/load.py", line 847, in load_dataset builder_instance.download_and_prepare( File "/Users/brady/anaconda3/envs/elysium/lib/python3.8/site-packages/datasets/builder.py", line 615, in download_and_prepare self._download_and_prepare( File "/Users/brady/anaconda3/envs/elysium/lib/python3.8/site-packages/datasets/builder.py", line 675, in _download_and_prepare verify_checksums( File "/Users/brady/anaconda3/envs/elysium/lib/python3.8/site-packages/datasets/utils/info_utils.py", line 40, in verify_checksums raise NonMatchingChecksumError(error_msg + str(bad_urls)) datasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://github.com/budzianowski/multiwoz/raw/master/data/MultiWOZ_2.2/dialog_acts.json', 'https://github.com/budzianowski/multiwoz/raw/master/data/MultiWOZ_2.2/test/dialogues_001.json'] ## Environment info - `datasets` version: 1.11.0 - Platform: macOS-10.15.7-x86_64-i386-64bit - Python version: 3.8.10 - PyArrow version: 5.0.0
2022-03-15T16:07:02Z
https://github.com/huggingface/datasets/issues/2957
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2957/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2956/comments
https://api.github.com/repos/huggingface/datasets/issues/2956/timeline
null
null
null
I_kwDODunzps473H-_
open
[]
null
2,956
{ "avatar_url": "https://avatars.githubusercontent.com/u/55560583?v=4", "events_url": "https://api.github.com/users/SaulLu/events{/privacy}", "followers_url": "https://api.github.com/users/SaulLu/followers", "following_url": "https://api.github.com/users/SaulLu/following{/other_user}", "gists_url": "https://api.github.com/users/SaulLu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SaulLu", "id": 55560583, "login": "SaulLu", "node_id": "MDQ6VXNlcjU1NTYwNTgz", "organizations_url": "https://api.github.com/users/SaulLu/orgs", "received_events_url": "https://api.github.com/users/SaulLu/received_events", "repos_url": "https://api.github.com/users/SaulLu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SaulLu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SaulLu/subscriptions", "type": "User", "url": "https://api.github.com/users/SaulLu" }
Cache problem in the `load_dataset` method for local compressed file(s)
https://api.github.com/repos/huggingface/datasets/issues/2956/events
null
https://api.github.com/repos/huggingface/datasets/issues/2956/labels{/name}
2021-09-22T13:34:32Z
null
false
null
null
1,004,306,367
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2956
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug Cache problem in the `load_dataset` method: when modifying a compressed file in a local folder `load_dataset` doesn't detect the change and load the previous version. ## Steps to reproduce the bug To test it directly, I have prepared a [Google Colaboratory notebook](https://colab.research.google.com/drive/11Em_Amoc-aPGhSBIkSHU2AvEh24nVayy?usp=sharing) that shows this behavior. For this example, I have created a toy dataset at: https://huggingface.co/datasets/SaulLu/toy_struc_dataset This dataset is composed of two versions: - v1 on commit `a6beb46` which has a single example `{'id': 1, 'value': {'tag': 'a', 'value': 1}}` in file `train.jsonl.gz` - v2 on commit `e7935f4` (`main` head) which has a single example `{'attr': 1, 'id': 1, 'value': 'a'}` in file `train.jsonl.gz` With a terminal, we can start to get the v1 version of the dataset ```bash git lfs install git clone https://huggingface.co/datasets/SaulLu/toy_struc_dataset cd toy_struc_dataset git checkout a6beb46 ``` Then we can load it with python and look at the content: ```python from datasets import load_dataset path = "/content/toy_struc_dataset" dataset = load_dataset(path, data_files={"train": "*.jsonl.gz"}) print(dataset["train"][0]) ``` Output ``` {'id': 1, 'value': {'tag': 'a', 'value': 1}} # This is the example in v1 ``` With a terminal, we can now start to get the v1 version of the dataset ```bash git checkout main ``` Then we can load it with python and look at the content: ```python from datasets import load_dataset path = "/content/toy_struc_dataset" dataset = load_dataset(path, data_files={"train": "*.jsonl.gz"}) print(dataset["train"][0]) ``` Output ``` {'id': 1, 'value': {'tag': 'a', 'value': 1}} # This is the example in v1 (not v2) ``` ## Expected results The last output should have been ``` {"id":1, "value": "a", "attr": 1} # This is the example in v2 ``` ## Ideas As discussed offline with Quentin, if the cache hash was ever sensitive to changes in a compressed file we would probably not have the problem anymore. This situation leads me to suggest 2 other features: - to also have an `load_from_cache_file` argument in the "load_dataset" method - to reorganize the cache so that we can delete the caches related to a dataset (cf issue #ToBeFilledSoon) And thanks again for this great library :hugs: ## Environment info - `datasets` version: 1.12.1 - Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.12 - PyArrow version: 3.0.0
2023-08-31T16:49:01Z
https://github.com/huggingface/datasets/issues/2956
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2956/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2955/comments
https://api.github.com/repos/huggingface/datasets/issues/2955/timeline
2021-09-24T10:36:05Z
null
null
PR_kwDODunzps4sHuRu
closed
[]
false
2,955
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update legacy Python image for CI tests in Linux
https://api.github.com/repos/huggingface/datasets/issues/2955/events
null
https://api.github.com/repos/huggingface/datasets/issues/2955/labels{/name}
2021-09-22T08:25:27Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2955.diff", "html_url": "https://github.com/huggingface/datasets/pull/2955", "merged_at": "2021-09-24T10:36:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/2955.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2955" }
1,003,999,469
[]
https://api.github.com/repos/huggingface/datasets/issues/2955
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Instead of legacy, use next-generation convenience images, built from the ground up with CI, efficiency, and determinism in mind. Here are some of the highlights: - Faster spin-up time - In Docker terminology, these next-gen images will generally have fewer and smaller layers. Using these new images will lead to faster image downloads when a build starts, and a higher likelihood that the image is already cached on the host. - Improved reliability and stability - The existing legacy convenience images are rebuilt practically every day with potential changes from upstream that we cannot always test fast enough. This leads to frequent breaking changes, which is not the best environment for stable, deterministic builds. Next-gen images will only be rebuilt for security and critical-bugs, leading to more stable and deterministic images. More info: https://circleci.com/docs/2.0/circleci-images
2021-09-24T10:36:05Z
https://github.com/huggingface/datasets/pull/2955
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2955/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2954/comments
https://api.github.com/repos/huggingface/datasets/issues/2954/timeline
2021-09-28T06:55:51Z
null
null
PR_kwDODunzps4sHa8O
closed
[]
false
2,954
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Run tests in parallel
https://api.github.com/repos/huggingface/datasets/issues/2954/events
null
https://api.github.com/repos/huggingface/datasets/issues/2954/labels{/name}
2021-09-22T07:00:44Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2954.diff", "html_url": "https://github.com/huggingface/datasets/pull/2954", "merged_at": "2021-09-28T06:55:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/2954.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2954" }
1,003,904,803
[]
https://api.github.com/repos/huggingface/datasets/issues/2954
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Run CI tests in parallel to speed up the test suite. Speed up results: - Linux: from `7m 30s` to `5m 32s` - Windows: from `13m 52s` to `11m 10s`
2021-09-28T06:55:51Z
https://github.com/huggingface/datasets/pull/2954
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2954/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2953/comments
https://api.github.com/repos/huggingface/datasets/issues/2953/timeline
2021-10-21T15:16:43Z
null
completed
I_kwDODunzps47xQC1
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,953
{ "avatar_url": "https://avatars.githubusercontent.com/u/55323451?v=4", "events_url": "https://api.github.com/users/JamieSlome/events{/privacy}", "followers_url": "https://api.github.com/users/JamieSlome/followers", "following_url": "https://api.github.com/users/JamieSlome/following{/other_user}", "gists_url": "https://api.github.com/users/JamieSlome/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JamieSlome", "id": 55323451, "login": "JamieSlome", "node_id": "MDQ6VXNlcjU1MzIzNDUx", "organizations_url": "https://api.github.com/users/JamieSlome/orgs", "received_events_url": "https://api.github.com/users/JamieSlome/received_events", "repos_url": "https://api.github.com/users/JamieSlome/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JamieSlome/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JamieSlome/subscriptions", "type": "User", "url": "https://api.github.com/users/JamieSlome" }
Trying to get in touch regarding a security issue
https://api.github.com/repos/huggingface/datasets/issues/2953/events
null
https://api.github.com/repos/huggingface/datasets/issues/2953/labels{/name}
2021-09-21T15:58:13Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,002,766,517
[]
https://api.github.com/repos/huggingface/datasets/issues/2953
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Hey there! I'd like to report a security issue but cannot find contact instructions on your repository. If not a hassle, might you kindly add a `SECURITY.md` file with an email, or another contact method? GitHub [recommends](https://docs.github.com/en/code-security/getting-started/adding-a-security-policy-to-your-repository) this best practice to ensure security issues are responsibly disclosed, and it would serve as a simple instruction for security researchers in the future. Thank you for your consideration, and I look forward to hearing from you! (cc @huntr-helper)
2021-10-21T15:16:43Z
https://github.com/huggingface/datasets/issues/2953
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2953/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2952/comments
https://api.github.com/repos/huggingface/datasets/issues/2952/timeline
2021-09-21T15:30:44Z
null
null
PR_kwDODunzps4sDU8S
closed
[]
false
2,952
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix missing conda deps
https://api.github.com/repos/huggingface/datasets/issues/2952/events
null
https://api.github.com/repos/huggingface/datasets/issues/2952/labels{/name}
2021-09-21T15:23:01Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2952.diff", "html_url": "https://github.com/huggingface/datasets/pull/2952", "merged_at": "2021-09-21T15:30:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/2952.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2952" }
1,002,704,096
[]
https://api.github.com/repos/huggingface/datasets/issues/2952
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
`aiohttp` was added as a dependency in #2662 but was missing for the conda build, which causes the 1.12.0 and 1.12.1 to fail. Fix #2932.
2021-09-22T04:39:59Z
https://github.com/huggingface/datasets/pull/2952
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2952/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2951/comments
https://api.github.com/repos/huggingface/datasets/issues/2951/timeline
2021-09-21T10:14:32Z
null
null
PR_kwDODunzps4r-lGs
closed
[]
false
2,951
{ "avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4", "events_url": "https://api.github.com/users/Rocketknight1/events{/privacy}", "followers_url": "https://api.github.com/users/Rocketknight1/followers", "following_url": "https://api.github.com/users/Rocketknight1/following{/other_user}", "gists_url": "https://api.github.com/users/Rocketknight1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Rocketknight1", "id": 12866554, "login": "Rocketknight1", "node_id": "MDQ6VXNlcjEyODY2NTU0", "organizations_url": "https://api.github.com/users/Rocketknight1/orgs", "received_events_url": "https://api.github.com/users/Rocketknight1/received_events", "repos_url": "https://api.github.com/users/Rocketknight1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Rocketknight1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Rocketknight1/subscriptions", "type": "User", "url": "https://api.github.com/users/Rocketknight1" }
Dummy labels no longer on by default in `to_tf_dataset`
https://api.github.com/repos/huggingface/datasets/issues/2951/events
null
https://api.github.com/repos/huggingface/datasets/issues/2951/labels{/name}
2021-09-20T18:26:59Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2951.diff", "html_url": "https://github.com/huggingface/datasets/pull/2951", "merged_at": "2021-09-21T10:14:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/2951.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2951" }
1,001,267,888
[]
https://api.github.com/repos/huggingface/datasets/issues/2951
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
After more experimentation, I think I have a way to do things that doesn't depend on adding `dummy_labels` - they were quite a hacky solution anyway!
2021-09-21T14:00:57Z
https://github.com/huggingface/datasets/pull/2951
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2951/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2950/comments
https://api.github.com/repos/huggingface/datasets/issues/2950/timeline
2021-09-20T15:28:01Z
null
null
PR_kwDODunzps4r-AKu
closed
[]
false
2,950
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix fn kwargs in filter
https://api.github.com/repos/huggingface/datasets/issues/2950/events
null
https://api.github.com/repos/huggingface/datasets/issues/2950/labels{/name}
2021-09-20T15:10:26Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2950.diff", "html_url": "https://github.com/huggingface/datasets/pull/2950", "merged_at": "2021-09-20T15:28:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/2950.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2950" }
1,001,085,353
[]
https://api.github.com/repos/huggingface/datasets/issues/2950
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
#2836 broke the `fn_kwargs` parameter of `filter`, as mentioned in https://github.com/huggingface/datasets/issues/2927 I fixed that and added a test to make sure it doesn't happen again (for either map or filter) Fix #2927
2021-09-20T16:22:59Z
https://github.com/huggingface/datasets/pull/2950
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2950/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2949/comments
https://api.github.com/repos/huggingface/datasets/issues/2949/timeline
2021-10-01T15:39:29Z
null
null
PR_kwDODunzps4r90Pt
closed
[]
false
2,949
{ "avatar_url": "https://avatars.githubusercontent.com/u/1706443?v=4", "events_url": "https://api.github.com/users/shirte/events{/privacy}", "followers_url": "https://api.github.com/users/shirte/followers", "following_url": "https://api.github.com/users/shirte/following{/other_user}", "gists_url": "https://api.github.com/users/shirte/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/shirte", "id": 1706443, "login": "shirte", "node_id": "MDQ6VXNlcjE3MDY0NDM=", "organizations_url": "https://api.github.com/users/shirte/orgs", "received_events_url": "https://api.github.com/users/shirte/received_events", "repos_url": "https://api.github.com/users/shirte/repos", "site_admin": false, "starred_url": "https://api.github.com/users/shirte/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shirte/subscriptions", "type": "User", "url": "https://api.github.com/users/shirte" }
Introduce web and wiki config in triviaqa dataset
https://api.github.com/repos/huggingface/datasets/issues/2949/events
null
https://api.github.com/repos/huggingface/datasets/issues/2949/labels{/name}
2021-09-20T14:17:23Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2949.diff", "html_url": "https://github.com/huggingface/datasets/pull/2949", "merged_at": "2021-10-01T15:39:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/2949.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2949" }
1,001,026,680
[]
https://api.github.com/repos/huggingface/datasets/issues/2949
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
The TriviaQA paper suggests that the two subsets (Wikipedia and Web) should be treated differently. There are also different leaderboards for the two sets on CodaLab. For that reason, introduce additional builder configs in the trivia_qa dataset.
2021-10-05T13:20:52Z
https://github.com/huggingface/datasets/pull/2949
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2949/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2948/comments
https://api.github.com/repos/huggingface/datasets/issues/2948/timeline
2021-09-20T13:18:28Z
null
null
PR_kwDODunzps4r9PdV
closed
[]
false
2,948
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix minor URL format in scitldr dataset
https://api.github.com/repos/huggingface/datasets/issues/2948/events
null
https://api.github.com/repos/huggingface/datasets/issues/2948/labels{/name}
2021-09-20T11:11:32Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2948.diff", "html_url": "https://github.com/huggingface/datasets/pull/2948", "merged_at": "2021-09-20T13:18:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/2948.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2948" }
1,000,844,077
[]
https://api.github.com/repos/huggingface/datasets/issues/2948
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
While investigating issue #2918, I found this minor format issues in the URLs (if runned in a Windows machine).
2021-09-20T13:18:28Z
https://github.com/huggingface/datasets/pull/2948
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2948/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2947/comments
https://api.github.com/repos/huggingface/datasets/issues/2947/timeline
2021-09-20T13:43:02Z
null
null
PR_kwDODunzps4r9GIP
closed
[]
false
2,947
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Don't use old, incompatible cache for the new `filter`
https://api.github.com/repos/huggingface/datasets/issues/2947/events
null
https://api.github.com/repos/huggingface/datasets/issues/2947/labels{/name}
2021-09-20T10:18:59Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2947.diff", "html_url": "https://github.com/huggingface/datasets/pull/2947", "merged_at": "2021-09-20T13:43:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/2947.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2947" }
1,000,798,338
[]
https://api.github.com/repos/huggingface/datasets/issues/2947
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
#2836 changed `Dataset.filter` and the resulting data that are stored in the cache are different and incompatible with the ones of the previous `filter` implementation. However the caching mechanism wasn't able to differentiate between the old and the new implementation of filter (only the method name was taken into account). This is an issue because anyone that update `datasets` and re-runs some code that uses `filter` would see an error, because the cache would try to load an incompatible `filter` result. To fix this I added the notion of versioning for dataset transform in the caching mechanism, and bumped the version of the `filter` implementation to 2.0.0 This way the new `filter` outputs are now considered different from the old ones from the caching point of view. This should fix #2943 cc @anton-l
2021-09-20T16:25:09Z
https://github.com/huggingface/datasets/pull/2947
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2947/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2946/comments
https://api.github.com/repos/huggingface/datasets/issues/2946/timeline
2021-09-20T09:35:59Z
null
null
PR_kwDODunzps4r89f8
closed
[]
false
2,946
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Update meteor score from nltk update
https://api.github.com/repos/huggingface/datasets/issues/2946/events
null
https://api.github.com/repos/huggingface/datasets/issues/2946/labels{/name}
2021-09-20T09:28:46Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2946.diff", "html_url": "https://github.com/huggingface/datasets/pull/2946", "merged_at": "2021-09-20T09:35:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/2946.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2946" }
1,000,754,824
[]
https://api.github.com/repos/huggingface/datasets/issues/2946
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
It looks like there were issues in NLTK on the way the METEOR score was computed. A fix was added in NLTK at https://github.com/nltk/nltk/pull/2763, and therefore the scoring function no longer returns the same values. I updated the score of the example in the docs
2021-09-20T09:35:59Z
https://github.com/huggingface/datasets/pull/2946
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2946/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2945/comments
https://api.github.com/repos/huggingface/datasets/issues/2945/timeline
2021-09-20T12:00:16Z
null
completed
I_kwDODunzps47pFLz
closed
[]
null
2,945
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Protect master branch
https://api.github.com/repos/huggingface/datasets/issues/2945/events
null
https://api.github.com/repos/huggingface/datasets/issues/2945/labels{/name}
2021-09-20T06:47:01Z
null
false
null
null
1,000,624,883
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2945
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.: - 00cc036fea7c7745cfe722360036ed306796a3f2 - 13ae8c98602bbad8197de3b9b425f4c78f582af1 - ... I propose to protect our master branch, so that we avoid we can accidentally make this kind of mistakes in the future: - [x] For Pull Requests using GitHub, allow only squash merging, so that only a single commit per Pull Request is merged into the master branch - Currently, simple merge commits are already disabled - I propose to disable rebase merging as well - ~~Protect the master branch from direct pushes (to avoid accidentally pushing of merge commits)~~ - ~~This protection would reject direct pushes to master branch~~ - ~~If so, for each release (when we need to commit directly to the master branch), we should previously disable the protection and re-enable it again after the release~~ - [x] Protect the master branch only from direct pushing of **merge commits** - GitHub offers the possibility to protect the master branch only from merge commits (which are the ones that introduce all the commits from the feature branch into the master branch). - No need to disable/re-enable this protection on each release This purpose of this Issue is to open a discussion about this problem and to agree in a solution.
2021-09-20T12:01:27Z
https://github.com/huggingface/datasets/issues/2945
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2945/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2944/comments
https://api.github.com/repos/huggingface/datasets/issues/2944/timeline
2021-10-08T15:31:53Z
null
completed
I_kwDODunzps47oxhy
closed
[]
null
2,944
{ "avatar_url": "https://avatars.githubusercontent.com/u/31893406?v=4", "events_url": "https://api.github.com/users/cccntu/events{/privacy}", "followers_url": "https://api.github.com/users/cccntu/followers", "following_url": "https://api.github.com/users/cccntu/following{/other_user}", "gists_url": "https://api.github.com/users/cccntu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cccntu", "id": 31893406, "login": "cccntu", "node_id": "MDQ6VXNlcjMxODkzNDA2", "organizations_url": "https://api.github.com/users/cccntu/orgs", "received_events_url": "https://api.github.com/users/cccntu/received_events", "repos_url": "https://api.github.com/users/cccntu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cccntu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cccntu/subscriptions", "type": "User", "url": "https://api.github.com/users/cccntu" }
Add `remove_columns` to `IterableDataset `
https://api.github.com/repos/huggingface/datasets/issues/2944/events
null
https://api.github.com/repos/huggingface/datasets/issues/2944/labels{/name}
2021-09-20T04:01:00Z
null
false
null
null
1,000,544,370
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "7057ff", "default": true, "description": "Good for newcomers", "id": 1935892877, "name": "good first issue", "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue" } ]
https://api.github.com/repos/huggingface/datasets/issues/2944
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. ```python from datasets import load_dataset dataset = load_dataset("c4", 'realnewslike', streaming =True, split='train') dataset = dataset.remove_columns('url') ``` ``` AttributeError: 'IterableDataset' object has no attribute 'remove_columns' ``` **Describe the solution you'd like** It would be nice to have `.remove_columns()` to match the `Datasets` api. **Describe alternatives you've considered** This can be done with a single call to `.map()`, I can try to help add this. 🤗
2021-10-08T15:31:53Z
https://github.com/huggingface/datasets/issues/2944
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2944/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2943/comments
https://api.github.com/repos/huggingface/datasets/issues/2943/timeline
2021-09-20T16:25:42Z
null
completed
I_kwDODunzps47oDUr
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
2,943
{ "avatar_url": "https://avatars.githubusercontent.com/u/26864830?v=4", "events_url": "https://api.github.com/users/anton-l/events{/privacy}", "followers_url": "https://api.github.com/users/anton-l/followers", "following_url": "https://api.github.com/users/anton-l/following{/other_user}", "gists_url": "https://api.github.com/users/anton-l/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/anton-l", "id": 26864830, "login": "anton-l", "node_id": "MDQ6VXNlcjI2ODY0ODMw", "organizations_url": "https://api.github.com/users/anton-l/orgs", "received_events_url": "https://api.github.com/users/anton-l/received_events", "repos_url": "https://api.github.com/users/anton-l/repos", "site_admin": false, "starred_url": "https://api.github.com/users/anton-l/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/anton-l/subscriptions", "type": "User", "url": "https://api.github.com/users/anton-l" }
Backwards compatibility broken for cached datasets that use `.filter()`
https://api.github.com/repos/huggingface/datasets/issues/2943/events
null
https://api.github.com/repos/huggingface/datasets/issues/2943/labels{/name}
2021-09-19T16:16:37Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
1,000,355,115
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2943
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)}` Related feature: https://github.com/huggingface/datasets/pull/2836 :question: This is probably a `wontfix` bug, since it can be solved by simply cleaning the related cache dirs, but the workaround could be useful for someone googling the error :) ## Workaround Remove the cache for the given dataset, e.g. `rm -rf ~/.cache/huggingface/datasets/librispeech_asr`. ## Steps to reproduce the bug 1. Delete `~/.cache/huggingface/datasets/librispeech_asr` if it exists. 2. `pip install datasets==1.11.0` and run the following snippet: ```python from datasets import load_dataset ids = ["1272-141231-0000"] ds = load_dataset("patrickvonplaten/librispeech_asr_dummy", "clean", split="validation") ds = ds.filter(lambda x: x["id"] in ids) ``` 3. `pip install datasets==1.12.1` and re-run the code again ## Expected results Same result as with the previous `datasets` version. ## Actual results ```bash Reusing dataset librispeech_asr (./.cache/huggingface/datasets/librispeech_asr/clean/2.1.0/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1) Loading cached processed dataset at ./.cache/huggingface/datasets/librispeech_asr/clean/2.1.0/468ec03677f46a8714ac6b5b64dba02d246a228d92cbbad7f3dc190fa039eab1/cache-cd1c29844fdbc87a.arrow Traceback (most recent call last): File "./repos/transformers/src/transformers/models/wav2vec2/try_dataset.py", line 5, in <module> ds = ds.filter(lambda x: x["id"] in ids) File "./envs/transformers/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 185, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "./envs/transformers/lib/python3.8/site-packages/datasets/fingerprint.py", line 398, in wrapper out = func(self, *args, **kwargs) File "./envs/transformers/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 2169, in filter indices = self.map( File "./envs/transformers/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 1686, in map return self._map_single( File "./envs/transformers/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 185, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "./envs/transformers/lib/python3.8/site-packages/datasets/fingerprint.py", line 398, in wrapper out = func(self, *args, **kwargs) File "./envs/transformers/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 1896, in _map_single return Dataset.from_file(cache_file_name, info=info, split=self.split) File "./envs/transformers/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 343, in from_file return cls( File "./envs/transformers/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 282, in __init__ self.info.features = self.info.features.reorder_fields_as(inferred_features) File "./envs/transformers/lib/python3.8/site-packages/datasets/features.py", line 1151, in reorder_fields_as return Features(recursive_reorder(self, other)) File "./envs/transformers/lib/python3.8/site-packages/datasets/features.py", line 1140, in recursive_reorder raise ValueError(f"Keys mismatch: between {source} and {target}" + stack_position) ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='int64', id=None), 'chapter_id': Value(dtype='int64', id=None), 'id': Value(dtype='string', id=None)} Process finished with exit code 1 ``` ## Environment info - `datasets` version: 1.12.1 - Platform: Linux-5.11.0-34-generic-x86_64-with-glibc2.17 - Python version: 3.8.10 - PyArrow version: 5.0.0
2021-09-20T16:25:43Z
https://github.com/huggingface/datasets/issues/2943
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2943/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2942/comments
https://api.github.com/repos/huggingface/datasets/issues/2942/timeline
2021-09-24T10:39:54Z
null
null
PR_kwDODunzps4r7tY6
closed
[]
false
2,942
{ "avatar_url": "https://avatars.githubusercontent.com/u/13545154?v=4", "events_url": "https://api.github.com/users/Hazoom/events{/privacy}", "followers_url": "https://api.github.com/users/Hazoom/followers", "following_url": "https://api.github.com/users/Hazoom/following{/other_user}", "gists_url": "https://api.github.com/users/Hazoom/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Hazoom", "id": 13545154, "login": "Hazoom", "node_id": "MDQ6VXNlcjEzNTQ1MTU0", "organizations_url": "https://api.github.com/users/Hazoom/orgs", "received_events_url": "https://api.github.com/users/Hazoom/received_events", "repos_url": "https://api.github.com/users/Hazoom/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Hazoom/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hazoom/subscriptions", "type": "User", "url": "https://api.github.com/users/Hazoom" }
Add SEDE dataset
https://api.github.com/repos/huggingface/datasets/issues/2942/events
null
https://api.github.com/repos/huggingface/datasets/issues/2942/labels{/name}
2021-09-19T13:11:24Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2942.diff", "html_url": "https://github.com/huggingface/datasets/pull/2942", "merged_at": "2021-09-24T10:39:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/2942.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2942" }
1,000,309,765
[]
https://api.github.com/repos/huggingface/datasets/issues/2942
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
This PR adds the SEDE dataset for the task of realistic Text-to-SQL, following the instructions of how to add a database and a dataset card. Please see our paper for more details: https://arxiv.org/abs/2106.05006
2021-09-24T10:39:55Z
https://github.com/huggingface/datasets/pull/2942
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2942/reactions" }
true