comments_url
stringlengths
70
70
timeline_url
stringlengths
70
70
closed_at
stringlengths
20
20
βŒ€
performed_via_github_app
null
state_reason
stringclasses
3 values
node_id
stringlengths
18
32
state
stringclasses
2 values
assignees
listlengths
0
4
draft
bool
2 classes
number
int64
1.61k
6.73k
user
dict
title
stringlengths
1
290
events_url
stringlengths
68
68
milestone
dict
labels_url
stringlengths
75
75
created_at
stringlengths
20
20
active_lock_reason
null
locked
bool
1 class
assignee
dict
pull_request
dict
id
int64
771M
2.18B
labels
listlengths
0
4
url
stringlengths
61
61
comments
listlengths
0
30
repository_url
stringclasses
1 value
author_association
stringclasses
3 values
body
stringlengths
0
228k
βŒ€
updated_at
stringlengths
20
20
html_url
stringlengths
49
51
reactions
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/4462/comments
https://api.github.com/repos/huggingface/datasets/issues/4462/timeline
null
null
reopened
I_kwDODunzps5LZ5Qz
open
[]
null
4,462
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
BigBench: NonMatchingSplitsSizesError when passing a dataset configuration parameter
https://api.github.com/repos/huggingface/datasets/issues/4462/events
null
https://api.github.com/repos/huggingface/datasets/issues/4462/labels{/name}
2022-06-08T17:31:24Z
null
false
null
null
1,265,079,347
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4462
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
As noticed in https://github.com/huggingface/datasets/pull/4125 when a dataset config class has a parameter that reduces the number of examples (e.g. named `max_examples`), then loading the dataset and passing `max_examples` raises `NonMatchingSplitsSizesError`. This is because it will check for expected the number of examples of the config with the same name without taking into account the `max_examples` parameter. This can be fixed by checking the expected number of examples using the **config id** instead of name. Indeed the config id corresponds to the config name + an optional suffix that depends on the config parameters
2022-07-05T07:39:55Z
https://github.com/huggingface/datasets/issues/4462
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4462/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4461/comments
https://api.github.com/repos/huggingface/datasets/issues/4461/timeline
2022-06-08T14:41:00Z
null
completed
I_kwDODunzps5LY1LD
closed
[]
null
4,461
{ "avatar_url": "https://avatars.githubusercontent.com/u/59248970?v=4", "events_url": "https://api.github.com/users/AlexNLP/events{/privacy}", "followers_url": "https://api.github.com/users/AlexNLP/followers", "following_url": "https://api.github.com/users/AlexNLP/following{/other_user}", "gists_url": "https://api.github.com/users/AlexNLP/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/AlexNLP", "id": 59248970, "login": "AlexNLP", "node_id": "MDQ6VXNlcjU5MjQ4OTcw", "organizations_url": "https://api.github.com/users/AlexNLP/orgs", "received_events_url": "https://api.github.com/users/AlexNLP/received_events", "repos_url": "https://api.github.com/users/AlexNLP/repos", "site_admin": false, "starred_url": "https://api.github.com/users/AlexNLP/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AlexNLP/subscriptions", "type": "User", "url": "https://api.github.com/users/AlexNLP" }
AttributeError: module 'datasets' has no attribute 'load_dataset'
https://api.github.com/repos/huggingface/datasets/issues/4461/events
null
https://api.github.com/repos/huggingface/datasets/issues/4461/labels{/name}
2022-06-08T13:59:20Z
null
false
null
null
1,264,800,451
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4461
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug I have piped install datasets, but this package doesn't have these attributes: load_dataset, load_metric. ## Environment info - `datasets` version: 1.9.0 - Platform: Linux-5.13.0-44-generic-x86_64-with-debian-bullseye-sid - Python version: 3.6.13 - PyArrow version: 6.0.1
2024-02-12T18:33:47Z
https://github.com/huggingface/datasets/issues/4461
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4461/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4460/comments
https://api.github.com/repos/huggingface/datasets/issues/4460/timeline
2022-07-26T19:04:21Z
null
null
PR_kwDODunzps45UHIs
closed
[]
false
4,460
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Drop Python 3.6 support
https://api.github.com/repos/huggingface/datasets/issues/4460/events
null
https://api.github.com/repos/huggingface/datasets/issues/4460/labels{/name}
2022-06-08T12:10:18Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4460.diff", "html_url": "https://github.com/huggingface/datasets/pull/4460", "merged_at": "2022-07-26T19:04:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/4460.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4460" }
1,264,644,205
[]
https://api.github.com/repos/huggingface/datasets/issues/4460
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Remove the fallback imports/checks in the code needed for Python 3.6 and update the requirements/CI files. Also, use Python types for the NumPy dtype wherever possible to avoid deprecation warnings in newer NumPy versions.
2022-07-26T19:16:39Z
https://github.com/huggingface/datasets/pull/4460
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4460/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4459/comments
https://api.github.com/repos/huggingface/datasets/issues/4459/timeline
2022-06-08T12:27:13Z
null
null
PR_kwDODunzps45UFc8
closed
[]
false
4,459
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Add and fix language tags for udhr dataset
https://api.github.com/repos/huggingface/datasets/issues/4459/events
null
https://api.github.com/repos/huggingface/datasets/issues/4459/labels{/name}
2022-06-08T12:03:42Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4459.diff", "html_url": "https://github.com/huggingface/datasets/pull/4459", "merged_at": "2022-06-08T12:27:13Z", "patch_url": "https://github.com/huggingface/datasets/pull/4459.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4459" }
1,264,636,481
[]
https://api.github.com/repos/huggingface/datasets/issues/4459
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Related to #4362.
2022-06-08T12:36:24Z
https://github.com/huggingface/datasets/pull/4459
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4459/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4457/comments
https://api.github.com/repos/huggingface/datasets/issues/4457/timeline
2022-06-14T15:59:08Z
null
null
PR_kwDODunzps45QZCU
closed
[]
false
4,457
{ "avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4", "events_url": "https://api.github.com/users/Rocketknight1/events{/privacy}", "followers_url": "https://api.github.com/users/Rocketknight1/followers", "following_url": "https://api.github.com/users/Rocketknight1/following{/other_user}", "gists_url": "https://api.github.com/users/Rocketknight1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Rocketknight1", "id": 12866554, "login": "Rocketknight1", "node_id": "MDQ6VXNlcjEyODY2NTU0", "organizations_url": "https://api.github.com/users/Rocketknight1/orgs", "received_events_url": "https://api.github.com/users/Rocketknight1/received_events", "repos_url": "https://api.github.com/users/Rocketknight1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Rocketknight1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Rocketknight1/subscriptions", "type": "User", "url": "https://api.github.com/users/Rocketknight1" }
First draft of the docs for TF + Datasets
https://api.github.com/repos/huggingface/datasets/issues/4457/events
null
https://api.github.com/repos/huggingface/datasets/issues/4457/labels{/name}
2022-06-07T16:06:48Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4457.diff", "html_url": "https://github.com/huggingface/datasets/pull/4457", "merged_at": "2022-06-14T15:59:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/4457.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4457" }
1,263,531,911
[ { "color": "0075ca", "default": true, "description": "Improvements or additions to documentation", "id": 1935892861, "name": "documentation", "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation" } ]
https://api.github.com/repos/huggingface/datasets/issues/4457
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
I might cc a few of the other TF people to take a look when this is closer to being finished, but it's still a draft for now.
2022-06-14T16:08:41Z
https://github.com/huggingface/datasets/pull/4457
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4457/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4456/comments
https://api.github.com/repos/huggingface/datasets/issues/4456/timeline
null
null
null
I_kwDODunzps5LS4jp
open
[]
null
4,456
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Workflow for Tabular data
https://api.github.com/repos/huggingface/datasets/issues/4456/events
null
https://api.github.com/repos/huggingface/datasets/issues/4456/labels{/name}
2022-06-07T12:48:22Z
null
false
null
null
1,263,241,449
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "c5def5", "default": false, "description": "Generic discussion on the library", "id": 2067400324, "name": "generic discussion", "node_id": "MDU6TGFiZWwyMDY3NDAwMzI0", "url": "https://api.github.com/repos/huggingface/datasets/labels/generic%20discussion" } ]
https://api.github.com/repos/huggingface/datasets/issues/4456
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Tabular data are treated very differently than data for NLP, audio, vision, etc. and therefore the worflow for tabular data in `datasets` is not ideal. For example for tabular data, it is common to use pandas/spark/dask to process the data, and then load the data into X and y (X is an array of features and y an array of labels), then train_test_split and finally feed the data to a machine learning model. In `datasets` the workflow is different: we use load_dataset, then map, then train_test_split (if we only have a train split) and we end up with columnar dataset splits, not formatted as X and y. Right now, it is already possible to convert a dataset from and to pandas, but there are still many things that could improve the workflow for tabular data: - be able to load the data into X and y - be able to load a dataset from the output of spark or dask (as far as I know it's usually csv or parquet files on S3/GCS/HDFS etc.) - support "unsplit" datasets explicitly, instead of putting everything in "train" by default cc @adrinjalali @merveenoyan feel free to complete/correct this :) Feel free to also share ideas of APIs that would be super intuitive in your opinion !
2023-03-06T08:53:55Z
https://github.com/huggingface/datasets/issues/4456
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 1, "heart": 0, "hooray": 1, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4456/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4455/comments
https://api.github.com/repos/huggingface/datasets/issues/4455/timeline
2022-06-08T07:16:17Z
null
null
PR_kwDODunzps45O5F9
closed
[]
false
4,455
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update data URLs in fever dataset
https://api.github.com/repos/huggingface/datasets/issues/4455/events
null
https://api.github.com/repos/huggingface/datasets/issues/4455/labels{/name}
2022-06-07T10:40:54Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4455.diff", "html_url": "https://github.com/huggingface/datasets/pull/4455", "merged_at": "2022-06-08T07:16:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/4455.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4455" }
1,263,089,067
[]
https://api.github.com/repos/huggingface/datasets/issues/4455
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
As stated in their website, data owners updated their URLs on 28/04/2022. This PR updates the data URLs. Fix #4452.
2022-06-08T07:24:54Z
https://github.com/huggingface/datasets/pull/4455
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4455/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4454/comments
https://api.github.com/repos/huggingface/datasets/issues/4454/timeline
2022-06-07T11:53:11Z
null
completed
I_kwDODunzps5LQuQd
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" } ]
null
4,454
{ "avatar_url": "https://avatars.githubusercontent.com/u/18160852?v=4", "events_url": "https://api.github.com/users/WithYouTo/events{/privacy}", "followers_url": "https://api.github.com/users/WithYouTo/followers", "following_url": "https://api.github.com/users/WithYouTo/following{/other_user}", "gists_url": "https://api.github.com/users/WithYouTo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/WithYouTo", "id": 18160852, "login": "WithYouTo", "node_id": "MDQ6VXNlcjE4MTYwODUy", "organizations_url": "https://api.github.com/users/WithYouTo/orgs", "received_events_url": "https://api.github.com/users/WithYouTo/received_events", "repos_url": "https://api.github.com/users/WithYouTo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/WithYouTo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/WithYouTo/subscriptions", "type": "User", "url": "https://api.github.com/users/WithYouTo" }
Dataset Viewer issue for Yaxin/SemEval2015
https://api.github.com/repos/huggingface/datasets/issues/4454/events
null
https://api.github.com/repos/huggingface/datasets/issues/4454/labels{/name}
2022-06-07T03:31:46Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
null
1,262,674,973
[ { "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists", "id": 1935892865, "name": "duplicate", "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate" }, { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
https://api.github.com/repos/huggingface/datasets/issues/4454
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
### Link _No response_ ### Description the link could not visit ### Owner _No response_
2022-06-07T11:53:11Z
https://github.com/huggingface/datasets/issues/4454
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4454/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4453/comments
https://api.github.com/repos/huggingface/datasets/issues/4453/timeline
2022-06-09T08:34:16Z
null
completed
I_kwDODunzps5LQuC5
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
4,453
{ "avatar_url": "https://avatars.githubusercontent.com/u/18160852?v=4", "events_url": "https://api.github.com/users/WithYouTo/events{/privacy}", "followers_url": "https://api.github.com/users/WithYouTo/followers", "following_url": "https://api.github.com/users/WithYouTo/following{/other_user}", "gists_url": "https://api.github.com/users/WithYouTo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/WithYouTo", "id": 18160852, "login": "WithYouTo", "node_id": "MDQ6VXNlcjE4MTYwODUy", "organizations_url": "https://api.github.com/users/WithYouTo/orgs", "received_events_url": "https://api.github.com/users/WithYouTo/received_events", "repos_url": "https://api.github.com/users/WithYouTo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/WithYouTo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/WithYouTo/subscriptions", "type": "User", "url": "https://api.github.com/users/WithYouTo" }
Dataset Viewer issue for Yaxin/SemEval2015
https://api.github.com/repos/huggingface/datasets/issues/4453/events
null
https://api.github.com/repos/huggingface/datasets/issues/4453/labels{/name}
2022-06-07T03:30:08Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,262,674,105
[]
https://api.github.com/repos/huggingface/datasets/issues/4453
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
### Link _No response_ ### Description _No response_ ### Owner _No response_
2022-06-09T08:34:16Z
https://github.com/huggingface/datasets/issues/4453
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4453/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4452/comments
https://api.github.com/repos/huggingface/datasets/issues/4452/timeline
2022-06-08T07:16:16Z
null
completed
I_kwDODunzps5LQKx2
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
4,452
{ "avatar_url": "https://avatars.githubusercontent.com/u/5347982?v=4", "events_url": "https://api.github.com/users/santhnm2/events{/privacy}", "followers_url": "https://api.github.com/users/santhnm2/followers", "following_url": "https://api.github.com/users/santhnm2/following{/other_user}", "gists_url": "https://api.github.com/users/santhnm2/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/santhnm2", "id": 5347982, "login": "santhnm2", "node_id": "MDQ6VXNlcjUzNDc5ODI=", "organizations_url": "https://api.github.com/users/santhnm2/orgs", "received_events_url": "https://api.github.com/users/santhnm2/received_events", "repos_url": "https://api.github.com/users/santhnm2/repos", "site_admin": false, "starred_url": "https://api.github.com/users/santhnm2/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/santhnm2/subscriptions", "type": "User", "url": "https://api.github.com/users/santhnm2" }
Trying to load FEVER dataset results in NonMatchingChecksumError
https://api.github.com/repos/huggingface/datasets/issues/4452/events
null
https://api.github.com/repos/huggingface/datasets/issues/4452/labels{/name}
2022-06-06T23:13:15Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,262,529,654
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4452
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Trying to load the `fever` dataset fails with `datasets.utils.info_utils.NonMatchingChecksumError`. I tried with `download_mode="force_redownload"` but that did not fix the error. I also tried with `ignore_verification=True` but then that raised a `json.decoder.JSONDecodeError`. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset('fever', 'v1.0') # Fails with NonMatchingChecksumError dataset = load_dataset('fever', 'v1.0', download_mode="force_redownload") # Fails with NonMatchingChecksumError dataset = load_dataset('fever', 'v1.0', ignore_verification=True)` # Fails with JSONDecodeError ``` ## Expected results I expect this call to return with no error raised. ## Actual results With `ignore_verification=False`: ``` *** datasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://s3-eu-west-1.amazonaws.com/fever.public/train.jsonl', 'https://s3-eu-west-1.amazonaws.com/fever.public/shared_task_dev.jsonl', 'https://s3-eu-west-1.amazonaws.com/fever.public/shared_task_dev_public.jsonl', 'https://s3-eu-west-1.amazonaws.com/fever.public/shared_task_test.jsonl', 'https://s3-eu-west-1.amazonaws.com/fever.public/paper_dev.jsonl', 'https://s3-eu-west-1.amazonaws.com/fever.public/paper_test.jsonl'] ``` With `ignore_verification=True`: ``` *** json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0) ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.2.3.dev0 - Platform: Linux-4.15.0-50-generic-x86_64-with-glibc2.10 - Python version: 3.8.13 - PyArrow version: 8.0.0 - Pandas version: 1.4.2
2022-12-15T13:36:40Z
https://github.com/huggingface/datasets/issues/4452
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4452/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4451/comments
https://api.github.com/repos/huggingface/datasets/issues/4451/timeline
2022-06-07T17:14:44Z
null
null
PR_kwDODunzps45LkGc
closed
[]
false
4,451
{ "avatar_url": "https://avatars.githubusercontent.com/u/8917831?v=4", "events_url": "https://api.github.com/users/JohnGiorgi/events{/privacy}", "followers_url": "https://api.github.com/users/JohnGiorgi/followers", "following_url": "https://api.github.com/users/JohnGiorgi/following{/other_user}", "gists_url": "https://api.github.com/users/JohnGiorgi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JohnGiorgi", "id": 8917831, "login": "JohnGiorgi", "node_id": "MDQ6VXNlcjg5MTc4MzE=", "organizations_url": "https://api.github.com/users/JohnGiorgi/orgs", "received_events_url": "https://api.github.com/users/JohnGiorgi/received_events", "repos_url": "https://api.github.com/users/JohnGiorgi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JohnGiorgi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JohnGiorgi/subscriptions", "type": "User", "url": "https://api.github.com/users/JohnGiorgi" }
Use newer version of multi-news with fixes
https://api.github.com/repos/huggingface/datasets/issues/4451/events
null
https://api.github.com/repos/huggingface/datasets/issues/4451/labels{/name}
2022-06-06T16:57:08Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4451.diff", "html_url": "https://github.com/huggingface/datasets/pull/4451", "merged_at": "2022-06-07T17:14:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/4451.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4451" }
1,262,103,323
[]
https://api.github.com/repos/huggingface/datasets/issues/4451
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Closes #4430.
2022-06-07T17:40:01Z
https://github.com/huggingface/datasets/pull/4451
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4451/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4450/comments
https://api.github.com/repos/huggingface/datasets/issues/4450/timeline
2022-06-06T14:43:03Z
null
null
PR_kwDODunzps45Kzwh
closed
[]
false
4,450
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Update README.md of fquad
https://api.github.com/repos/huggingface/datasets/issues/4450/events
null
https://api.github.com/repos/huggingface/datasets/issues/4450/labels{/name}
2022-06-06T13:52:41Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4450.diff", "html_url": "https://github.com/huggingface/datasets/pull/4450", "merged_at": "2022-06-06T14:43:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/4450.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4450" }
1,261,878,324
[]
https://api.github.com/repos/huggingface/datasets/issues/4450
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
null
2022-06-06T14:51:49Z
https://github.com/huggingface/datasets/pull/4450
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4450/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4449/comments
https://api.github.com/repos/huggingface/datasets/issues/4449/timeline
2022-06-06T15:44:50Z
null
completed
I_kwDODunzps5LLVX2
closed
[]
null
4,449
{ "avatar_url": "https://avatars.githubusercontent.com/u/87345839?v=4", "events_url": "https://api.github.com/users/Aeckard45/events{/privacy}", "followers_url": "https://api.github.com/users/Aeckard45/followers", "following_url": "https://api.github.com/users/Aeckard45/following{/other_user}", "gists_url": "https://api.github.com/users/Aeckard45/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Aeckard45", "id": 87345839, "login": "Aeckard45", "node_id": "MDQ6VXNlcjg3MzQ1ODM5", "organizations_url": "https://api.github.com/users/Aeckard45/orgs", "received_events_url": "https://api.github.com/users/Aeckard45/received_events", "repos_url": "https://api.github.com/users/Aeckard45/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Aeckard45/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Aeckard45/subscriptions", "type": "User", "url": "https://api.github.com/users/Aeckard45" }
Rj
https://api.github.com/repos/huggingface/datasets/issues/4449/events
null
https://api.github.com/repos/huggingface/datasets/issues/4449/labels{/name}
2022-06-06T02:24:32Z
null
false
null
null
1,261,262,326
[]
https://api.github.com/repos/huggingface/datasets/issues/4449
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
import android.content.DialogInterface; import android.database.Cursor; import android.os.Bundle; import android.view.View; import android.widget.ArrayAdapter; import android.widget.Button; import android.widget.EditText; import android.widget.Toast; import androidx.appcompat.app.AlertDialog; import androidx.appcompat.app.AppCompatActivity; public class MainActivity extends AppCompatActivity { private EditText editTextID; private EditText editTextName; private EditText editTextNum; private String name; private int number; private String ID; private dbHelper db; @Override protected void onCreate(Bundle savedInstanceState) { super.onCreate(savedInstanceState); setContentView(R.layout.activity_main); db = new dbHelper(this); editTextID = findViewById(R.id.editText1); editTextName = findViewById(R.id.editText2); editTextNum = findViewById(R.id.editText3); Button buttonSave = findViewById(R.id.button); Button buttonRead = findViewById(R.id.button2); Button buttonUpdate = findViewById(R.id.button3); Button buttonDelete = findViewById(R.id.button4); Button buttonSearch = findViewById(R.id.button5); Button buttonDeleteAll = findViewById(R.id.button6); buttonSave.setOnClickListener(new View.OnClickListener() { @Override public void onClick(View v) { name = editTextName.getText().toString(); String num = editTextNum.getText().toString(); if (name.isEmpty() || num.isEmpty()) { Toast.makeText(MainActivity.this, "Cannot Submit Empty Fields", Toast.LENGTH_SHORT).show(); } else { number = Integer.parseInt(num); try { // Insert Data db.insertData(name, number); // Clear the fields editTextID.getText().clear(); editTextName.getText().clear(); editTextNum.getText().clear(); } catch (Exception e) { e.printStackTrace(); } } } }); buttonRead.setOnClickListener(new View.OnClickListener() { @Override public void onClick(View v) { final ArrayAdapter<String> adapter = new ArrayAdapter<>(MainActivity.this, android.R.layout.simple_list_item_1); String name; String num; String id; try { Cursor cursor = db.readData(); if (cursor != null && cursor.getCount() > 0) { while (cursor.moveToNext()) { id = cursor.getString(0); // get data in column index 0 name = cursor.getString(1); // get data in column index 1 num = cursor.getString(2); // get data in column index 2 // Add SQLite data to listView adapter.add("ID :- " + id + "\n" + "Name :- " + name + "\n" + "Number :- " + num + "\n\n"); } } else { adapter.add("No Data"); } cursor.close(); } catch (Exception e) { e.printStackTrace(); } // show the saved data in alertDialog AlertDialog.Builder builder = new AlertDialog.Builder(MainActivity.this); builder.setTitle("SQLite saved data"); builder.setIcon(R.mipmap.app_icon_foreground); builder.setAdapter(adapter, new DialogInterface.OnClickListener() { @Override public void onClick(DialogInterface dialog, int which) { } }); builder.setPositiveButton("OK", new DialogInterface.OnClickListener() { @Override public void onClick(DialogInterface dialog, int which) { dialog.cancel(); } }); AlertDialog dialog = builder.create(); dialog.show(); } }); buttonUpdate.setOnClickListener(new View.OnClickListener() { @Override public void onClick(View v) { name = editTextName.getText().toString(); String num = editTextNum.getText().toString(); ID = editTextID.getText().toString(); if (name.isEmpty() || num.isEmpty() || ID.isEmpty()) { Toast.makeText(MainActivity.this, "Cannot Submit Empty Fields", Toast.LENGTH_SHORT).show(); } else { number = Integer.parseInt(num); try { // Update Data db.updateData(ID, name, number); // Clear the fields editTextID.getText().clear(); editTextName.getText().clear(); editTextNum.getText().clear(); } catch (Exception e) { e.printStackTrace(); } } } }); buttonDelete.setOnClickListener(new View.OnClickListener() { @Override public void onClick(View v) { ID = editTextID.getText().toString(); if (ID.isEmpty()) { Toast.makeText(MainActivity.this, "Please enter the ID", Toast.LENGTH_SHORT).show(); } else { try { // Delete Data db.deleteData(ID); // Clear the fields editTextID.getText().clear(); editTextName.getText().clear(); editTextNum.getText().clear(); } catch (Exception e) { e.printStackTrace(); } } } }); buttonDeleteAll.setOnClickListener(new View.OnClickListener() { @Override public void onClick(View v) { // Delete all data // You can simply delete all the data by calling this method --> db.deleteAllData(); // You can try this also AlertDialog.Builder builder = new AlertDialog.Builder(MainActivity.this); builder.setIcon(R.mipmap.app_icon_foreground); builder.setTitle("Delete All Data"); builder.setCancelable(false); builder.setMessage("Do you really need to delete your all data ?"); builder.setPositiveButton("Yes", new DialogInterface.OnClickListener() { @Override public void onClick(DialogInterface dialog, int which) { // User confirmed , now you can delete the data db.deleteAllData(); // Clear the fields editTextID.getText().clear(); editTextName.getText().clear(); editTextNum.getText().clear(); } }); builder.setNegativeButton("No", new DialogInterface.OnClickListener() { @Override public void onClick(DialogInterface dialog, int which) { // user not confirmed dialog.cancel(); } }); AlertDialog dialog = builder.create(); dialog.show(); } }); buttonSearch.setOnClickListener(new View.OnClickListener() { @Override public void onClick(View v) { ID = editTextID.getText().toString(); if (ID.isEmpty()) { Toast.makeText(MainActivity.this, "Please enter the ID", Toast.LENGTH_SHORT).show(); } else { try { // Search data Cursor cursor = db.searchData(ID); if (cursor.moveToFirst()) { editTextName.setText(cursor.getString(1)); editTextNum.setText(cursor.getString(2)); Toast.makeText(MainActivity.this, "Data successfully searched", Toast.LENGTH_SHORT).show(); } else { Toast.makeText(MainActivity.this, "ID not found", Toast.LENGTH_SHORT).show(); editTextNum.setText("ID Not found"); editTextName.setText("ID not found"); } cursor.close(); } catch (Exception e) { e.printStackTrace(); } } } }); } }
2022-06-06T15:44:50Z
https://github.com/huggingface/datasets/issues/4449
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4449/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4448/comments
https://api.github.com/repos/huggingface/datasets/issues/4448/timeline
null
null
null
I_kwDODunzps5LKNDx
open
[]
null
4,448
{ "avatar_url": "https://avatars.githubusercontent.com/u/57996478?v=4", "events_url": "https://api.github.com/users/yuvalkirstain/events{/privacy}", "followers_url": "https://api.github.com/users/yuvalkirstain/followers", "following_url": "https://api.github.com/users/yuvalkirstain/following{/other_user}", "gists_url": "https://api.github.com/users/yuvalkirstain/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yuvalkirstain", "id": 57996478, "login": "yuvalkirstain", "node_id": "MDQ6VXNlcjU3OTk2NDc4", "organizations_url": "https://api.github.com/users/yuvalkirstain/orgs", "received_events_url": "https://api.github.com/users/yuvalkirstain/received_events", "repos_url": "https://api.github.com/users/yuvalkirstain/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yuvalkirstain/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yuvalkirstain/subscriptions", "type": "User", "url": "https://api.github.com/users/yuvalkirstain" }
New Preprocessing Feature - Deduplication [Request]
https://api.github.com/repos/huggingface/datasets/issues/4448/events
null
https://api.github.com/repos/huggingface/datasets/issues/4448/labels{/name}
2022-06-05T05:32:56Z
null
false
null
null
1,260,966,129
[ { "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists", "id": 1935892865, "name": "duplicate", "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate" }, { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/4448
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
**Is your feature request related to a problem? Please describe.** Many large datasets are full of duplications and it has been shown that deduplicating datasets can lead to better performance while training, and more truthful evaluation at test-time. A feature that allows one to easily deduplicate a dataset can be cool! **Describe the solution you'd like** We can define a function and keep only the first/last data-point that yields the value according to this function. **Describe alternatives you've considered** The clear alternative is to repeat a clear boilerplate every time someone want to deduplicate a dataset.
2023-12-12T07:52:40Z
https://github.com/huggingface/datasets/issues/4448
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4448/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4447/comments
https://api.github.com/repos/huggingface/datasets/issues/4447/timeline
2022-06-06T15:41:37Z
null
null
PR_kwDODunzps45E4A-
closed
[]
false
4,447
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Minor fixes/improvements in `scene_parse_150` card
https://api.github.com/repos/huggingface/datasets/issues/4447/events
null
https://api.github.com/repos/huggingface/datasets/issues/4447/labels{/name}
2022-06-03T15:22:34Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4447.diff", "html_url": "https://github.com/huggingface/datasets/pull/4447", "merged_at": "2022-06-06T15:41:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/4447.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4447" }
1,260,041,805
[]
https://api.github.com/repos/huggingface/datasets/issues/4447
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Add `paperswithcode_id` and fix some links in the `scene_parse_150` card.
2022-06-06T15:50:25Z
https://github.com/huggingface/datasets/pull/4447
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4447/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4446/comments
https://api.github.com/repos/huggingface/datasets/issues/4446/timeline
2022-06-03T16:01:29Z
null
null
PR_kwDODunzps45E1Qb
closed
[]
false
4,446
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Add missing kwargs to docstrings
https://api.github.com/repos/huggingface/datasets/issues/4446/events
null
https://api.github.com/repos/huggingface/datasets/issues/4446/labels{/name}
2022-06-03T15:10:27Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4446.diff", "html_url": "https://github.com/huggingface/datasets/pull/4446", "merged_at": "2022-06-03T16:01:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/4446.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4446" }
1,260,028,995
[]
https://api.github.com/repos/huggingface/datasets/issues/4446
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
null
2022-06-03T16:10:09Z
https://github.com/huggingface/datasets/pull/4446
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4446/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4445/comments
https://api.github.com/repos/huggingface/datasets/issues/4445/timeline
2022-06-03T14:27:09Z
null
null
PR_kwDODunzps45EjtA
closed
[]
false
4,445
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix missing args in docstring of load_dataset_builder
https://api.github.com/repos/huggingface/datasets/issues/4445/events
null
https://api.github.com/repos/huggingface/datasets/issues/4445/labels{/name}
2022-06-03T13:55:50Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4445.diff", "html_url": "https://github.com/huggingface/datasets/pull/4445", "merged_at": "2022-06-03T14:27:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/4445.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4445" }
1,259,947,568
[]
https://api.github.com/repos/huggingface/datasets/issues/4445
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Currently, the docstring of `load_dataset_builder` only contains the first parameter `path` (no other): - https://huggingface.co/docs/datasets/v2.2.1/en/package_reference/loading_methods#datasets.load_dataset_builder.path
2022-06-03T14:35:32Z
https://github.com/huggingface/datasets/pull/4445
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4445/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4444/comments
https://api.github.com/repos/huggingface/datasets/issues/4444/timeline
2022-06-03T10:52:46Z
null
null
PR_kwDODunzps45D2XX
closed
[]
false
4,444
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix kwargs in docstrings
https://api.github.com/repos/huggingface/datasets/issues/4444/events
null
https://api.github.com/repos/huggingface/datasets/issues/4444/labels{/name}
2022-06-03T10:29:02Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4444.diff", "html_url": "https://github.com/huggingface/datasets/pull/4444", "merged_at": "2022-06-03T10:52:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/4444.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4444" }
1,259,738,209
[]
https://api.github.com/repos/huggingface/datasets/issues/4444
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
To fix the rendering of `**kwargs` in docstrings, a parentheses must be added afterwards. See: - huggingface/doc-builder/issues/235
2022-06-03T11:01:28Z
https://github.com/huggingface/datasets/pull/4444
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4444/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4443/comments
https://api.github.com/repos/huggingface/datasets/issues/4443/timeline
null
null
null
I_kwDODunzps5LFBE-
open
[]
null
4,443
{ "avatar_url": "https://avatars.githubusercontent.com/u/32382826?v=4", "events_url": "https://api.github.com/users/ZYMXIXI/events{/privacy}", "followers_url": "https://api.github.com/users/ZYMXIXI/followers", "following_url": "https://api.github.com/users/ZYMXIXI/following{/other_user}", "gists_url": "https://api.github.com/users/ZYMXIXI/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ZYMXIXI", "id": 32382826, "login": "ZYMXIXI", "node_id": "MDQ6VXNlcjMyMzgyODI2", "organizations_url": "https://api.github.com/users/ZYMXIXI/orgs", "received_events_url": "https://api.github.com/users/ZYMXIXI/received_events", "repos_url": "https://api.github.com/users/ZYMXIXI/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ZYMXIXI/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZYMXIXI/subscriptions", "type": "User", "url": "https://api.github.com/users/ZYMXIXI" }
Dataset Viewer issue for openclimatefix/nimrod-uk-1km
https://api.github.com/repos/huggingface/datasets/issues/4443/events
null
https://api.github.com/repos/huggingface/datasets/issues/4443/labels{/name}
2022-06-03T08:17:16Z
null
false
null
null
1,259,606,334
[]
https://api.github.com/repos/huggingface/datasets/issues/4443
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
### Link _No response_ ### Description _No response_ ### Owner _No response_
2023-09-25T12:15:08Z
https://github.com/huggingface/datasets/issues/4443
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4443/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4442/comments
https://api.github.com/repos/huggingface/datasets/issues/4442/timeline
2022-06-07T18:50:37Z
null
completed
I_kwDODunzps5LBIxc
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" } ]
null
4,442
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
Dataset Viewer issue for amazon_polarity
https://api.github.com/repos/huggingface/datasets/issues/4442/events
null
https://api.github.com/repos/huggingface/datasets/issues/4442/labels{/name}
2022-06-02T19:18:38Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
null
1,258,589,276
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
https://api.github.com/repos/huggingface/datasets/issues/4442
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
### Link https://huggingface.co/datasets/amazon_polarity/viewer/amazon_polarity/test ### Description For some reason the train split is OK but the test split is not for this dataset: ``` Server error Status code: 400 Exception: FileNotFoundError Message: [Errno 2] No such file or directory: '/cache/modules/datasets_modules/datasets/amazon_polarity/__init__.py' ``` ### Owner No
2022-06-07T18:50:37Z
https://github.com/huggingface/datasets/issues/4442
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4442/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4441/comments
https://api.github.com/repos/huggingface/datasets/issues/4441/timeline
2022-06-07T18:50:55Z
null
completed
I_kwDODunzps5LBDvQ
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" } ]
null
4,441
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
Dataset Viewer issue for aeslc
https://api.github.com/repos/huggingface/datasets/issues/4441/events
null
https://api.github.com/repos/huggingface/datasets/issues/4441/labels{/name}
2022-06-02T18:57:12Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
null
1,258,568,656
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
https://api.github.com/repos/huggingface/datasets/issues/4441
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
### Link https://huggingface.co/datasets/aeslc ### Description The dataset viewer can't find `dataset_infos.json` in it's cache: ``` Server error Status code: 400 Exception: FileNotFoundError Message: [Errno 2] No such file or directory: '/cache/modules/datasets_modules/datasets/aeslc/eb8e30234cf984a58ebe9f205674597ac1db2ec91e7321cd7f36864f7e3671b8/dataset_infos.json' ``` ### Owner No
2022-06-07T18:50:55Z
https://github.com/huggingface/datasets/issues/4441
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4441/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4440/comments
https://api.github.com/repos/huggingface/datasets/issues/4440/timeline
2022-06-23T16:23:02Z
null
null
PR_kwDODunzps44_io_
closed
[]
false
4,440
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
Update docs around audio and vision
https://api.github.com/repos/huggingface/datasets/issues/4440/events
null
https://api.github.com/repos/huggingface/datasets/issues/4440/labels{/name}
2022-06-02T17:42:03Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4440.diff", "html_url": "https://github.com/huggingface/datasets/pull/4440", "merged_at": "2022-06-23T16:23:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/4440.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4440" }
1,258,494,469
[ { "color": "0075ca", "default": true, "description": "Improvements or additions to documentation", "id": 1935892861, "name": "documentation", "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation" } ]
https://api.github.com/repos/huggingface/datasets/issues/4440
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
As part of the strategy to center the docs around the different modalities, this PR updates the quickstart to include audio and vision examples. This improves the developer experience by making audio and vision content more discoverable, enabling users working in these modalities to also quickly get started without digging too deeply into the docs. Other changes include: - Moved the installation guide to the Get Started section because it should be part of a user's onboarding to the library before exploring tutorials or how-to's. - Updated the native TF code at creating a `tf.data.Dataset` because it was throwing an error. The `to_tensor()` bit was redundant and removing it fixed the error (please double-check me here!). - Added some UI components to the quickstart so it's easier for users to navigate directly to the relevant section with context about what to expect. - Reverted to the code tabs for content that don't have any framework-specific text. I think this saves space compared to the code blocks. We'll still use the code blocks if the `torch` text is different from the `tf` text. Let me know what you think, especially if we should include some code samples for training a model in the audio/vision sections. I left this out since we already showed it in the NLP section. I want to keep the focus on using Datasets to load and process a dataset, and not so much the training part. Maybe we can add links to the Transformers docs instead?
2022-06-23T16:33:19Z
https://github.com/huggingface/datasets/pull/4440
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4440/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4439/comments
https://api.github.com/repos/huggingface/datasets/issues/4439/timeline
2022-06-03T08:44:16Z
null
completed
I_kwDODunzps5LAi4_
closed
[]
null
4,439
{ "avatar_url": "https://avatars.githubusercontent.com/u/13925685?v=4", "events_url": "https://api.github.com/users/drscotthawley/events{/privacy}", "followers_url": "https://api.github.com/users/drscotthawley/followers", "following_url": "https://api.github.com/users/drscotthawley/following{/other_user}", "gists_url": "https://api.github.com/users/drscotthawley/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/drscotthawley", "id": 13925685, "login": "drscotthawley", "node_id": "MDQ6VXNlcjEzOTI1Njg1", "organizations_url": "https://api.github.com/users/drscotthawley/orgs", "received_events_url": "https://api.github.com/users/drscotthawley/received_events", "repos_url": "https://api.github.com/users/drscotthawley/repos", "site_admin": false, "starred_url": "https://api.github.com/users/drscotthawley/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/drscotthawley/subscriptions", "type": "User", "url": "https://api.github.com/users/drscotthawley" }
TIMIT won't load after manual download: Errors about files that don't exist
https://api.github.com/repos/huggingface/datasets/issues/4439/events
null
https://api.github.com/repos/huggingface/datasets/issues/4439/labels{/name}
2022-06-02T16:35:56Z
null
false
null
null
1,258,434,111
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4439
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug I get the message from HuggingFace that it must be downloaded manually. From the URL provided in the message, I got to UPenn page for manual download. (UPenn apparently want $250? for the dataset??) ...So, ok, I obtained a copy from a friend and also a smaller version from Kaggle. But in both cases the HF dataloader fails; it is looking for files that don't exist anywhere in the dataset: it is looking for files with lower-case letters like "**test*" (all the filenames in both my copies are uppercase) and certain file extensions that exclude the .DOC which is provided in TIMIT: ## Steps to reproduce the bug ```python data = load_dataset('timit_asr', 'clean')['train'] ``` ## Expected results The dataset should load with no errors. ## Actual results This error message: ``` File "/home/ubuntu/envs/data2vec/lib/python3.9/site-packages/datasets/data_files.py", line 201, in resolve_patterns_locally_or_by_urls raise FileNotFoundError(error_msg) FileNotFoundError: Unable to resolve any data file that matches '['**test*', '**eval*']' at /home/ubuntu/datasets/timit with any supported extension ['csv', 'tsv', 'json', 'jsonl', 'parquet', 'txt', 'blp', 'bmp', 'dib', 'bufr', 'cur', 'pcx', 'dcx', 'dds', 'ps', 'eps', 'fit', 'fits', 'fli', 'flc', 'ftc', 'ftu', 'gbr', 'gif', 'grib', 'h5', 'hdf', 'png', 'apng', 'jp2', 'j2k', 'jpc', 'jpf', 'jpx', 'j2c', 'icns', 'ico', 'im', 'iim', 'tif', 'tiff', 'jfif', 'jpe', 'jpg', 'jpeg', 'mpg', 'mpeg', 'msp', 'pcd', 'pxr', 'pbm', 'pgm', 'ppm', 'pnm', 'psd', 'bw', 'rgb', 'rgba', 'sgi', 'ras', 'tga', 'icb', 'vda', 'vst', 'webp', 'wmf', 'emf', 'xbm', 'xpm', 'zip'] ``` But this is a strange sort of error: why is it looking for lower-case file names when all the TIMIT dataset filenames are uppercase? Why does it exclude .DOC files when the only parts of the TIMIT data set with "TEST" in them have ".DOC" extensions? ...I wonder, how was anyone able to get this to work in the first place? The files in the dataset look like the following: ``` Β³ PHONCODE.DOC Β³ PROMPTS.TXT Β³ SPKRINFO.TXT Β³ SPKRSENT.TXT Β³ TESTSET.DOC ``` ...so why are these being excluded by the dataset loader? ## Environment info - `datasets` version: 2.2.2 - Platform: Linux-5.4.0-1060-aws-x86_64-with-glibc2.27 - Python version: 3.9.9 - PyArrow version: 8.0.0 - Pandas version: 1.4.2
2022-06-03T08:44:17Z
https://github.com/huggingface/datasets/issues/4439
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4439/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4438/comments
https://api.github.com/repos/huggingface/datasets/issues/4438/timeline
2022-06-02T16:32:27Z
null
null
PR_kwDODunzps44-vhC
closed
[]
false
4,438
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix docstring of inspect_dataset
https://api.github.com/repos/huggingface/datasets/issues/4438/events
null
https://api.github.com/repos/huggingface/datasets/issues/4438/labels{/name}
2022-06-02T14:21:10Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4438.diff", "html_url": "https://github.com/huggingface/datasets/pull/4438", "merged_at": "2022-06-02T16:32:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/4438.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4438" }
1,258,255,394
[]
https://api.github.com/repos/huggingface/datasets/issues/4438
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
As pointed out by @sgugger: - huggingface/doc-builder/issues/235
2022-06-02T16:40:55Z
https://github.com/huggingface/datasets/pull/4438
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4438/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4437/comments
https://api.github.com/repos/huggingface/datasets/issues/4437/timeline
2022-06-06T15:41:25Z
null
null
PR_kwDODunzps44-uRW
closed
[]
false
4,437
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Add missing columns to `blended_skill_talk`
https://api.github.com/repos/huggingface/datasets/issues/4437/events
null
https://api.github.com/repos/huggingface/datasets/issues/4437/labels{/name}
2022-06-02T14:16:26Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4437.diff", "html_url": "https://github.com/huggingface/datasets/pull/4437", "merged_at": "2022-06-06T15:41:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/4437.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4437" }
1,258,249,582
[]
https://api.github.com/repos/huggingface/datasets/issues/4437
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Adds the missing columns to `blended_skill_talk` to align the loading logic with [ParlAI](https://github.com/facebookresearch/ParlAI/blob/main/parlai/tasks/blended_skill_talk/build.py). Fix #4426
2022-06-06T15:49:56Z
https://github.com/huggingface/datasets/pull/4437
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4437/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4436/comments
https://api.github.com/repos/huggingface/datasets/issues/4436/timeline
2022-06-02T09:24:27Z
null
null
PR_kwDODunzps449FsU
closed
[]
false
4,436
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix directory names for LDC data in timit_asr dataset
https://api.github.com/repos/huggingface/datasets/issues/4436/events
null
https://api.github.com/repos/huggingface/datasets/issues/4436/labels{/name}
2022-06-02T06:45:04Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4436.diff", "html_url": "https://github.com/huggingface/datasets/pull/4436", "merged_at": "2022-06-02T09:24:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/4436.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4436" }
1,257,758,834
[]
https://api.github.com/repos/huggingface/datasets/issues/4436
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Related to: - #4422
2022-06-02T09:32:56Z
https://github.com/huggingface/datasets/pull/4436
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4436/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4435/comments
https://api.github.com/repos/huggingface/datasets/issues/4435/timeline
2022-06-02T23:59:18Z
null
completed
I_kwDODunzps5K89_o
closed
[]
null
4,435
{ "avatar_url": "https://avatars.githubusercontent.com/u/2789441?v=4", "events_url": "https://api.github.com/users/mihail911/events{/privacy}", "followers_url": "https://api.github.com/users/mihail911/followers", "following_url": "https://api.github.com/users/mihail911/following{/other_user}", "gists_url": "https://api.github.com/users/mihail911/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mihail911", "id": 2789441, "login": "mihail911", "node_id": "MDQ6VXNlcjI3ODk0NDE=", "organizations_url": "https://api.github.com/users/mihail911/orgs", "received_events_url": "https://api.github.com/users/mihail911/received_events", "repos_url": "https://api.github.com/users/mihail911/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mihail911/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mihail911/subscriptions", "type": "User", "url": "https://api.github.com/users/mihail911" }
Load a local cached dataset that has been modified
https://api.github.com/repos/huggingface/datasets/issues/4435/events
null
https://api.github.com/repos/huggingface/datasets/issues/4435/labels{/name}
2022-06-02T01:51:49Z
null
false
null
null
1,257,496,552
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4435
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug I have loaded a dataset as follows: ``` d = load_dataset("emotion", split="validation") ``` Afterwards I make some modifications to the dataset via a `map` call: ``` d.map(some_update_func, cache_file_name=modified_dataset) ``` This generates a cached version of the dataset on my local system in the same directory as the original download of the data (/path/to/cache). Running an `ls` returns: ``` modified_dataset dataset_info.json emotion-test.arrow emotion-train.arrow emotion-validation.arrow ``` as expected. However, when I try to load up the modified cached dataset via a call to ``` modified = load_dataset("emotion", split="validation", data_files="/path/to/cache/modified_dataset") ``` it simply redownloads a new version of the dataset and dumps to a new cache rather than loading up the original modified dataset: ``` Using custom data configuration validation-cdbf51685638421b Downloading and preparing dataset emotion/validation to ... ``` How am I supposed to load the original modified local cache copy of the dataset? ## Environment info - `datasets` version: 2.2.2 - Platform: Linux-5.4.0-113-generic-x86_64-with-glibc2.17 - Python version: 3.8.13 - PyArrow version: 8.0.0 - Pandas version: 1.4.2
2022-06-02T23:59:26Z
https://github.com/huggingface/datasets/issues/4435
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4435/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4434/comments
https://api.github.com/repos/huggingface/datasets/issues/4434/timeline
2022-06-07T09:24:09Z
null
null
PR_kwDODunzps443mAr
closed
[]
false
4,434
{ "avatar_url": "https://avatars.githubusercontent.com/u/2529049?v=4", "events_url": "https://api.github.com/users/silverriver/events{/privacy}", "followers_url": "https://api.github.com/users/silverriver/followers", "following_url": "https://api.github.com/users/silverriver/following{/other_user}", "gists_url": "https://api.github.com/users/silverriver/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/silverriver", "id": 2529049, "login": "silverriver", "node_id": "MDQ6VXNlcjI1MjkwNDk=", "organizations_url": "https://api.github.com/users/silverriver/orgs", "received_events_url": "https://api.github.com/users/silverriver/received_events", "repos_url": "https://api.github.com/users/silverriver/repos", "site_admin": false, "starred_url": "https://api.github.com/users/silverriver/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/silverriver/subscriptions", "type": "User", "url": "https://api.github.com/users/silverriver" }
Fix dummy dataset generation script for handling nested types of _URLs
https://api.github.com/repos/huggingface/datasets/issues/4434/events
null
https://api.github.com/repos/huggingface/datasets/issues/4434/labels{/name}
2022-06-01T14:53:15Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4434.diff", "html_url": "https://github.com/huggingface/datasets/pull/4434", "merged_at": "2022-06-07T09:24:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/4434.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4434" }
1,256,207,321
[]
https://api.github.com/repos/huggingface/datasets/issues/4434
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
It seems that when user specify nested _URLs structures in their dataset script. An error will be raised when generating dummy dataset. I think the types of all elements in `dummy_data_dict.values()` should be checked because they may have different types. Linked to issue #4428 PS: I am not sure whether my code fix this issue in a proper way.
2022-06-07T12:08:28Z
https://github.com/huggingface/datasets/pull/4434
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4434/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4433/comments
https://api.github.com/repos/huggingface/datasets/issues/4433/timeline
2022-06-09T10:26:07Z
null
null
PR_kwDODunzps442P5L
closed
[]
false
4,433
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Fix script fetching and local path handling in `inspect_dataset` and `inspect_metric`
https://api.github.com/repos/huggingface/datasets/issues/4433/events
null
https://api.github.com/repos/huggingface/datasets/issues/4433/labels{/name}
2022-06-01T12:09:56Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4433.diff", "html_url": "https://github.com/huggingface/datasets/pull/4433", "merged_at": "2022-06-09T10:26:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/4433.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4433" }
1,255,830,758
[]
https://api.github.com/repos/huggingface/datasets/issues/4433
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Fix #4348
2022-06-09T10:34:54Z
https://github.com/huggingface/datasets/pull/4433
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 1, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4433/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4432/comments
https://api.github.com/repos/huggingface/datasets/issues/4432/timeline
2022-06-02T17:35:15Z
null
null
PR_kwDODunzps441JmK
closed
[]
false
4,432
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix builder docstring
https://api.github.com/repos/huggingface/datasets/issues/4432/events
null
https://api.github.com/repos/huggingface/datasets/issues/4432/labels{/name}
2022-06-01T09:45:30Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4432.diff", "html_url": "https://github.com/huggingface/datasets/pull/4432", "merged_at": "2022-06-02T17:35:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/4432.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4432" }
1,255,523,720
[]
https://api.github.com/repos/huggingface/datasets/issues/4432
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Currently, the args of `DatasetBuilder` do not appear in the docs: https://huggingface.co/docs/datasets/v2.1.0/en/package_reference/builder_classes#datasets.DatasetBuilder
2022-06-02T17:43:47Z
https://github.com/huggingface/datasets/pull/4432
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4432/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4431/comments
https://api.github.com/repos/huggingface/datasets/issues/4431/timeline
2022-06-11T12:31:16Z
null
null
PR_kwDODunzps44x5aG
closed
[]
false
4,431
{ "avatar_url": "https://avatars.githubusercontent.com/u/2529049?v=4", "events_url": "https://api.github.com/users/silverriver/events{/privacy}", "followers_url": "https://api.github.com/users/silverriver/followers", "following_url": "https://api.github.com/users/silverriver/following{/other_user}", "gists_url": "https://api.github.com/users/silverriver/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/silverriver", "id": 2529049, "login": "silverriver", "node_id": "MDQ6VXNlcjI1MjkwNDk=", "organizations_url": "https://api.github.com/users/silverriver/orgs", "received_events_url": "https://api.github.com/users/silverriver/received_events", "repos_url": "https://api.github.com/users/silverriver/repos", "site_admin": false, "starred_url": "https://api.github.com/users/silverriver/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/silverriver/subscriptions", "type": "User", "url": "https://api.github.com/users/silverriver" }
Add personaldialog datasets
https://api.github.com/repos/huggingface/datasets/issues/4431/events
null
https://api.github.com/repos/huggingface/datasets/issues/4431/labels{/name}
2022-06-01T01:20:40Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4431.diff", "html_url": "https://github.com/huggingface/datasets/pull/4431", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4431.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4431" }
1,254,618,948
[]
https://api.github.com/repos/huggingface/datasets/issues/4431
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
It seems that all tests are passed
2022-06-11T12:40:23Z
https://github.com/huggingface/datasets/pull/4431
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4431/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4430/comments
https://api.github.com/repos/huggingface/datasets/issues/4430/timeline
2022-06-07T17:14:44Z
null
completed
I_kwDODunzps5KxNEv
closed
[]
null
4,430
{ "avatar_url": "https://avatars.githubusercontent.com/u/8917831?v=4", "events_url": "https://api.github.com/users/JohnGiorgi/events{/privacy}", "followers_url": "https://api.github.com/users/JohnGiorgi/followers", "following_url": "https://api.github.com/users/JohnGiorgi/following{/other_user}", "gists_url": "https://api.github.com/users/JohnGiorgi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JohnGiorgi", "id": 8917831, "login": "JohnGiorgi", "node_id": "MDQ6VXNlcjg5MTc4MzE=", "organizations_url": "https://api.github.com/users/JohnGiorgi/orgs", "received_events_url": "https://api.github.com/users/JohnGiorgi/received_events", "repos_url": "https://api.github.com/users/JohnGiorgi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JohnGiorgi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JohnGiorgi/subscriptions", "type": "User", "url": "https://api.github.com/users/JohnGiorgi" }
Add ability to load newer, cleaner version of Multi-News
https://api.github.com/repos/huggingface/datasets/issues/4430/events
null
https://api.github.com/repos/huggingface/datasets/issues/4430/labels{/name}
2022-05-31T21:00:44Z
null
false
null
null
1,254,412,591
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/4430
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
**Is your feature request related to a problem? Please describe.** The [Multi-News dataloader points to the original version of the Multi-News dataset](https://github.com/huggingface/datasets/blob/12540dd75015678ec6019f258d811ee107439a73/datasets/multi_news/multi_news.py#L47), but this has [known errors in it](https://github.com/Alex-Fabbri/Multi-News/issues/11). There exists a [newer version which fixes some of these issues](https://drive.google.com/open?id=1jwBzXBVv8sfnFrlzPnSUBHEEAbpIUnFq). Unfortunately I don't think you can just replace this old URL with the new one, otherwise this could lead to issues with reproducibility. **Describe the solution you'd like** Add a new version to the Multi-News dataloader that points to the updated dataset which has fixes for some known issues. **Describe alternatives you've considered** Replace the current URL to the original version to the dataset with the URL to the version with fixes. **Additional context** Would be happy to make a PR for this, could someone maybe point me to another dataloader that has multiple versions so I can see how this is handled in `datasets`?
2022-06-07T17:14:44Z
https://github.com/huggingface/datasets/issues/4430
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4430/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4429/comments
https://api.github.com/repos/huggingface/datasets/issues/4429/timeline
2022-06-08T11:31:45Z
null
null
PR_kwDODunzps44whxN
closed
[]
false
4,429
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update builder docstring for deprecated/added arguments
https://api.github.com/repos/huggingface/datasets/issues/4429/events
null
https://api.github.com/repos/huggingface/datasets/issues/4429/labels{/name}
2022-05-31T17:37:25Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4429.diff", "html_url": "https://github.com/huggingface/datasets/pull/4429", "merged_at": "2022-06-08T11:31:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/4429.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4429" }
1,254,184,358
[]
https://api.github.com/repos/huggingface/datasets/issues/4429
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This PR updates the builder docstring with deprecated/added directives for arguments name/config_name. Follow up of: - #4414 - huggingface/doc-builder#233 First merge: - #4432
2022-06-08T11:40:18Z
https://github.com/huggingface/datasets/pull/4429
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4429/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4428/comments
https://api.github.com/repos/huggingface/datasets/issues/4428/timeline
2022-06-07T09:24:09Z
null
completed
I_kwDODunzps5Kv_AS
closed
[]
null
4,428
{ "avatar_url": "https://avatars.githubusercontent.com/u/2529049?v=4", "events_url": "https://api.github.com/users/silverriver/events{/privacy}", "followers_url": "https://api.github.com/users/silverriver/followers", "following_url": "https://api.github.com/users/silverriver/following{/other_user}", "gists_url": "https://api.github.com/users/silverriver/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/silverriver", "id": 2529049, "login": "silverriver", "node_id": "MDQ6VXNlcjI1MjkwNDk=", "organizations_url": "https://api.github.com/users/silverriver/orgs", "received_events_url": "https://api.github.com/users/silverriver/received_events", "repos_url": "https://api.github.com/users/silverriver/repos", "site_admin": false, "starred_url": "https://api.github.com/users/silverriver/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/silverriver/subscriptions", "type": "User", "url": "https://api.github.com/users/silverriver" }
Errors when building dummy data if you use nested _URLS
https://api.github.com/repos/huggingface/datasets/issues/4428/events
null
https://api.github.com/repos/huggingface/datasets/issues/4428/labels{/name}
2022-05-31T16:10:57Z
null
false
null
null
1,254,092,818
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4428
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug When making dummy data with the `datasets-cli dummy_data` tool, an error will be raised if you use a nested _URLS in your dataset script. Traceback (most recent call last): File "/home/name/LCCC/datasets/src/datasets/commands/datasets_cli.py", line 43, in <module> main() File "/home/name/LCCC/datasets/src/datasets/commands/datasets_cli.py", line 39, in main service.run() File "/home/name/LCCC/datasets/src/datasets/commands/dummy_data.py", line 311, in run self._autogenerate_dummy_data( File "/home/name/LCCC/datasets/src/datasets/commands/dummy_data.py", line 337, in _autogenerate_dummy_data dataset_builder._split_generators(dl_manager) File "/home/name/.cache/huggingface/modules/datasets_modules/datasets/personal_dialog/559332bced5eeafa7f7efc2a7c10ce02cee2a8116bbab4611c35a50ba2715b77/personal_dialog.py", line 108, in _split_generators data_dir = dl_manager.download_and_extract(urls) File "/home/name/LCCC/datasets/src/datasets/commands/dummy_data.py", line 56, in download_and_extract dummy_output = self.mock_download_manager.download(url_or_urls) File "/home/name/LCCC/datasets/src/datasets/download/mock_download_manager.py", line 130, in download return self.download_and_extract(data_url) File "/home/name/LCCC/datasets/src/datasets/download/mock_download_manager.py", line 122, in download_and_extract return self.create_dummy_data_dict(dummy_file, data_url) File "/home/name/LCCC/datasets/src/datasets/download/mock_download_manager.py", line 165, in create_dummy_data_dict if isinstance(first_value, str) and len(set(dummy_data_dict.values())) < len(dummy_data_dict.values()): TypeError: unhashable type: 'list' ## Steps to reproduce the bug You can use my dataset script implemented here: https://github.com/silverriver/datasets/blob/2ecd36760c40b8e29b1137cd19b5bad0e19c76fd/datasets/personal_dialog/personal_dialog.py ```python datasets_cli dummy_data datasets/personal_dialog --auto_generate ``` You can change https://github.com/silverriver/datasets/blob/2ecd36760c40b8e29b1137cd19b5bad0e19c76fd/datasets/personal_dialog/personal_dialog.py#L54 to ``` "train": "https://huggingface.co/datasets/silver/personal_dialog/resolve/main/dev_random.jsonl.gz" ``` before runing the above script to avoid downloading a large training data. ## Expected results The dummy data should be generated ## Actual results An error is raised. It seems that in https://github.com/huggingface/datasets/blob/12540dd75015678ec6019f258d811ee107439a73/src/datasets/download/mock_download_manager.py#L165 We only check if the first item of dummy_data_dict.values() is str. However, dummy_data_dict.values() may have the type of [str, list, list]. A simple fix would be changing https://github.com/huggingface/datasets/blob/12540dd75015678ec6019f258d811ee107439a73/src/datasets/download/mock_download_manager.py#L165 to ```python if all([isinstance(value, str) for value in dummy_data_dict.values()]) and len(set(dummy_data_dict.values())) < len(dummy_data_dict.values()): ``` But I don't know if this kinds of change may bring any side effect since I am not sure about the detail logic here. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - Platform: Linux - Python version: Python 3.9.10 - PyArrow version: 7.0.0
2022-06-07T09:24:09Z
https://github.com/huggingface/datasets/issues/4428
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4428/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4427/comments
https://api.github.com/repos/huggingface/datasets/issues/4427/timeline
2022-06-01T12:29:02Z
null
null
PR_kwDODunzps44vyGg
closed
[]
false
4,427
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Add HF.co for PRs/Issues for specific datasets
https://api.github.com/repos/huggingface/datasets/issues/4427/events
null
https://api.github.com/repos/huggingface/datasets/issues/4427/labels{/name}
2022-05-31T14:31:21Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4427.diff", "html_url": "https://github.com/huggingface/datasets/pull/4427", "merged_at": "2022-06-01T12:29:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/4427.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4427" }
1,253,959,313
[]
https://api.github.com/repos/huggingface/datasets/issues/4427
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
As in https://github.com/huggingface/transformers/pull/17485, issues and PR for datasets under a namespace have to be on the HF Hub
2022-06-01T12:37:42Z
https://github.com/huggingface/datasets/pull/4427
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4427/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4426/comments
https://api.github.com/repos/huggingface/datasets/issues/4426/timeline
2022-06-03T16:25:25Z
null
completed
I_kwDODunzps5KvM1P
closed
[]
null
4,426
{ "avatar_url": "https://avatars.githubusercontent.com/u/22641583?v=4", "events_url": "https://api.github.com/users/DrMatters/events{/privacy}", "followers_url": "https://api.github.com/users/DrMatters/followers", "following_url": "https://api.github.com/users/DrMatters/following{/other_user}", "gists_url": "https://api.github.com/users/DrMatters/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/DrMatters", "id": 22641583, "login": "DrMatters", "node_id": "MDQ6VXNlcjIyNjQxNTgz", "organizations_url": "https://api.github.com/users/DrMatters/orgs", "received_events_url": "https://api.github.com/users/DrMatters/received_events", "repos_url": "https://api.github.com/users/DrMatters/repos", "site_admin": false, "starred_url": "https://api.github.com/users/DrMatters/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DrMatters/subscriptions", "type": "User", "url": "https://api.github.com/users/DrMatters" }
Add loading variable number of columns for different splits
https://api.github.com/repos/huggingface/datasets/issues/4426/events
null
https://api.github.com/repos/huggingface/datasets/issues/4426/labels{/name}
2022-05-31T13:40:16Z
null
false
null
null
1,253,887,311
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/4426
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
**Is your feature request related to a problem? Please describe.** The original dataset `blended_skill_talk` consists of different sets of columns for the different splits: (test/valid) splits have additional data column `label_candidates` that the (train) doesn't have. When loading such data, an exception occurs at table.py:cast_table_to_schema, because of mismatched columns.
2022-06-03T16:25:25Z
https://github.com/huggingface/datasets/issues/4426
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4426/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4425/comments
https://api.github.com/repos/huggingface/datasets/issues/4425/timeline
2022-06-01T14:06:51Z
null
null
PR_kwDODunzps44uuDq
closed
[]
false
4,425
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Make extensions case-insensitive in timit_asr dataset
https://api.github.com/repos/huggingface/datasets/issues/4425/events
null
https://api.github.com/repos/huggingface/datasets/issues/4425/labels{/name}
2022-05-31T10:10:04Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4425.diff", "html_url": "https://github.com/huggingface/datasets/pull/4425", "merged_at": "2022-06-01T14:06:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/4425.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4425" }
1,253,641,604
[]
https://api.github.com/repos/huggingface/datasets/issues/4425
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Related to #4422.
2022-06-01T14:15:30Z
https://github.com/huggingface/datasets/pull/4425
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4425/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4424/comments
https://api.github.com/repos/huggingface/datasets/issues/4424/timeline
2022-05-31T13:42:31Z
null
null
PR_kwDODunzps44uZBD
closed
[]
false
4,424
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix DuplicatedKeysError in timit_asr dataset
https://api.github.com/repos/huggingface/datasets/issues/4424/events
null
https://api.github.com/repos/huggingface/datasets/issues/4424/labels{/name}
2022-05-31T08:47:45Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4424.diff", "html_url": "https://github.com/huggingface/datasets/pull/4424", "merged_at": "2022-05-31T13:42:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/4424.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4424" }
1,253,542,488
[]
https://api.github.com/repos/huggingface/datasets/issues/4424
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix #4422.
2022-05-31T13:50:50Z
https://github.com/huggingface/datasets/pull/4424
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4424/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4423/comments
https://api.github.com/repos/huggingface/datasets/issues/4423/timeline
2022-06-11T12:31:42Z
null
null
PR_kwDODunzps44trdP
closed
[]
false
4,423
{ "avatar_url": "https://avatars.githubusercontent.com/u/2529049?v=4", "events_url": "https://api.github.com/users/silverriver/events{/privacy}", "followers_url": "https://api.github.com/users/silverriver/followers", "following_url": "https://api.github.com/users/silverriver/following{/other_user}", "gists_url": "https://api.github.com/users/silverriver/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/silverriver", "id": 2529049, "login": "silverriver", "node_id": "MDQ6VXNlcjI1MjkwNDk=", "organizations_url": "https://api.github.com/users/silverriver/orgs", "received_events_url": "https://api.github.com/users/silverriver/received_events", "repos_url": "https://api.github.com/users/silverriver/repos", "site_admin": false, "starred_url": "https://api.github.com/users/silverriver/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/silverriver/subscriptions", "type": "User", "url": "https://api.github.com/users/silverriver" }
Add new dataset MMChat
https://api.github.com/repos/huggingface/datasets/issues/4423/events
null
https://api.github.com/repos/huggingface/datasets/issues/4423/labels{/name}
2022-05-31T04:45:07Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4423.diff", "html_url": "https://github.com/huggingface/datasets/pull/4423", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4423.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4423" }
1,253,326,023
[]
https://api.github.com/repos/huggingface/datasets/issues/4423
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Hi, I am adding a new dataset MMChat. It seems that all tests are passed
2022-06-11T12:40:52Z
https://github.com/huggingface/datasets/pull/4423
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4423/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4422/comments
https://api.github.com/repos/huggingface/datasets/issues/4422/timeline
2022-05-31T13:42:31Z
null
completed
I_kwDODunzps5KsX-P
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
4,422
{ "avatar_url": "https://avatars.githubusercontent.com/u/992795?v=4", "events_url": "https://api.github.com/users/bhaddow/events{/privacy}", "followers_url": "https://api.github.com/users/bhaddow/followers", "following_url": "https://api.github.com/users/bhaddow/following{/other_user}", "gists_url": "https://api.github.com/users/bhaddow/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bhaddow", "id": 992795, "login": "bhaddow", "node_id": "MDQ6VXNlcjk5Mjc5NQ==", "organizations_url": "https://api.github.com/users/bhaddow/orgs", "received_events_url": "https://api.github.com/users/bhaddow/received_events", "repos_url": "https://api.github.com/users/bhaddow/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bhaddow/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhaddow/subscriptions", "type": "User", "url": "https://api.github.com/users/bhaddow" }
Cannot load timit_asr data set
https://api.github.com/repos/huggingface/datasets/issues/4422/events
null
https://api.github.com/repos/huggingface/datasets/issues/4422/labels{/name}
2022-05-30T22:00:22Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,253,146,511
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4422
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug I am trying to load the timit_asr data set. I have tried with a copy from the LDC, and a copy from deepai. In both cases they fail with a "duplicate key" error. With the LDC version I have to convert the file extensions all to upper-case before I can load it at all. ## Steps to reproduce the bug ```python timit = datasets.load_dataset("timit_asr", data_dir = "/path/to/dataset") # Sample code to reproduce the bug ``` ## Expected results The data set should load without error. It worked for me before the LDC url change. ## Actual results ``` datasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET ! Found duplicate Key: SA1 Keys should be unique and deterministic in nature ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - `datasets` version: 2.2.2 - Platform: Linux-5.4.0-90-generic-x86_64-with-glibc2.17 - Python version: 3.8.12 - PyArrow version: 8.0.0 - Pandas version: 1.4.2
2022-06-02T06:34:05Z
https://github.com/huggingface/datasets/issues/4422
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4422/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4421/comments
https://api.github.com/repos/huggingface/datasets/issues/4421/timeline
2022-06-06T15:22:50Z
null
null
PR_kwDODunzps44szxR
closed
[]
false
4,421
{ "avatar_url": "https://avatars.githubusercontent.com/u/2910707?v=4", "events_url": "https://api.github.com/users/osyvokon/events{/privacy}", "followers_url": "https://api.github.com/users/osyvokon/followers", "following_url": "https://api.github.com/users/osyvokon/following{/other_user}", "gists_url": "https://api.github.com/users/osyvokon/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/osyvokon", "id": 2910707, "login": "osyvokon", "node_id": "MDQ6VXNlcjI5MTA3MDc=", "organizations_url": "https://api.github.com/users/osyvokon/orgs", "received_events_url": "https://api.github.com/users/osyvokon/received_events", "repos_url": "https://api.github.com/users/osyvokon/repos", "site_admin": false, "starred_url": "https://api.github.com/users/osyvokon/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/osyvokon/subscriptions", "type": "User", "url": "https://api.github.com/users/osyvokon" }
Add extractor for bzip2-compressed files
https://api.github.com/repos/huggingface/datasets/issues/4421/events
null
https://api.github.com/repos/huggingface/datasets/issues/4421/labels{/name}
2022-05-30T19:19:40Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4421.diff", "html_url": "https://github.com/huggingface/datasets/pull/4421", "merged_at": "2022-06-06T15:22:49Z", "patch_url": "https://github.com/huggingface/datasets/pull/4421.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4421" }
1,253,059,467
[]
https://api.github.com/repos/huggingface/datasets/issues/4421
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This change enables loading bzipped datasets, just like any other compressed dataset.
2022-06-06T15:22:50Z
https://github.com/huggingface/datasets/pull/4421
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4421/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4420/comments
https://api.github.com/repos/huggingface/datasets/issues/4420/timeline
2023-07-11T09:33:17Z
null
completed
I_kwDODunzps5Kq0in
closed
[]
null
4,420
{ "avatar_url": "https://avatars.githubusercontent.com/u/40303490?v=4", "events_url": "https://api.github.com/users/gullabi/events{/privacy}", "followers_url": "https://api.github.com/users/gullabi/followers", "following_url": "https://api.github.com/users/gullabi/following{/other_user}", "gists_url": "https://api.github.com/users/gullabi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/gullabi", "id": 40303490, "login": "gullabi", "node_id": "MDQ6VXNlcjQwMzAzNDkw", "organizations_url": "https://api.github.com/users/gullabi/orgs", "received_events_url": "https://api.github.com/users/gullabi/received_events", "repos_url": "https://api.github.com/users/gullabi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/gullabi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gullabi/subscriptions", "type": "User", "url": "https://api.github.com/users/gullabi" }
Metric evaluation problems in multi-node, shared file system
https://api.github.com/repos/huggingface/datasets/issues/4420/events
null
https://api.github.com/repos/huggingface/datasets/issues/4420/labels{/name}
2022-05-30T13:24:05Z
null
false
null
null
1,252,739,239
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4420
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Metric evaluation fails in multi-node within a shared file system, because the master process cannot find the lock files from other nodes. (This issue was originally mentioned in the transformers repo https://github.com/huggingface/transformers/issues/17412) ## Steps to reproduce the bug 1. clone [this huggingface model](https://huggingface.co/PereLluis13/wav2vec2-xls-r-300m-ca-lm) and replace the `run_speech_recognition_ctc.py` script with the version in the gist [here](https://gist.github.com/gullabi/3f66094caa8db1c1e615dd35bd67ec71#file-run_speech_recognition_ctc-py). 2. Setup the `venv` according to the requirements of the model file plus `datasets==2.0.0`, `transformers==4.18.0` and `torch==1.9.0` 3. Launch the runner in a distributed environment which has a shared file system for two nodes, preferably with SLURM. Example [here](https://gist.github.com/gullabi/3f66094caa8db1c1e615dd35bd67ec71) Specifically for the datasets, for the distributed setup the `load_metric` is called as: ``` process_id=int(os.environ["RANK"]) num_process=int(os.environ["WORLD_SIZE"]) eval_metrics = {metric: load_metric(metric, process_id=process_id, num_process=num_process, experiment_id="slurm") for metric in data_args.eval_metrics} ``` ## Expected results The training should not fail, due to the failure of the `Metric.compute()` step. ## Actual results For the test I am executing the world size is 4, with 2 GPUs in 2 nodes. However the process is not finding the necessary lock files ``` File "/gpfs/projects/bsc88/speech/asr/wav2vec2-xls-r-300m-ca-lm/run_speech_recognition_ctc.py", line 841, in <module> main() File "/gpfs/projects/bsc88/speech/asr/wav2vec2-xls-r-300m-ca-lm/run_speech_recognition_ctc.py", line 792, in main train_result = trainer.train(resume_from_checkpoint=checkpoint) File "/gpfs/projects/bsc88/projects/speech-tech-resources/venv_amd_speech/lib/python3.7/site-packages/transformers/trainer.py", line 1497, in train self._maybe_log_save_evaluate(tr_loss, model, trial, epoch, ignore_keys_for_eval) File "/gpfs/projects/bsc88/projects/speech-tech-resources/venv_amd_speech/lib/python3.7/site-packages/transformers/trainer.py", line 1624, in _maybe_log_save_evaluate metrics = self.evaluate(ignore_keys=ignore_keys_for_eval) File "/gpfs/projects/bsc88/projects/speech-tech-resources/venv_amd_speech/lib/python3.7/site-packages/transformers/trainer.py", line 2291, in evaluate metric_key_prefix=metric_key_prefix, File "/gpfs/projects/bsc88/projects/speech-tech-resources/venv_amd_speech/lib/python3.7/site-packages/transformers/trainer.py", line 2535, in evaluation_loop metrics = self.compute_metrics(EvalPrediction(predictions=all_preds, label_ids=all_labels)) File "/gpfs/projects/bsc88/speech/asr/wav2vec2-xls-r-300m-ca-lm/run_speech_recognition_ctc.py", line 742, in compute_metrics metrics = {k: v.compute(predictions=pred_str, references=label_str) for k, v in eval_metrics.items()} File "/gpfs/projects/bsc88/speech/asr/wav2vec2-xls-r-300m-ca-lm/run_speech_recognition_ctc.py", line 742, in <dictcomp> metrics = {k: v.compute(predictions=pred_str, references=label_str) for k, v in eval_metrics.items()} File "/gpfs/projects/bsc88/projects/speech-tech-resources/venv_amd_speech/lib/python3.7/site-packages/datasets/metric.py", line 419, in compute self.add_batch(**inputs) File "/gpfs/projects/bsc88/projects/speech-tech-resources/venv_amd_speech/lib/python3.7/site-packages/datasets/metric.py", line 465, in add_batch self._init_writer() File "/gpfs/projects/bsc88/projects/speech-tech-resources/venv_amd_speech/lib/python3.7/site-packages/datasets/metric.py", line 552, in _init_writer self._check_rendez_vous() # wait for master to be ready and to let everyone go File "/gpfs/projects/bsc88/projects/speech-tech-resources/venv_amd_speech/lib/python3.7/site-packages/datasets/metric.py", line 342, in _check_rendez_vous ) from None ValueError: Expected to find locked file /home/bsc88/bsc88474/.cache/huggingface/metrics/wer/default/slurm-4-0.arrow.lock from process 3 but it doesn't exist. ``` When I look at the cache directory, I can see all the lock files in principle: ``` /home/bsc88/bsc88474/.cache/huggingface/metrics/wer/default/slurm-4-0.arrow /home/bsc88/bsc88474/.cache/huggingface/metrics/wer/default/slurm-4-0.arrow.lock /home/bsc88/bsc88474/.cache/huggingface/metrics/wer/default/slurm-4-1.arrow /home/bsc88/bsc88474/.cache/huggingface/metrics/wer/default/slurm-4-1.arrow.lock /home/bsc88/bsc88474/.cache/huggingface/metrics/wer/default/slurm-4-2.arrow /home/bsc88/bsc88474/.cache/huggingface/metrics/wer/default/slurm-4-2.arrow.lock /home/bsc88/bsc88474/.cache/huggingface/metrics/wer/default/slurm-4-3.arrow /home/bsc88/bsc88474/.cache/huggingface/metrics/wer/default/slurm-4-3.arrow.lock /home/bsc88/bsc88474/.cache/huggingface/metrics/wer/default/slurm-4-rdv.lock ``` I see that there was another related issue here https://github.com/huggingface/datasets/issues/1942, but it seems to have resolved via https://github.com/huggingface/datasets/pull/1966. Let me know if there is problem with how I am calling the `load_metric` or whether I need to make changes to the `.compute()` steps. ## Environment info - `datasets` version: 2.0.0 - Platform: Linux-4.18.0-147.8.1.el8_1.x86_64-x86_64-with-centos-8.1.1911-Core - Python version: 3.7.4 - PyArrow version: 7.0.0 - Pandas version: 1.3.0
2023-07-11T09:33:18Z
https://github.com/huggingface/datasets/issues/4420
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4420/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4419/comments
https://api.github.com/repos/huggingface/datasets/issues/4419/timeline
2022-09-30T16:01:37Z
null
completed
I_kwDODunzps5Kqfdg
closed
[]
null
4,419
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
Update `unittest` assertions over tuples from `assertEqual` to `assertTupleEqual`
https://api.github.com/repos/huggingface/datasets/issues/4419/events
null
https://api.github.com/repos/huggingface/datasets/issues/4419/labels{/name}
2022-05-30T12:13:18Z
null
false
null
null
1,252,652,896
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/4419
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
**Is your feature request related to a problem? Please describe.** So this is more a readability improvement rather than a proposal, wouldn't it be better to use `assertTupleEqual` over the tuples rather than `assertEqual`? As `unittest` added that function in `v3.1`, as detailed at https://docs.python.org/3/library/unittest.html#unittest.TestCase.assertTupleEqual, so maybe it's worth updating. Find an example of an `assertEqual` over a tuple in πŸ€— `datasets` unit tests over an `ArrowDataset` at https://github.com/huggingface/datasets/blob/0bb47271910c8a0b628dba157988372307fca1d2/tests/test_arrow_dataset.py#L570 **Describe the solution you'd like** Start slowly replacing all the `assertEqual` statements with `assertTupleEqual` if the assertion is done over a Python tuple, as we're doing with the Python lists using `assertListEqual` rather than `assertEqual`. **Additional context** If so, please let me know and I'll try to go over the tests and create a PR if applicable, otherwise, if you consider this should stay as `assertEqual` rather than `assertSequenceEqual` feel free to close this issue! Thanks πŸ€—
2022-09-30T16:01:37Z
https://github.com/huggingface/datasets/issues/4419
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4419/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4418/comments
https://api.github.com/repos/huggingface/datasets/issues/4418/timeline
2022-05-30T14:58:18Z
null
null
PR_kwDODunzps44q9pG
closed
[]
false
4,418
{ "avatar_url": "https://avatars.githubusercontent.com/u/2529049?v=4", "events_url": "https://api.github.com/users/silverriver/events{/privacy}", "followers_url": "https://api.github.com/users/silverriver/followers", "following_url": "https://api.github.com/users/silverriver/following{/other_user}", "gists_url": "https://api.github.com/users/silverriver/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/silverriver", "id": 2529049, "login": "silverriver", "node_id": "MDQ6VXNlcjI1MjkwNDk=", "organizations_url": "https://api.github.com/users/silverriver/orgs", "received_events_url": "https://api.github.com/users/silverriver/received_events", "repos_url": "https://api.github.com/users/silverriver/repos", "site_admin": false, "starred_url": "https://api.github.com/users/silverriver/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/silverriver/subscriptions", "type": "User", "url": "https://api.github.com/users/silverriver" }
Add dataset MMChat
https://api.github.com/repos/huggingface/datasets/issues/4418/events
null
https://api.github.com/repos/huggingface/datasets/issues/4418/labels{/name}
2022-05-30T10:10:40Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4418.diff", "html_url": "https://github.com/huggingface/datasets/pull/4418", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4418.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4418" }
1,252,506,268
[]
https://api.github.com/repos/huggingface/datasets/issues/4418
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2022-05-30T14:58:18Z
https://github.com/huggingface/datasets/pull/4418
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4418/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4417/comments
https://api.github.com/repos/huggingface/datasets/issues/4417/timeline
2022-09-16T14:44:19Z
null
completed
I_kwDODunzps5Knvuj
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" } ]
null
4,417
{ "avatar_url": "https://avatars.githubusercontent.com/u/32235549?v=4", "events_url": "https://api.github.com/users/StephennFernandes/events{/privacy}", "followers_url": "https://api.github.com/users/StephennFernandes/followers", "following_url": "https://api.github.com/users/StephennFernandes/following{/other_user}", "gists_url": "https://api.github.com/users/StephennFernandes/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/StephennFernandes", "id": 32235549, "login": "StephennFernandes", "node_id": "MDQ6VXNlcjMyMjM1NTQ5", "organizations_url": "https://api.github.com/users/StephennFernandes/orgs", "received_events_url": "https://api.github.com/users/StephennFernandes/received_events", "repos_url": "https://api.github.com/users/StephennFernandes/repos", "site_admin": false, "starred_url": "https://api.github.com/users/StephennFernandes/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/StephennFernandes/subscriptions", "type": "User", "url": "https://api.github.com/users/StephennFernandes" }
how to convert a dict generator into a huggingface dataset.
https://api.github.com/repos/huggingface/datasets/issues/4417/events
null
https://api.github.com/repos/huggingface/datasets/issues/4417/labels{/name}
2022-05-29T16:28:27Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
null
1,251,933,091
[ { "color": "d876e3", "default": true, "description": "Further information is requested", "id": 1935892912, "name": "question", "node_id": "MDU6TGFiZWwxOTM1ODkyOTEy", "url": "https://api.github.com/repos/huggingface/datasets/labels/question" } ]
https://api.github.com/repos/huggingface/datasets/issues/4417
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
### Link _No response_ ### Description Hey there, I have used seqio to get a well distributed mixture of samples from multiple dataset. However the resultant output from seqio is a python generator dict, which I cannot produce back into huggingface dataset. The generator contains all the samples needed for training the model but I cannot convert it into a huggingface dataset. The code looks like this: ``` for ex in seqio_data: print(ex[β€œtext”]) ``` I need to convert the seqio_data (generator) into huggingface dataset. the complete seqio code goes here: ``` import functools import seqio import tensorflow as tf import t5.data from datasets import load_dataset from t5.data import postprocessors from t5.data import preprocessors from t5.evaluation import metrics from seqio import FunctionDataSource, utils TaskRegistry = seqio.TaskRegistry def gen_dataset(split, shuffle=False, seed=None, column="text", dataset_params=None): dataset = load_dataset(**dataset_params) if shuffle: if seed: dataset = dataset.shuffle(seed=seed) else: dataset = dataset.shuffle() while True: for item in dataset[str(split)]: yield item[column] def dataset_fn(split, shuffle_files, seed=None, dataset_params=None): return tf.data.Dataset.from_generator( functools.partial(gen_dataset, split, shuffle_files, seed, dataset_params=dataset_params), output_signature=tf.TensorSpec(shape=(), dtype=tf.string, name=dataset_name) ) @utils.map_over_dataset def target_to_key(x, key_map, target_key): """Assign the value from the dataset to target_key in key_map""" return {**key_map, target_key: x} dataset_name = 'oscar-corpus/OSCAR-2109' subset= 'mr' dataset_params = {"path": dataset_name, "language":subset, "use_auth_token":True} dataset_shapes = None TaskRegistry.add( "oscar_marathi_corpus", source=seqio.FunctionDataSource( dataset_fn=functools.partial(dataset_fn, dataset_params=dataset_params), splits=("train", "validation"), caching_permitted=False, num_input_examples=dataset_shapes, ), preprocessors=[ functools.partial( target_to_key, key_map={ "targets": None, }, target_key="targets")], output_features={"targets": seqio.Feature(vocabulary=seqio.PassThroughVocabulary, add_eos=False, dtype=tf.string, rank=0)}, metric_fns=[] ) dataset = seqio.get_mixture_or_task("oscar_marathi_corpus").get_dataset( sequence_length=None, split="train", shuffle=True, num_epochs=1, shard_info=seqio.ShardInfo(index=0, num_shards=10), use_cached=False, seed=42 ) for _, ex in zip(range(5), dataset): print(ex['targets'].numpy().decode()) ``` ### Owner _No response_
2022-09-16T14:44:19Z
https://github.com/huggingface/datasets/issues/4417
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4417/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4416/comments
https://api.github.com/repos/huggingface/datasets/issues/4416/timeline
2022-06-02T09:13:46Z
null
null
PR_kwDODunzps44o7sF
closed
[]
false
4,416
{ "avatar_url": "https://avatars.githubusercontent.com/u/2529049?v=4", "events_url": "https://api.github.com/users/silverriver/events{/privacy}", "followers_url": "https://api.github.com/users/silverriver/followers", "following_url": "https://api.github.com/users/silverriver/following{/other_user}", "gists_url": "https://api.github.com/users/silverriver/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/silverriver", "id": 2529049, "login": "silverriver", "node_id": "MDQ6VXNlcjI1MjkwNDk=", "organizations_url": "https://api.github.com/users/silverriver/orgs", "received_events_url": "https://api.github.com/users/silverriver/received_events", "repos_url": "https://api.github.com/users/silverriver/repos", "site_admin": false, "starred_url": "https://api.github.com/users/silverriver/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/silverriver/subscriptions", "type": "User", "url": "https://api.github.com/users/silverriver" }
Add LCCC dataset
https://api.github.com/repos/huggingface/datasets/issues/4416/events
null
https://api.github.com/repos/huggingface/datasets/issues/4416/labels{/name}
2022-05-29T12:27:19Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4416.diff", "html_url": "https://github.com/huggingface/datasets/pull/4416", "merged_at": "2022-06-02T09:13:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/4416.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4416" }
1,251,875,763
[]
https://api.github.com/repos/huggingface/datasets/issues/4416
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Hi, I am trying to add a new dataset lccc. All tests are passed.
2022-06-15T10:28:59Z
https://github.com/huggingface/datasets/pull/4416
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4416/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4415/comments
https://api.github.com/repos/huggingface/datasets/issues/4415/timeline
2022-06-07T12:33:52Z
null
null
PR_kwDODunzps44mIJk
closed
[]
false
4,415
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Update `dataset_infos.json` with new split info in `Dataset.push_to_hub` to avoid verification error
https://api.github.com/repos/huggingface/datasets/issues/4415/events
null
https://api.github.com/repos/huggingface/datasets/issues/4415/labels{/name}
2022-05-27T17:03:42Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4415.diff", "html_url": "https://github.com/huggingface/datasets/pull/4415", "merged_at": "2022-06-07T12:33:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/4415.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4415" }
1,251,002,981
[]
https://api.github.com/repos/huggingface/datasets/issues/4415
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Update `dataset_infos.json` when pushing splits one by one via `Dataset.push_to_hub` to avoid the splits verification error. TODO: ~~- [ ] handle token + `{Audio, Image}.embed_storage`~~ - [x] tests
2022-06-07T12:42:25Z
https://github.com/huggingface/datasets/pull/4415
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4415/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4414/comments
https://api.github.com/repos/huggingface/datasets/issues/4414/timeline
2022-05-31T14:58:51Z
null
null
PR_kwDODunzps44klhY
closed
[]
false
4,414
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Rename DatasetBuilder config_name
https://api.github.com/repos/huggingface/datasets/issues/4414/events
null
https://api.github.com/repos/huggingface/datasets/issues/4414/labels{/name}
2022-05-27T09:28:02Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4414.diff", "html_url": "https://github.com/huggingface/datasets/pull/4414", "merged_at": "2022-05-31T14:58:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/4414.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4414" }
1,250,546,888
[]
https://api.github.com/repos/huggingface/datasets/issues/4414
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This PR renames the DatasetBuilder keyword argument `name` to `config_name` so that: - it avoids confusion with the attribute `DatasetBuilder.name`, which is different - it aligns with the Dataset property name `config_name`, defined in `DatasetInfoMixin.config_name` Other simpler possibility could be to rename it to just `config` instead. Please note I have only renamed this argument of DatasetBuilder because I think this refactoring has a low impact on users: we can assume this is not a public facing parameter, but private or related to the inners of our library. It would have a major impact to rename it also in: - load_dataset - load_dataset_builder: although this could also be assumed as inners... - in our CLI commands Besides the naming of `name`, I also find really confusing the naming of `path` in `load_dataset`. IMHO, they should have a more simpler and precise meaning (currently, they are too vague). I would propose (maybe for next major release): ``` load_dataset(dataset, config,... ``` instead of ``` load_dataset(path, name,... ```
2022-05-31T15:07:21Z
https://github.com/huggingface/datasets/pull/4414
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4414/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4413/comments
https://api.github.com/repos/huggingface/datasets/issues/4413/timeline
2022-06-15T07:30:46Z
null
completed
I_kwDODunzps5KhXNu
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" } ]
null
4,413
{ "avatar_url": "https://avatars.githubusercontent.com/u/24966039?v=4", "events_url": "https://api.github.com/users/dgcnz/events{/privacy}", "followers_url": "https://api.github.com/users/dgcnz/followers", "following_url": "https://api.github.com/users/dgcnz/following{/other_user}", "gists_url": "https://api.github.com/users/dgcnz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dgcnz", "id": 24966039, "login": "dgcnz", "node_id": "MDQ6VXNlcjI0OTY2MDM5", "organizations_url": "https://api.github.com/users/dgcnz/orgs", "received_events_url": "https://api.github.com/users/dgcnz/received_events", "repos_url": "https://api.github.com/users/dgcnz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dgcnz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dgcnz/subscriptions", "type": "User", "url": "https://api.github.com/users/dgcnz" }
Dataset Viewer issue for ett
https://api.github.com/repos/huggingface/datasets/issues/4413/events
null
https://api.github.com/repos/huggingface/datasets/issues/4413/labels{/name}
2022-05-27T02:12:35Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
null
1,250,259,822
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
https://api.github.com/repos/huggingface/datasets/issues/4413
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
### Link https://huggingface.co/datasets/ett ### Description Timestamp is not JSON serializable. ``` Status code: 500 Exception: Status500Error Message: Type is not JSON serializable: Timestamp ``` ### Owner No
2022-06-15T07:30:46Z
https://github.com/huggingface/datasets/issues/4413
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4413/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4412/comments
https://api.github.com/repos/huggingface/datasets/issues/4412/timeline
2022-06-01T13:04:16Z
null
null
PR_kwDODunzps44hFvq
closed
[]
false
4,412
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Skip hidden files/directories in data files resolution and `iter_files`
https://api.github.com/repos/huggingface/datasets/issues/4412/events
null
https://api.github.com/repos/huggingface/datasets/issues/4412/labels{/name}
2022-05-26T12:10:28Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4412.diff", "html_url": "https://github.com/huggingface/datasets/pull/4412", "merged_at": "2022-06-01T13:04:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/4412.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4412" }
1,249,490,179
[]
https://api.github.com/repos/huggingface/datasets/issues/4412
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Fix #4115
2022-06-15T17:11:25Z
https://github.com/huggingface/datasets/pull/4412
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4412/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4411/comments
https://api.github.com/repos/huggingface/datasets/issues/4411/timeline
2022-06-14T16:01:56Z
null
null
PR_kwDODunzps44g_yL
closed
[]
false
4,411
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
Update `_format_columns` in `remove_columns`
https://api.github.com/repos/huggingface/datasets/issues/4411/events
null
https://api.github.com/repos/huggingface/datasets/issues/4411/labels{/name}
2022-05-26T11:40:06Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4411.diff", "html_url": "https://github.com/huggingface/datasets/pull/4411", "merged_at": "2022-06-14T16:01:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/4411.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4411" }
1,249,462,390
[]
https://api.github.com/repos/huggingface/datasets/issues/4411
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
As explained at #4398, when calling `dataset.add_faiss_index` under certain conditions when calling a sequence of operations `cast_column`, `map`, and `remove_columns`, it fails as it's trying to look for already removed columns. So on, after testing some possible fixes, it seems that setting the dataset format right after removing the columns seems to be working fine, so I had to add a call to `.set_format` in the `remove_columns` function. Hope this helps!
2022-06-14T19:05:37Z
https://github.com/huggingface/datasets/pull/4411
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4411/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4410/comments
https://api.github.com/repos/huggingface/datasets/issues/4410/timeline
2022-05-26T06:40:12Z
null
null
PR_kwDODunzps44f_Td
closed
[]
false
4,410
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Remove Google Drive URL in spider dataset
https://api.github.com/repos/huggingface/datasets/issues/4410/events
null
https://api.github.com/repos/huggingface/datasets/issues/4410/labels{/name}
2022-05-26T06:17:35Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4410.diff", "html_url": "https://github.com/huggingface/datasets/pull/4410", "merged_at": "2022-05-26T06:40:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/4410.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4410" }
1,249,148,457
[]
https://api.github.com/repos/huggingface/datasets/issues/4410
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
The `spider` dataset is distributed under the [CC BY-SA 4.0](https://creativecommons.org/licenses/by-sa/4.0/legalcode) license. Fix #4401.
2022-05-26T06:48:42Z
https://github.com/huggingface/datasets/pull/4410
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4410/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4409/comments
https://api.github.com/repos/huggingface/datasets/issues/4409/timeline
2022-07-07T13:16:09Z
null
null
PR_kwDODunzps44fxiH
closed
[]
false
4,409
{ "avatar_url": "https://avatars.githubusercontent.com/u/34292279?v=4", "events_url": "https://api.github.com/users/YooSungHyun/events{/privacy}", "followers_url": "https://api.github.com/users/YooSungHyun/followers", "following_url": "https://api.github.com/users/YooSungHyun/following{/other_user}", "gists_url": "https://api.github.com/users/YooSungHyun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/YooSungHyun", "id": 34292279, "login": "YooSungHyun", "node_id": "MDQ6VXNlcjM0MjkyMjc5", "organizations_url": "https://api.github.com/users/YooSungHyun/orgs", "received_events_url": "https://api.github.com/users/YooSungHyun/received_events", "repos_url": "https://api.github.com/users/YooSungHyun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/YooSungHyun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/YooSungHyun/subscriptions", "type": "User", "url": "https://api.github.com/users/YooSungHyun" }
Update: add using pcm bytes (#4323)
https://api.github.com/repos/huggingface/datasets/issues/4409/events
null
https://api.github.com/repos/huggingface/datasets/issues/4409/labels{/name}
2022-05-26T04:26:36Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4409.diff", "html_url": "https://github.com/huggingface/datasets/pull/4409", "merged_at": "2022-07-07T13:16:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/4409.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4409" }
1,249,083,179
[]
https://api.github.com/repos/huggingface/datasets/issues/4409
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
first of all, please look #4323 why i can not use {"path","array","sampling_rate"} because sf.write(format="wav") and sf.read(BytesIO) is changed my pcm data value maybe, i think wav got header but, pcm is not. and variable naming, pcm data is "byte" type. so, "array" name is not fair i think so, i use scipy lib and numpy (that is huggingface dependency) and refer to @lhoestq answered, 1. encode -> using sampling_rate and pcm byte -> wav style byte (scipy.wavfile.write to byte) 2. byte converting using fairseq style pcm audio read [FileAudioDataset](https://github.com/facebookresearch/fairseq/blob/main/fairseq/data/audio/raw_audio_dataset.py) 4. decode -> read wavfile.read that way is not screw up my pcm byte to float data, and another audio type(wav) safety please check!
2022-07-07T13:27:29Z
https://github.com/huggingface/datasets/pull/4409
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4409/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4408/comments
https://api.github.com/repos/huggingface/datasets/issues/4408/timeline
2022-05-25T20:36:47Z
null
null
PR_kwDODunzps44ecNI
closed
[]
false
4,408
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Update imagenet gate
https://api.github.com/repos/huggingface/datasets/issues/4408/events
null
https://api.github.com/repos/huggingface/datasets/issues/4408/labels{/name}
2022-05-25T20:32:19Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4408.diff", "html_url": "https://github.com/huggingface/datasets/pull/4408", "merged_at": "2022-05-25T20:36:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/4408.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4408" }
1,248,687,574
[]
https://api.github.com/repos/huggingface/datasets/issues/4408
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
null
2022-05-25T20:45:11Z
https://github.com/huggingface/datasets/pull/4408
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4408/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4407/comments
https://api.github.com/repos/huggingface/datasets/issues/4407/timeline
2022-06-07T18:39:16Z
null
completed
I_kwDODunzps5KbTgi
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" } ]
null
4,407
{ "avatar_url": "https://avatars.githubusercontent.com/u/39762734?v=4", "events_url": "https://api.github.com/users/jiangwangyi/events{/privacy}", "followers_url": "https://api.github.com/users/jiangwangyi/followers", "following_url": "https://api.github.com/users/jiangwangyi/following{/other_user}", "gists_url": "https://api.github.com/users/jiangwangyi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jiangwangyi", "id": 39762734, "login": "jiangwangyi", "node_id": "MDQ6VXNlcjM5NzYyNzM0", "organizations_url": "https://api.github.com/users/jiangwangyi/orgs", "received_events_url": "https://api.github.com/users/jiangwangyi/received_events", "repos_url": "https://api.github.com/users/jiangwangyi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jiangwangyi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jiangwangyi/subscriptions", "type": "User", "url": "https://api.github.com/users/jiangwangyi" }
Dataset Viewer issue for conll2012_ontonotesv5
https://api.github.com/repos/huggingface/datasets/issues/4407/events
null
https://api.github.com/repos/huggingface/datasets/issues/4407/labels{/name}
2022-05-25T20:18:33Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
null
1,248,671,778
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
https://api.github.com/repos/huggingface/datasets/issues/4407
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
### Link https://huggingface.co/datasets/conll2012_ontonotesv5 ### Description Dataset viewer outage. ### Owner No
2022-06-07T18:39:16Z
https://github.com/huggingface/datasets/issues/4407
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4407/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4406/comments
https://api.github.com/repos/huggingface/datasets/issues/4406/timeline
2022-05-27T14:51:23Z
null
null
PR_kwDODunzps44ePLU
closed
[]
false
4,406
{ "avatar_url": "https://avatars.githubusercontent.com/u/58078086?v=4", "events_url": "https://api.github.com/users/lbourdois/events{/privacy}", "followers_url": "https://api.github.com/users/lbourdois/followers", "following_url": "https://api.github.com/users/lbourdois/following{/other_user}", "gists_url": "https://api.github.com/users/lbourdois/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lbourdois", "id": 58078086, "login": "lbourdois", "node_id": "MDQ6VXNlcjU4MDc4MDg2", "organizations_url": "https://api.github.com/users/lbourdois/orgs", "received_events_url": "https://api.github.com/users/lbourdois/received_events", "repos_url": "https://api.github.com/users/lbourdois/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lbourdois/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lbourdois/subscriptions", "type": "User", "url": "https://api.github.com/users/lbourdois" }
Improve language tag for PIAF dataset
https://api.github.com/repos/huggingface/datasets/issues/4406/events
null
https://api.github.com/repos/huggingface/datasets/issues/4406/labels{/name}
2022-05-25T19:41:55Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4406.diff", "html_url": "https://github.com/huggingface/datasets/pull/4406", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4406.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4406" }
1,248,626,622
[]
https://api.github.com/repos/huggingface/datasets/issues/4406
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Hi, As pointed out by @lhoestq in this discussion (https://huggingface.co/datasets/asi/wikitext_fr/discussions/1), it is not yet possible to edit datasets outside of a namespace with the Hub PR feature and that you have to go through GitHub. This modification should allow better referencing since only the xx language tags are currently taken into account and not the xx-xx.
2022-05-27T14:51:23Z
https://github.com/huggingface/datasets/pull/4406
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4406/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4405/comments
https://api.github.com/repos/huggingface/datasets/issues/4405/timeline
2022-06-07T14:27:20Z
null
completed
I_kwDODunzps5Ka7qH
closed
[]
null
4,405
{ "avatar_url": "https://avatars.githubusercontent.com/u/39762734?v=4", "events_url": "https://api.github.com/users/jiangwangyi/events{/privacy}", "followers_url": "https://api.github.com/users/jiangwangyi/followers", "following_url": "https://api.github.com/users/jiangwangyi/following{/other_user}", "gists_url": "https://api.github.com/users/jiangwangyi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jiangwangyi", "id": 39762734, "login": "jiangwangyi", "node_id": "MDQ6VXNlcjM5NzYyNzM0", "organizations_url": "https://api.github.com/users/jiangwangyi/orgs", "received_events_url": "https://api.github.com/users/jiangwangyi/received_events", "repos_url": "https://api.github.com/users/jiangwangyi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jiangwangyi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jiangwangyi/subscriptions", "type": "User", "url": "https://api.github.com/users/jiangwangyi" }
[TypeError: Couldn't cast array of type] Cannot process dataset in v2.2.2
https://api.github.com/repos/huggingface/datasets/issues/4405/events
null
https://api.github.com/repos/huggingface/datasets/issues/4405/labels{/name}
2022-05-25T18:56:43Z
null
false
null
null
1,248,574,087
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4405
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug I am trying to process the [conll2012_ontonotesv5](https://huggingface.co/datasets/conll2012_ontonotesv5) dataset in `datasets` v2.2.2 and am running into a type error when casting the features. ## Steps to reproduce the bug ```python import os from typing import ( List, Dict, ) from collections import ( defaultdict, ) from dataclasses import ( dataclass, ) from datasets import ( load_dataset, ) @dataclass class ConllConverter: path: str name: str cache_dir: str def __post_init__( self, ): self.dataset = load_dataset( path=self.path, name=self.name, cache_dir=self.cache_dir, ) def convert( self, ): class_label = self.dataset["train"].features["sentences"][0]["named_entities"].feature # label_set = list(set([ # label.split("-")[1] if label != "O" else label for label in class_label.names # ])) def prepare_chunk(token, entity): assert len(token) == len(entity) # Sequence length length = len(token) # Variable used entity_chunk = defaultdict(list) idx = flag = 0 # While loop while idx < length: if entity[idx] == "O": flag += 1 idx += 1 else: iob_tp, lab_tp = entity[idx].split("-") assert iob_tp == "B" idx += 1 while idx < length and entity[idx].startswith("I-"): idx += 1 entity_chunk[lab_tp].append(token[flag: idx]) flag = idx entity_chunk = dict(entity_chunk) # for label in label_set: # if label != "O" and label not in entity_chunk.keys(): # entity_chunk[label] = None return entity_chunk def prepare_features( batch: Dict[str, List], ) -> Dict[str, List]: sentence = [ sent for doc_sent in batch["sentences"] for sent in doc_sent ] feature = { "sentence": list(), } for sent in sentence: token = sent["words"] entity = class_label.int2str(sent["named_entities"]) entity_chunk = prepare_chunk(token, entity) sent_feat = { "token": token, "entity": entity, "entity_chunk": entity_chunk, } feature["sentence"].append(sent_feat) return feature column_names = self.dataset.column_names["train"] dataset = self.dataset.map( function=prepare_features, with_indices=False, batched=True, batch_size=3, remove_columns=column_names, num_proc=1, ) dataset.save_to_disk( dataset_dict_path=os.path.join("data", self.path, self.name) ) if __name__ == "__main__": converter = ConllConverter( path="conll2012_ontonotesv5", name="english_v4", cache_dir="cache", ) converter.convert() ``` ## Expected results I want to use the dataset to perform NER task and to change the label list into a {Entity Type: list of spans} format. ## Actual results <details> <summary>Traceback</summary> ```python Traceback (most recent call last): | 0/81 [00:00<?, ?ba/s] File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/multiprocess/pool.py", line 125, in worker result = (True, func(*args, **kwds)) File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 532, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 499, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/fingerprint.py", line 458, in wrapper out = func(self, *args, **kwargs) File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 2751, in _map_single writer.write_batch(batch) File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/arrow_writer.py", line 503, in write_batch arrays.append(pa.array(typed_sequence)) File "pyarrow/array.pxi", line 230, in pyarrow.lib.array File "pyarrow/array.pxi", line 110, in pyarrow.lib._handle_arrow_array_protocol File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/arrow_writer.py", line 198, in __arrow_array__ out = cast_array_to_feature(out, type, allow_number_to_str=not self.trying_type) File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/table.py", line 1675, in wrapper return func(array, *args, **kwargs) File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/table.py", line 1793, in cast_array_to_feature arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()] File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/table.py", line 1793, in <listcomp> arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()] File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/table.py", line 1675, in wrapper return func(array, *args, **kwargs) File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/table.py", line 1844, in cast_array_to_feature raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{feature}") TypeError: Couldn't cast array of type struct<CARDINAL: list<item: list<item: string>>, DATE: list<item: list<item: string>>, EVENT: list<item: list<item: string>>, FAC: list<item: list<item: string>>, GPE: list<item: list<item: string>>, LANGUAGE: list<item: list<item: string>>, LAW: list<item: list<item: string>>, LOC: list<item: list<item: string>>, MONEY: list<item: list<item: string>>, NORP: list<item: list<item: string>>, ORDINAL: list<item: list<item: string>>, ORG: list<item: list<item: string>>, PERCENT: list<item: list<item: string>>, PERSON: list<item: list<item: string>>, QUANTITY: list<item: list<item: string>>, TIME: list<item: list<item: string>>, WORK_OF_ART: list<item: list<item: string>>> to {'CARDINAL': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'DATE': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'EVENT': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'FAC': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'GPE': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'LAW': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'LOC': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'MONEY': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'NORP': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'ORDINAL': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'ORG': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'PERCENT': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'PERSON': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'PRODUCT': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'QUANTITY': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'TIME': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'WORK_OF_ART': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None)} """ The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home2/jiangwangyi/workspace/work/Entity/dataconverter.py", line 110, in <module> converter.convert() File "/home2/jiangwangyi/workspace/work/Entity/dataconverter.py", line 91, in convert dataset = self.dataset.map( File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/dataset_dict.py", line 770, in map { File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/dataset_dict.py", line 771, in <dictcomp> k: dataset.map( File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 2459, in map transformed_shards[index] = async_result.get() File "/home2/jiangwangyi/miniconda3/lib/python3.9/site-packages/multiprocess/pool.py", line 771, in get raise self._value TypeError: Couldn't cast array of type struct<CARDINAL: list<item: list<item: string>>, DATE: list<item: list<item: string>>, EVENT: list<item: list<item: string>>, FAC: list<item: list<item: string>>, GPE: list<item: list<item: string>>, LANGUAGE: list<item: list<item: string>>, LAW: list<item: list<item: string>>, LOC: list<item: list<item: string>>, MONEY: list<item: list<item: string>>, NORP: list<item: list<item: string>>, ORDINAL: list<item: list<item: string>>, ORG: list<item: list<item: string>>, PERCENT: list<item: list<item: string>>, PERSON: list<item: list<item: string>>, QUANTITY: list<item: list<item: string>>, TIME: list<item: list<item: string>>, WORK_OF_ART: list<item: list<item: string>>> to {'CARDINAL': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'DATE': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'EVENT': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'FAC': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'GPE': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'LAW': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'LOC': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'MONEY': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'NORP': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'ORDINAL': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'ORG': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'PERCENT': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'PERSON': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'PRODUCT': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'QUANTITY': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'TIME': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None), 'WORK_OF_ART': Sequence(feature=Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), length=-1, id=None)} ``` </details> ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.2.2 - Platform: Ubuntu 18.04 - Python version: 3.9.7 - PyArrow version: 7.0.0
2022-06-07T14:27:20Z
https://github.com/huggingface/datasets/issues/4405
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4405/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4404/comments
https://api.github.com/repos/huggingface/datasets/issues/4404/timeline
2022-06-16T10:47:53Z
null
completed
I_kwDODunzps5Ka7Xj
closed
[]
null
4,404
{ "avatar_url": "https://avatars.githubusercontent.com/u/36440?v=4", "events_url": "https://api.github.com/users/f4hy/events{/privacy}", "followers_url": "https://api.github.com/users/f4hy/followers", "following_url": "https://api.github.com/users/f4hy/following{/other_user}", "gists_url": "https://api.github.com/users/f4hy/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/f4hy", "id": 36440, "login": "f4hy", "node_id": "MDQ6VXNlcjM2NDQw", "organizations_url": "https://api.github.com/users/f4hy/orgs", "received_events_url": "https://api.github.com/users/f4hy/received_events", "repos_url": "https://api.github.com/users/f4hy/repos", "site_admin": false, "starred_url": "https://api.github.com/users/f4hy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/f4hy/subscriptions", "type": "User", "url": "https://api.github.com/users/f4hy" }
Dataset should have a `.name` field
https://api.github.com/repos/huggingface/datasets/issues/4404/events
null
https://api.github.com/repos/huggingface/datasets/issues/4404/labels{/name}
2022-05-25T18:56:08Z
null
false
null
null
1,248,572,899
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/4404
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
**Is your feature request related to a problem? Please describe.** If building pipelines that can evaluate on more than one dataset, it would be nice to be able to log results of things like `Evaluating on {dataset.name}` or `results for {dataset.name} are: {results}` Without some way of concisely identifying a dataset from the dataset object, tools which might run on more than one dataset must be passed the dataset object _and_ the name/id of the dataset being used. **Describe the solution you'd like** The DatasetInfo class should have a `name` field which is the name of a dataset. then for a given dataset if it evolves in time the `version` can be updated but its different versions of the same dataset with a unique `name`. The name could then all be accessed by `dataset.name` **Describe alternatives you've considered** For my own purposes I am considering making `NamedDataset[Dataset]` where the subclass just has a .name field. **Additional context** My guess is that most usecases are not working with more than one dataset in a given pipeline so a name is not really needed. This has surprised me though as one of the advantages of a standard dataset interface is to be able to build pipelines which can be passed in a dataset and separate responsibilities of the dataset loading from the train or eval pipeline.
2022-09-13T15:09:30Z
https://github.com/huggingface/datasets/issues/4404
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4404/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4403/comments
https://api.github.com/repos/huggingface/datasets/issues/4403/timeline
2022-05-31T08:25:02Z
null
null
PR_kwDODunzps44dcpl
closed
[]
false
4,403
{ "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "events_url": "https://api.github.com/users/thomasw21/events{/privacy}", "followers_url": "https://api.github.com/users/thomasw21/followers", "following_url": "https://api.github.com/users/thomasw21/following{/other_user}", "gists_url": "https://api.github.com/users/thomasw21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thomasw21", "id": 24695242, "login": "thomasw21", "node_id": "MDQ6VXNlcjI0Njk1MjQy", "organizations_url": "https://api.github.com/users/thomasw21/orgs", "received_events_url": "https://api.github.com/users/thomasw21/received_events", "repos_url": "https://api.github.com/users/thomasw21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thomasw21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomasw21/subscriptions", "type": "User", "url": "https://api.github.com/users/thomasw21" }
Uncomment logging deactivation for ArrowBasedBuilder
https://api.github.com/repos/huggingface/datasets/issues/4403/events
null
https://api.github.com/repos/huggingface/datasets/issues/4403/labels{/name}
2022-05-25T16:46:15Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4403.diff", "html_url": "https://github.com/huggingface/datasets/pull/4403", "merged_at": "2022-05-31T08:25:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/4403.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4403" }
1,248,390,134
[]
https://api.github.com/repos/huggingface/datasets/issues/4403
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
null
2022-05-31T08:33:36Z
https://github.com/huggingface/datasets/pull/4403
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4403/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4402/comments
https://api.github.com/repos/huggingface/datasets/issues/4402/timeline
2022-05-25T15:08:03Z
null
null
PR_kwDODunzps44cdR5
closed
[]
false
4,402
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Skip identical files in `push_to_hub` instead of overwriting
https://api.github.com/repos/huggingface/datasets/issues/4402/events
null
https://api.github.com/repos/huggingface/datasets/issues/4402/labels{/name}
2022-05-25T13:12:51Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4402.diff", "html_url": "https://github.com/huggingface/datasets/pull/4402", "merged_at": "2022-05-25T15:08:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/4402.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4402" }
1,248,078,067
[]
https://api.github.com/repos/huggingface/datasets/issues/4402
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Skip identical files instead of overwriting them to save bandwidth and circumvent (user-side/server-side) errors, which can arise when working with large datasets due to long-lasting HTTP connections, by repeating calls to `push_to_hub` to resume an upload. To be able to check if an upload can be resumed, this PR modifies the shard naming scheme from: ``` data/{split}-[0-9][0-9][0-9][0-9][0-9]-of-[0-9][0-9][0-9][0-9][0-9].parquet ``` to: ``` data/{split}-[0-9][0-9][0-9][0-9][0-9]-of-[0-9][0-9][0-9][0-9][0-9]-<SHARD_FINGERPRINT>.parquet ``` cc @LysandreJik
2022-05-25T15:16:36Z
https://github.com/huggingface/datasets/pull/4402
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4402/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4401/comments
https://api.github.com/repos/huggingface/datasets/issues/4401/timeline
2022-05-26T06:40:12Z
null
completed
I_kwDODunzps5KXlQx
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
4,401
{ "avatar_url": "https://avatars.githubusercontent.com/u/81417777?v=4", "events_url": "https://api.github.com/users/OmarAlaaeldein/events{/privacy}", "followers_url": "https://api.github.com/users/OmarAlaaeldein/followers", "following_url": "https://api.github.com/users/OmarAlaaeldein/following{/other_user}", "gists_url": "https://api.github.com/users/OmarAlaaeldein/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/OmarAlaaeldein", "id": 81417777, "login": "OmarAlaaeldein", "node_id": "MDQ6VXNlcjgxNDE3Nzc3", "organizations_url": "https://api.github.com/users/OmarAlaaeldein/orgs", "received_events_url": "https://api.github.com/users/OmarAlaaeldein/received_events", "repos_url": "https://api.github.com/users/OmarAlaaeldein/repos", "site_admin": false, "starred_url": "https://api.github.com/users/OmarAlaaeldein/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/OmarAlaaeldein/subscriptions", "type": "User", "url": "https://api.github.com/users/OmarAlaaeldein" }
"NonMatchingChecksumError" when importing 'spider' dataset
https://api.github.com/repos/huggingface/datasets/issues/4401/events
null
https://api.github.com/repos/huggingface/datasets/issues/4401/labels{/name}
2022-05-25T07:45:07Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,247,695,921
[ { "color": "8B51EF", "default": false, "description": "", "id": 4069435429, "name": "hosted-on-google-drive", "node_id": "LA_kwDODunzps7yjqgl", "url": "https://api.github.com/repos/huggingface/datasets/labels/hosted-on-google-drive" } ]
https://api.github.com/repos/huggingface/datasets/issues/4401
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug When importing 'spider' dataset [https://huggingface.co/datasets/spider] an error occurs ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset('spider') ``` ## Expected results Dataset object ## Actual results NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://drive.google.com/uc?export=download&id=1_AckYkinAnhqmRQtGsQgUKAnTHxxX5J0'] ## Environment info - `datasets` version: 2.2.2 - Platform: Windows-10-10.0.19041-SP0 - Python version: 3.7.11 - PyArrow version: 6.0.1 - Pandas version: 1.3.5
2022-05-26T06:40:12Z
https://github.com/huggingface/datasets/issues/4401
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4401/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4400/comments
https://api.github.com/repos/huggingface/datasets/issues/4400/timeline
2022-05-25T03:26:36Z
null
completed
I_kwDODunzps5KWeDN
closed
[]
null
4,400
{ "avatar_url": "https://avatars.githubusercontent.com/u/20658907?v=4", "events_url": "https://api.github.com/users/cailun01/events{/privacy}", "followers_url": "https://api.github.com/users/cailun01/followers", "following_url": "https://api.github.com/users/cailun01/following{/other_user}", "gists_url": "https://api.github.com/users/cailun01/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cailun01", "id": 20658907, "login": "cailun01", "node_id": "MDQ6VXNlcjIwNjU4OTA3", "organizations_url": "https://api.github.com/users/cailun01/orgs", "received_events_url": "https://api.github.com/users/cailun01/received_events", "repos_url": "https://api.github.com/users/cailun01/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cailun01/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cailun01/subscriptions", "type": "User", "url": "https://api.github.com/users/cailun01" }
load dataset wikitext-2-raw-v1 failed. Could not reach wikitext-2-raw-v1.py.
https://api.github.com/repos/huggingface/datasets/issues/4400/events
null
https://api.github.com/repos/huggingface/datasets/issues/4400/labels{/name}
2022-05-25T03:10:44Z
null
false
null
null
1,247,404,237
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4400
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Could not reach wikitext-2-raw-v1.py ## Steps to reproduce the bug ```python from datasets import load_dataset load_dataset("wikitext-2-raw-v1") ``` ## Expected results Download `wikitext-2-raw-v1` dataset successfully. ## Actual results ``` File "load_datasets.py", line 13, in <module> load_dataset("wikitext-2-raw-v1") File "/root/miniconda3/lib/python3.6/site-packages/datasets/load.py", line 1715, in load_dataset **config_kwargs, File "/root/miniconda3/lib/python3.6/site-packages/datasets/load.py", line 1536, in load_dataset_builder data_files=data_files, File "/root/miniconda3/lib/python3.6/site-packages/datasets/load.py", line 1282, in dataset_module_factory raise e1 from None File "/root/miniconda3/lib/python3.6/site-packages/datasets/load.py", line 1224, in dataset_module_factory dynamic_modules_path=dynamic_modules_path, File "/root/miniconda3/lib/python3.6/site-packages/datasets/load.py", line 559, in get_module local_path = self.download_loading_script(revision) File "/root/miniconda3/lib/python3.6/site-packages/datasets/load.py", line 539, in download_loading_script return cached_path(file_path, download_config=download_config) File "/root/miniconda3/lib/python3.6/site-packages/datasets/utils/file_utils.py", line 246, in cached_path download_desc=download_config.download_desc, File "/root/miniconda3/lib/python3.6/site-packages/datasets/utils/file_utils.py", line 582, in get_from_cache raise ConnectionError(f"Couldn't reach {url} ({repr(head_error)})") ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/2.2.2/datasets/wikitext-2-raw-v1/wikitext-2-raw-v1.py (ReadTimeout(ReadTimeoutError("HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Read timed out. (read timeout=100)",),)) ``` I tried to download wikitext-2-raw-v1.py by chrome and got: ![image](https://user-images.githubusercontent.com/20658907/170171595-0ca9f1da-c05a-4b57-861e-9530bfa3bdb9.png) ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.2.2 - Platform: CentOS 7 - Python version: 3.6 - PyArrow version: 3.0.0
2022-10-24T06:10:27Z
https://github.com/huggingface/datasets/issues/4400
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4400/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4399/comments
https://api.github.com/repos/huggingface/datasets/issues/4399/timeline
2022-09-12T15:30:43Z
null
completed
I_kwDODunzps5KUuvL
closed
[]
null
4,399
{ "avatar_url": "https://avatars.githubusercontent.com/u/40543?v=4", "events_url": "https://api.github.com/users/apohllo/events{/privacy}", "followers_url": "https://api.github.com/users/apohllo/followers", "following_url": "https://api.github.com/users/apohllo/following{/other_user}", "gists_url": "https://api.github.com/users/apohllo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/apohllo", "id": 40543, "login": "apohllo", "node_id": "MDQ6VXNlcjQwNTQz", "organizations_url": "https://api.github.com/users/apohllo/orgs", "received_events_url": "https://api.github.com/users/apohllo/received_events", "repos_url": "https://api.github.com/users/apohllo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/apohllo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/apohllo/subscriptions", "type": "User", "url": "https://api.github.com/users/apohllo" }
LocalDatasetModuleFactoryWithoutScript extracts invalid builder name
https://api.github.com/repos/huggingface/datasets/issues/4399/events
null
https://api.github.com/repos/huggingface/datasets/issues/4399/labels{/name}
2022-05-24T18:03:01Z
null
false
null
null
1,246,948,299
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "7057ff", "default": true, "description": "Good for newcomers", "id": 1935892877, "name": "good first issue", "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue" } ]
https://api.github.com/repos/huggingface/datasets/issues/4399
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug Trying to load a local dataset raises an error indicating that the config builder has to have a name. No error should be reported, since the call is completly valid. ## Steps to reproduce the bug ```python load_dataset("./data/some-dataset/", name="some-name") ``` ## Expected results The dataset should be loaded. ## Actual results ``` Traceback (most recent call last): File "train_lquad.py", line 19, in <module> load(tokenize_target_function, tokenize_target_function, {}, tokenizer) File "train_lquad.py", line 14, in load dataset = load_dataset("./data/lquad/", name="lquad") File "/net/pr2/scratch/people/plgapohl/python-3.8.6/lib/python3.8/site-packages/datasets/load.py", line 1708, in load_dataset builder_instance = load_dataset_builder( File "/net/pr2/scratch/people/plgapohl/python-3.8.6/lib/python3.8/site-packages/datasets/load.py", line 1560, in load_dataset_builder builder_instance: DatasetBuilder = builder_cls( File "/net/pr2/scratch/people/plgapohl/python-3.8.6/lib/python3.8/site-packages/datasets/builder.py", line 269, in __init__ self.config, self.config_id = self._create_builder_config( File "/net/pr2/scratch/people/plgapohl/python-3.8.6/lib/python3.8/site-packages/datasets/builder.py", line 403, in _create_builder_config raise ValueError(f"BuilderConfig must have a name, got {builder_config.name}") ValueError: BuilderConfig must have a name, got ``` ## Environment info - `datasets` version: 2.2.2 - Platform: Linux-4.18.0-348.20.1.el8_5.x86_64-x86_64-with-glibc2.2.5 - Python version: 3.8.6 - PyArrow version: 8.0.0 - Pandas version: 1.4.2 The error is probably in line 795 in load.py: ``` builder_kwargs = { "hash": hash, "data_files": data_files, "name": os.path.basename(self.path), "base_path": self.path, **builder_kwargs, } ``` `os.path.basename` for a directory returns an empty string, rather than the name of the directory.
2022-09-12T15:30:43Z
https://github.com/huggingface/datasets/issues/4399
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4399/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4398/comments
https://api.github.com/repos/huggingface/datasets/issues/4398/timeline
2022-06-14T16:01:56Z
null
completed
I_kwDODunzps5KTp_9
closed
[]
null
4,398
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
Calling `cast_column`/`remove_columns` and a sequence of `map` operations ends up making `faiss` fail with `ValueError`
https://api.github.com/repos/huggingface/datasets/issues/4398/events
null
https://api.github.com/repos/huggingface/datasets/issues/4398/labels{/name}
2022-05-24T14:41:34Z
null
false
null
null
1,246,666,749
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4398
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
First of all, sorry in advance for the unclear title, but this bug is weird to explain (at least for me), so I tried my best to summarize all the information in this issue. ## Describe the bug Calling a certain combination of operations over a πŸ€— `Dataset` and then trying to calculate the `faiss` index with `.add_faiss_index` ends up throwing an exception while trying to set the format back of a previously removed column. But this just happens over certain conditions... I'll present some scenarios below! ## Steps to reproduce the bug Assuming the following dataset named `sample.csv` with some IMDb data: ```csv id,title,summary 1877830,"The Batman","When a sadistic serial killer begins murdering key political figures in Gotham, Batman is forced to investigate the city's hidden corruption and question his family's involvement." 9419884,"Doctor Strange in the Multiverse of Madness","Doctor Strange teams up with a mysterious teenage girl from his dreams who can travel across multiverses, to battle multiple threats, including other-universe versions of himself, which threaten to wipe out millions across the multiverse. They seek help from Wanda the Scarlet Witch, Wong and others." 11138512,"The Northman","From visionary director Robert Eggers comes The Northman, an action-filled epic that follows a young Viking prince on his quest to avenge his father's murder." 1745960,"Top Gun: Maverick","After more than thirty years of service as one of the Navy's top aviators, Pete Mitchell is where he belongs, pushing the envelope as a courageous test pilot and dodging the advancement in rank that would ground him." ``` We'll be able to reproduce the bug using the following piece of code: ```python # Sample code to reproduce the bug from transformers import DPRContextEncoder, DPRContextEncoderTokenizer import torch torch.set_grad_enabled(False) ctx_encoder = DPRContextEncoder.from_pretrained("facebook/dpr-ctx_encoder-single-nq-base") ctx_tokenizer = DPRContextEncoderTokenizer.from_pretrained("facebook/dpr-ctx_encoder-single-nq-base") from datasets import load_dataset, Value ds = load_dataset("csv", data_files=["sample.csv"], split="train") ds = ds.cast_column("id", Value("int32")) # from `int64` to `int32` ds = ds.map(lambda x: {"inputs": f"{ctx_tokenizer.sep_token}".join(["title", "summary"])}) ds = ds.remove_columns(["title", "summary"]) def generate_embeddings(x): return {"embeddings": ctx_encoder(**ctx_tokenizer(x["inputs"], return_tensors="pt"))[0][0].numpy()} ds = ds.map(generate_embeddings) ds = ds.remove_columns("inputs") ds.add_faiss_index(column="embeddings") # It fails here! ``` The code above is an adaptation of https://huggingface.co/docs/datasets/faiss_es, for the sake of presenting the bug with a simple example. ## Expected results Ideally, the `faiss` index should be calculated over the πŸ€— `Dataset` and no exception should be triggered. ## Actual results But what happens instead is that a `ValueError: Columns ['inputs'] not in the dataset. Current columns in the dataset: ['id', 'embeddings']`, which makes no sense as that column has been previously dropped. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.2.2 - Platform: Linux-5.4.0-1074-azure-x86_64-with-glibc2.31 - Python version: 3.9.5 - PyArrow version: 8.0.0 - Pandas version: 1.4.2
2022-06-14T16:01:56Z
https://github.com/huggingface/datasets/issues/4398
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4398/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4397/comments
https://api.github.com/repos/huggingface/datasets/issues/4397/timeline
2022-05-25T13:54:08Z
null
null
PR_kwDODunzps44XcG3
closed
[]
false
4,397
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix dependency on dill version
https://api.github.com/repos/huggingface/datasets/issues/4397/events
null
https://api.github.com/repos/huggingface/datasets/issues/4397/labels{/name}
2022-05-24T13:54:23Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4397.diff", "html_url": "https://github.com/huggingface/datasets/pull/4397", "merged_at": "2022-05-25T13:54:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/4397.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4397" }
1,246,597,632
[]
https://api.github.com/repos/huggingface/datasets/issues/4397
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
We had to make a hotfix by pinning dill: - #4380 because from version 0.3.5, our custom `save_function` pickling function was raising an exception: - #4379 This PR fixes this by implementing our custom `save_function` depending on the version of dill. CC: @anivegesana This PR needs first being merged: - [x] #4384 - so that a circular import is fixed It is also convenient to merge first: - [x] #4385
2022-10-26T08:45:37Z
https://github.com/huggingface/datasets/pull/4397
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4397/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4396/comments
https://api.github.com/repos/huggingface/datasets/issues/4396/timeline
2022-05-24T05:41:00Z
null
null
PR_kwDODunzps44T0Di
closed
[]
false
4,396
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix URL in gem dataset for totto config
https://api.github.com/repos/huggingface/datasets/issues/4396/events
null
https://api.github.com/repos/huggingface/datasets/issues/4396/labels{/name}
2022-05-23T17:16:12Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4396.diff", "html_url": "https://github.com/huggingface/datasets/pull/4396", "merged_at": "2022-05-24T05:40:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/4396.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4396" }
1,245,479,399
[]
https://api.github.com/repos/huggingface/datasets/issues/4396
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
As commented in: - https://github.com/huggingface/datasets/issues/4386#issuecomment-1134902372 CC: @StevenTang1998
2022-05-24T05:49:11Z
https://github.com/huggingface/datasets/pull/4396
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4396/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4395/comments
https://api.github.com/repos/huggingface/datasets/issues/4395/timeline
2022-10-03T09:36:56Z
null
null
PR_kwDODunzps44TrBA
closed
[]
false
4,395
{ "avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4", "events_url": "https://api.github.com/users/nateraw/events{/privacy}", "followers_url": "https://api.github.com/users/nateraw/followers", "following_url": "https://api.github.com/users/nateraw/following{/other_user}", "gists_url": "https://api.github.com/users/nateraw/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nateraw", "id": 32437151, "login": "nateraw", "node_id": "MDQ6VXNlcjMyNDM3MTUx", "organizations_url": "https://api.github.com/users/nateraw/orgs", "received_events_url": "https://api.github.com/users/nateraw/received_events", "repos_url": "https://api.github.com/users/nateraw/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nateraw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nateraw/subscriptions", "type": "User", "url": "https://api.github.com/users/nateraw" }
Add Pascal VOC dataset
https://api.github.com/repos/huggingface/datasets/issues/4395/events
null
https://api.github.com/repos/huggingface/datasets/issues/4395/labels{/name}
2022-05-23T16:34:05Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4395.diff", "html_url": "https://github.com/huggingface/datasets/pull/4395", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4395.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4395" }
1,245,436,486
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
https://api.github.com/repos/huggingface/datasets/issues/4395
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This PR adds the Pascal VOC dataset in the same way TFDS has it added. I believe we can iterate on this dataset and in future versions include more data, such as segmentation masks, but for now I think it is a good idea to just add it the same way as TFDS to get a solid first version out there.
2023-09-24T09:37:05Z
https://github.com/huggingface/datasets/pull/4395
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4395/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4394/comments
https://api.github.com/repos/huggingface/datasets/issues/4394/timeline
null
null
null
I_kwDODunzps5KOJMZ
open
[]
null
4,394
{ "avatar_url": "https://avatars.githubusercontent.com/u/50416856?v=4", "events_url": "https://api.github.com/users/conan1024hao/events{/privacy}", "followers_url": "https://api.github.com/users/conan1024hao/followers", "following_url": "https://api.github.com/users/conan1024hao/following{/other_user}", "gists_url": "https://api.github.com/users/conan1024hao/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/conan1024hao", "id": 50416856, "login": "conan1024hao", "node_id": "MDQ6VXNlcjUwNDE2ODU2", "organizations_url": "https://api.github.com/users/conan1024hao/orgs", "received_events_url": "https://api.github.com/users/conan1024hao/received_events", "repos_url": "https://api.github.com/users/conan1024hao/repos", "site_admin": false, "starred_url": "https://api.github.com/users/conan1024hao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/conan1024hao/subscriptions", "type": "User", "url": "https://api.github.com/users/conan1024hao" }
trainer became extremely slow after reload dataset by `load_from_disk`
https://api.github.com/repos/huggingface/datasets/issues/4394/events
null
https://api.github.com/repos/huggingface/datasets/issues/4394/labels{/name}
2022-05-23T14:04:37Z
null
false
null
null
1,245,221,657
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4394
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Due to memory problem, I need to save my tokenized datasets locally by CPU and reload it by multi GPU for running training script. However, after I reload it by `load_from_disk` and start training, the speed is extremely slow. It says I need about 1500 hours with 8 A100 cards. Before this, I can run the whole script in one day with a single A100 card. Since I am try to pre-train a BERT, **my dataset is very large(29058165 rows)** ## Steps to reproduce the bug ```python tokenized_datasets.save_to_disk( "/pathto/dataset" ) tokenized_datasets = load_from_disk( "/pathto/dataset" ) trainer = Trainer( model=model, args=training_args, train_dataset=tokenized_datasets["train"] if training_args.do_train else None, eval_dataset=tokenized_datasets["validation"] if training_args.do_eval else None, tokenizer=tokenizer, data_collator=data_collator, ) train_result = trainer.train(resume_from_checkpoint=checkpoint) ``` ## Expected results Without the save and reload process, I only need about one day to run the whole script with one A100 card. ## Actual results ``` [INFO|trainer.py:1290] 2022-05-23 22:49:46,266 >> ***** Running training ***** [INFO|trainer.py:1291] 2022-05-23 22:49:46,266 >> Num examples = 29058165 [INFO|trainer.py:1292] 2022-05-23 22:49:46,266 >> Num Epochs = 5 [INFO|trainer.py:1293] 2022-05-23 22:49:46,266 >> Instantaneous batch size per device = 16 [INFO|trainer.py:1294] 2022-05-23 22:49:46,266 >> Total train batch size (w. parallel, distributed & accumulation) = 256 [INFO|trainer.py:1295] 2022-05-23 22:49:46,266 >> Gradient Accumulation steps = 2 [INFO|trainer.py:1296] 2022-05-23 22:49:46,266 >> Total optimization steps = 567540 0%| | 1/567540 [00:09<1544:49:04, 9.80s/it] 0%| | 2/567540 [00:17<1320:00:17, 8.37s/it] 0%| | 3/567540 [00:26<1393:10:17, 8.84s/it] 0%| | 4/567540 [00:34<1344:56:33, 8.53s/it] 0%| | 5/567540 [00:43<1359:36:12, 8.62s/it] ``` ## Environment info ``` torch 1.11.0+cu113 torchaudio 0.11.0+cu113 torchvision 0.12.0+cu113 transformers 4.18.0 datasets 2.2.2 ```
2023-11-23T07:40:30Z
https://github.com/huggingface/datasets/issues/4394
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4394/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4393/comments
https://api.github.com/repos/huggingface/datasets/issues/4393/timeline
2022-05-23T09:59:55Z
null
null
PR_kwDODunzps44RxWN
closed
[]
false
4,393
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update CI deprecated legacy image
https://api.github.com/repos/huggingface/datasets/issues/4393/events
null
https://api.github.com/repos/huggingface/datasets/issues/4393/labels{/name}
2022-05-23T09:35:42Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4393.diff", "html_url": "https://github.com/huggingface/datasets/pull/4393", "merged_at": "2022-05-23T09:59:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/4393.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4393" }
1,244,876,662
[]
https://api.github.com/repos/huggingface/datasets/issues/4393
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Now our CI still uses a deprecated legacy image: > You’re using a [deprecated Docker convenience image.](https://discuss.circleci.com/t/legacy-convenience-image-deprecation/41034) Upgrade to a next-gen Docker convenience image. This PR updates to next-generation convenience image. Related to: - #2955
2022-05-23T10:08:28Z
https://github.com/huggingface/datasets/pull/4393
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4393/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4392/comments
https://api.github.com/repos/huggingface/datasets/issues/4392/timeline
2022-05-23T15:08:32Z
null
null
PR_kwDODunzps44RtsX
closed
[]
false
4,392
{ "avatar_url": "https://avatars.githubusercontent.com/u/8264887?v=4", "events_url": "https://api.github.com/users/lvwerra/events{/privacy}", "followers_url": "https://api.github.com/users/lvwerra/followers", "following_url": "https://api.github.com/users/lvwerra/following{/other_user}", "gists_url": "https://api.github.com/users/lvwerra/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lvwerra", "id": 8264887, "login": "lvwerra", "node_id": "MDQ6VXNlcjgyNjQ4ODc=", "organizations_url": "https://api.github.com/users/lvwerra/orgs", "received_events_url": "https://api.github.com/users/lvwerra/received_events", "repos_url": "https://api.github.com/users/lvwerra/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lvwerra/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lvwerra/subscriptions", "type": "User", "url": "https://api.github.com/users/lvwerra" }
remove int documentation from logging docs
https://api.github.com/repos/huggingface/datasets/issues/4392/events
null
https://api.github.com/repos/huggingface/datasets/issues/4392/labels{/name}
2022-05-23T09:24:55Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4392.diff", "html_url": "https://github.com/huggingface/datasets/pull/4392", "merged_at": "2022-05-23T15:08:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/4392.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4392" }
1,244,859,971
[]
https://api.github.com/repos/huggingface/datasets/issues/4392
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Removes the `int` documentation from the [logging section](https://huggingface.co/docs/datasets/package_reference/logging_methods#levels) of the docs.
2022-05-23T15:16:55Z
https://github.com/huggingface/datasets/pull/4392
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4392/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4391/comments
https://api.github.com/repos/huggingface/datasets/issues/4391/timeline
2022-05-24T12:48:48Z
null
null
PR_kwDODunzps44RpGv
closed
[]
false
4,391
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
Refactor column mappings for question answering datasets
https://api.github.com/repos/huggingface/datasets/issues/4391/events
null
https://api.github.com/repos/huggingface/datasets/issues/4391/labels{/name}
2022-05-23T09:13:14Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4391.diff", "html_url": "https://github.com/huggingface/datasets/pull/4391", "merged_at": "2022-05-24T12:48:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/4391.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4391" }
1,244,839,185
[]
https://api.github.com/repos/huggingface/datasets/issues/4391
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This PR tweaks the keys in the metadata that are used to define the column mapping for question answering datasets. This is needed in order to faithfully reconstruct column names like `answers.text` and `answers.answer_start` from the keys in AutoTrain. As observed in https://github.com/huggingface/datasets/pull/4367 we cannot use periods `.` in the keys of the YAML tags, so a decision was made to use a flat mapping with underscores. For QA datasets, however, it's handy to be able to reconstruct the nesting -- hence this PR. cc @sashavor
2022-05-24T12:57:00Z
https://github.com/huggingface/datasets/pull/4391
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4391/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4390/comments
https://api.github.com/repos/huggingface/datasets/issues/4390/timeline
2022-06-01T09:19:25Z
null
null
PR_kwDODunzps44RoXs
closed
[]
false
4,390
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix metadata validation
https://api.github.com/repos/huggingface/datasets/issues/4390/events
null
https://api.github.com/repos/huggingface/datasets/issues/4390/labels{/name}
2022-05-23T09:11:20Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4390.diff", "html_url": "https://github.com/huggingface/datasets/pull/4390", "merged_at": "2022-06-01T09:19:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/4390.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4390" }
1,244,835,877
[]
https://api.github.com/repos/huggingface/datasets/issues/4390
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Since Python 3.8, the typing module: - raises an AttributeError when trying to access `__args__` on any type, e.g.: `List.__args__` - provides the `get_args` function instead: `get_args(List)` This PR implements a fix for Python >=3.8 whereas maintaining backward compatibility.
2022-06-01T09:27:52Z
https://github.com/huggingface/datasets/pull/4390
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4390/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4389/comments
https://api.github.com/repos/huggingface/datasets/issues/4389/timeline
2022-05-23T10:29:55Z
null
null
PR_kwDODunzps44RKMn
closed
[]
false
4,389
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix bug in gem dataset for wiki_auto_asset_turk config
https://api.github.com/repos/huggingface/datasets/issues/4389/events
null
https://api.github.com/repos/huggingface/datasets/issues/4389/labels{/name}
2022-05-23T07:19:49Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4389.diff", "html_url": "https://github.com/huggingface/datasets/pull/4389", "merged_at": "2022-05-23T10:29:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/4389.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4389" }
1,244,693,690
[]
https://api.github.com/repos/huggingface/datasets/issues/4389
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This PR fixes some URLs. Fix #4386.
2022-05-23T10:38:26Z
https://github.com/huggingface/datasets/pull/4389
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4389/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4388/comments
https://api.github.com/repos/huggingface/datasets/issues/4388/timeline
2022-05-25T05:16:15Z
null
null
PR_kwDODunzps44RAG1
closed
[]
false
4,388
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Set builder name from module instead of class
https://api.github.com/repos/huggingface/datasets/issues/4388/events
null
https://api.github.com/repos/huggingface/datasets/issues/4388/labels{/name}
2022-05-23T06:26:35Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4388.diff", "html_url": "https://github.com/huggingface/datasets/pull/4388", "merged_at": "2022-05-25T05:16:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/4388.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4388" }
1,244,645,158
[]
https://api.github.com/repos/huggingface/datasets/issues/4388
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Now the builder name attribute is set from from the builder class name. This PR sets the builder name attribute from the module name instead. Some motivating reasons: - The dataset ID is relevant and unique among all datasets and this is directly related to the repository name, i.e., the name of the directory containing the dataset - The name of the module (i.e. the file containing the loading loading script) is already relevant for loading: it must have the same name as its containing directory (related to the dataset ID), as we search for it using its directory name - On the other hand, the name of the builder class is not relevant for loading: in our code, we just search for a class which is subclass of `DatasetBuilder` (independently of its name). We do not put any constraint on the naming of the builder class and indeed it can have a name completely different from its module/direcotry/dataset_id IMO it makes more sense to align the caching directory name with the dataset_id/directory/module name instead of the builder class name. Fix #4381.
2022-05-25T05:24:43Z
https://github.com/huggingface/datasets/pull/4388
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4388/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4387/comments
https://api.github.com/repos/huggingface/datasets/issues/4387/timeline
2022-05-23T06:36:27Z
null
completed
I_kwDODunzps5KKDBp
closed
[]
null
4,387
{ "avatar_url": "https://avatars.githubusercontent.com/u/87345839?v=4", "events_url": "https://api.github.com/users/Aeckard45/events{/privacy}", "followers_url": "https://api.github.com/users/Aeckard45/followers", "following_url": "https://api.github.com/users/Aeckard45/following{/other_user}", "gists_url": "https://api.github.com/users/Aeckard45/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Aeckard45", "id": 87345839, "login": "Aeckard45", "node_id": "MDQ6VXNlcjg3MzQ1ODM5", "organizations_url": "https://api.github.com/users/Aeckard45/orgs", "received_events_url": "https://api.github.com/users/Aeckard45/received_events", "repos_url": "https://api.github.com/users/Aeckard45/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Aeckard45/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Aeckard45/subscriptions", "type": "User", "url": "https://api.github.com/users/Aeckard45" }
device/google/accessory/adk2012 - Git at Google
https://api.github.com/repos/huggingface/datasets/issues/4387/events
null
https://api.github.com/repos/huggingface/datasets/issues/4387/labels{/name}
2022-05-22T04:57:19Z
null
false
null
null
1,244,147,817
[]
https://api.github.com/repos/huggingface/datasets/issues/4387
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
"git clone https://android.googlesource.com/device/google/accessory/adk2012" https://android.googlesource.com/device/google/accessory/adk2012/#:~:text=git%20clone%20https%3A//android.googlesource.com/device/google/accessory/adk2012
2022-05-23T06:36:27Z
https://github.com/huggingface/datasets/issues/4387
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4387/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4386/comments
https://api.github.com/repos/huggingface/datasets/issues/4386/timeline
2022-05-23T10:29:55Z
null
completed
I_kwDODunzps5KJWhc
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
4,386
{ "avatar_url": "https://avatars.githubusercontent.com/u/37647985?v=4", "events_url": "https://api.github.com/users/StevenTang1998/events{/privacy}", "followers_url": "https://api.github.com/users/StevenTang1998/followers", "following_url": "https://api.github.com/users/StevenTang1998/following{/other_user}", "gists_url": "https://api.github.com/users/StevenTang1998/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/StevenTang1998", "id": 37647985, "login": "StevenTang1998", "node_id": "MDQ6VXNlcjM3NjQ3OTg1", "organizations_url": "https://api.github.com/users/StevenTang1998/orgs", "received_events_url": "https://api.github.com/users/StevenTang1998/received_events", "repos_url": "https://api.github.com/users/StevenTang1998/repos", "site_admin": false, "starred_url": "https://api.github.com/users/StevenTang1998/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/StevenTang1998/subscriptions", "type": "User", "url": "https://api.github.com/users/StevenTang1998" }
Bug for wiki_auto_asset_turk from GEM
https://api.github.com/repos/huggingface/datasets/issues/4386/events
null
https://api.github.com/repos/huggingface/datasets/issues/4386/labels{/name}
2022-05-21T12:31:30Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,243,965,532
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4386
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug The script of wiki_auto_asset_turk for GEM may be out of date. ## Steps to reproduce the bug ```python import datasets datasets.load_dataset('gem', 'wiki_auto_asset_turk') ``` ## Actual results ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/load.py", line 1731, in load_dataset builder_instance.download_and_prepare( File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/builder.py", line 640, in download_and_prepare self._download_and_prepare( File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/builder.py", line 1158, in _download_and_prepare super()._download_and_prepare(dl_manager, verify_infos, check_duplicate_keys=verify_infos) File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/builder.py", line 707, in _download_and_prepare split_generators = self._split_generators(dl_manager, **split_generators_kwargs) File "/home/tangtianyi/.cache/huggingface/modules/datasets_modules/datasets/gem/982a54473b12c6a6e40d4356e025fb7172a5bb2065e655e2c1af51f2b3cf4ca1/gem.py", line 538, in _split_generators dl_dir = dl_manager.download_and_extract(_URLs[self.config.name]) File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/utils/download_manager.py", line 416, in download_and_extract return self.extract(self.download(url_or_urls)) File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/utils/download_manager.py", line 294, in download downloaded_path_or_paths = map_nested( File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/utils/py_utils.py", line 351, in map_nested mapped = [ File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/utils/py_utils.py", line 352, in <listcomp> _single_map_nested((function, obj, types, None, True, None)) File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/utils/py_utils.py", line 288, in _single_map_nested return function(data_struct) File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/utils/download_manager.py", line 320, in _download return cached_path(url_or_filename, download_config=download_config) File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/utils/file_utils.py", line 234, in cached_path output_path = get_from_cache( File "/home/tangtianyi/miniconda3/lib/python3.8/site-packages/datasets/utils/file_utils.py", line 579, in get_from_cache raise FileNotFoundError(f"Couldn't find file at {url}") FileNotFoundError: Couldn't find file at https://github.com/facebookresearch/asset/raw/master/dataset/asset.test.orig ```
2022-05-24T05:55:52Z
https://github.com/huggingface/datasets/issues/4386
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4386/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4385/comments
https://api.github.com/repos/huggingface/datasets/issues/4385/timeline
2022-05-25T08:21:48Z
null
null
PR_kwDODunzps44OwXF
closed
[]
false
4,385
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Test dill
https://api.github.com/repos/huggingface/datasets/issues/4385/events
null
https://api.github.com/repos/huggingface/datasets/issues/4385/labels{/name}
2022-05-21T08:57:43Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4385.diff", "html_url": "https://github.com/huggingface/datasets/pull/4385", "merged_at": "2022-05-25T08:21:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/4385.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4385" }
1,243,921,287
[]
https://api.github.com/repos/huggingface/datasets/issues/4385
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Regression test for future releases of `dill`. Related to #4379.
2022-05-25T08:30:13Z
https://github.com/huggingface/datasets/pull/4385
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4385/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4384/comments
https://api.github.com/repos/huggingface/datasets/issues/4384/timeline
2022-05-25T10:43:43Z
null
null
PR_kwDODunzps44OwFr
closed
[]
false
4,384
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Refactor download
https://api.github.com/repos/huggingface/datasets/issues/4384/events
null
https://api.github.com/repos/huggingface/datasets/issues/4384/labels{/name}
2022-05-21T08:49:24Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4384.diff", "html_url": "https://github.com/huggingface/datasets/pull/4384", "merged_at": "2022-05-25T10:43:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/4384.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4384" }
1,243,919,748
[]
https://api.github.com/repos/huggingface/datasets/issues/4384
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This PR performs a refactoring of the download functionalities, by proposing a modular solution and moving them to their own package "download". Some motivating arguments: - understandability: from a logical partitioning of the library, it makes sense to have all download functionalities grouped together instead of scattered in a much larger directory containing many more different functionalities - abstraction: the level of abstraction of "download" (higher) is not the same as "utils" (lower); putting different levels of abstraction together, makes dependencies more intricate (potential circular dependencies) and the system more tightly coupled; when the levels of abstraction are clearly separated, the dependencies flow in a neat direction from higher to lower - architectural: "download" is a domain-specific functionality of our library/application (a dataset builder performs several actions: download, generate dataset and cache it); these functionalities are at the core of our library; on the other hand, "utils" are always a low-level set of functionalities, not directly related to our domain/business core logic (all libraries have "utils"), thus at the periphery of our lib architecture Also note that when a library is not architecturally designed following simple, neat, clean principles, this has a negative impact on extensibility, making more and more difficult to make enhancements. As a concrete example in this case, please see: https://app.circleci.com/pipelines/github/huggingface/datasets/12185/workflows/ff25a790-8e3f-45a1-aadd-9d79dfb73c4d/jobs/72860 - After an extension, a circular import is found - Diving into the cause of this circular import, see the dependency flow, which should be from higher to lower levels of abstraction: ``` ImportError while loading conftest '/home/circleci/datasets/tests/conftest.py'. tests/conftest.py:12: in <module> import datasets ../.pyenv/versions/3.6.15/lib/python3.6/site-packages/datasets/__init__.py:37: in <module> from .arrow_dataset import Dataset, concatenate_datasets ../.pyenv/versions/3.6.15/lib/python3.6/site-packages/datasets/arrow_dataset.py:59: in <module> from . import config ../.pyenv/versions/3.6.15/lib/python3.6/site-packages/datasets/config.py:8: in <module> from .utils.logging import get_logger ../.pyenv/versions/3.6.15/lib/python3.6/site-packages/datasets/utils/__init__.py:30: in <module> from .download_manager import DownloadConfig, DownloadManager, DownloadMode ../.pyenv/versions/3.6.15/lib/python3.6/site-packages/datasets/utils/download_manager.py:39: in <module> from .py_utils import NestedDataStructure, map_nested, size_str ../.pyenv/versions/3.6.15/lib/python3.6/site-packages/datasets/utils/py_utils.py:608: in <module> if config.DILL_VERSION < version.parse("0.3.5"): E AttributeError: module 'datasets.config' has no attribute 'DILL_VERSION' ``` Imports: - datasets - Dataset: lower level than datasets - config: lower level than Dataset - logger: lower level than config - DownloadManager: !!! HIGHER level of abstraction than logger!! Why when importing logger we require importing DownloadManager?!? - Logically, it does not make sense - This is due to an error in the design/architecture of our library: - To import the logger, we need to import it from `.utils.logging` - To import `.utils.logging` we need to import `.utils` - The import of `.utils` require the import of all its submodules defined in `utils.__init__.py`, among them: `.utils.download_manager`! When putting `logging` and `download_manager` both inside `utils`, in order to import `logging` we need to import `download_manager` first: this is a strong coupling between modules and moreover between modules at different levels of abstraction (to import a lower level module, we require to import a higher level module). Additionally, it is clear that is makes no sense that in order to import `logging` we require to import `download_manager` first.
2022-05-25T10:52:02Z
https://github.com/huggingface/datasets/pull/4384
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4384/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4383/comments
https://api.github.com/repos/huggingface/datasets/issues/4383/timeline
2022-05-21T19:20:13Z
null
completed
I_kwDODunzps5KI8BV
closed
[]
null
4,383
{ "avatar_url": "https://avatars.githubusercontent.com/u/99847861?v=4", "events_url": "https://api.github.com/users/AronCodes21/events{/privacy}", "followers_url": "https://api.github.com/users/AronCodes21/followers", "following_url": "https://api.github.com/users/AronCodes21/following{/other_user}", "gists_url": "https://api.github.com/users/AronCodes21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/AronCodes21", "id": 99847861, "login": "AronCodes21", "node_id": "U_kgDOBfOOtQ", "organizations_url": "https://api.github.com/users/AronCodes21/orgs", "received_events_url": "https://api.github.com/users/AronCodes21/received_events", "repos_url": "https://api.github.com/users/AronCodes21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/AronCodes21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AronCodes21/subscriptions", "type": "User", "url": "https://api.github.com/users/AronCodes21" }
L
https://api.github.com/repos/huggingface/datasets/issues/4383/events
null
https://api.github.com/repos/huggingface/datasets/issues/4383/labels{/name}
2022-05-21T03:47:58Z
null
false
null
null
1,243,856,981
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4383
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the L L ## Expected L A clear and concise lmll Specify the actual results or traceback. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - Platform: - Python version: - PyArrow version:
2022-05-21T19:20:13Z
https://github.com/huggingface/datasets/issues/4383
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4383/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4382/comments
https://api.github.com/repos/huggingface/datasets/issues/4382/timeline
2022-05-21T19:20:44Z
null
completed
I_kwDODunzps5KI30n
closed
[]
null
4,382
{ "avatar_url": "https://avatars.githubusercontent.com/u/87345839?v=4", "events_url": "https://api.github.com/users/Aeckard45/events{/privacy}", "followers_url": "https://api.github.com/users/Aeckard45/followers", "following_url": "https://api.github.com/users/Aeckard45/following{/other_user}", "gists_url": "https://api.github.com/users/Aeckard45/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Aeckard45", "id": 87345839, "login": "Aeckard45", "node_id": "MDQ6VXNlcjg3MzQ1ODM5", "organizations_url": "https://api.github.com/users/Aeckard45/orgs", "received_events_url": "https://api.github.com/users/Aeckard45/received_events", "repos_url": "https://api.github.com/users/Aeckard45/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Aeckard45/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Aeckard45/subscriptions", "type": "User", "url": "https://api.github.com/users/Aeckard45" }
First time trying
https://api.github.com/repos/huggingface/datasets/issues/4382/events
null
https://api.github.com/repos/huggingface/datasets/issues/4382/labels{/name}
2022-05-21T02:15:18Z
null
false
null
null
1,243,839,783
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
https://api.github.com/repos/huggingface/datasets/issues/4382
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Adding a Dataset - **Name:** *name of the dataset* - **Description:** *short description of the dataset (or link to social media or blog post)* - **Paper:** *link to the dataset paper if available* - **Data:** *link to the Github repository or current dataset location* - **Motivation:** *what are some good reasons to have this dataset* Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
2022-05-21T19:20:44Z
https://github.com/huggingface/datasets/issues/4382
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4382/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4381/comments
https://api.github.com/repos/huggingface/datasets/issues/4381/timeline
2022-05-25T05:16:15Z
null
completed
I_kwDODunzps5KHftP
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
4,381
{ "avatar_url": "https://avatars.githubusercontent.com/u/29777165?v=4", "events_url": "https://api.github.com/users/NouamaneTazi/events{/privacy}", "followers_url": "https://api.github.com/users/NouamaneTazi/followers", "following_url": "https://api.github.com/users/NouamaneTazi/following{/other_user}", "gists_url": "https://api.github.com/users/NouamaneTazi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NouamaneTazi", "id": 29777165, "login": "NouamaneTazi", "node_id": "MDQ6VXNlcjI5Nzc3MTY1", "organizations_url": "https://api.github.com/users/NouamaneTazi/orgs", "received_events_url": "https://api.github.com/users/NouamaneTazi/received_events", "repos_url": "https://api.github.com/users/NouamaneTazi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NouamaneTazi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NouamaneTazi/subscriptions", "type": "User", "url": "https://api.github.com/users/NouamaneTazi" }
Bug in caching 2 datasets both with the same builder class name
https://api.github.com/repos/huggingface/datasets/issues/4381/events
null
https://api.github.com/repos/huggingface/datasets/issues/4381/labels{/name}
2022-05-20T18:18:03Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,243,478,863
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4381
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Describe the bug The two datasets `mteb/mtop_intent` and `mteb/mtop_domain `use both the same cache folder `.cache/huggingface/datasets/mteb___mtop`. So if you first load `mteb/mtop_intent` then datasets will not load `mteb/mtop_domain`. If you delete this cache folder and flip the order how you load the two datasets , you will get the opposite datasets loaded (difference is here in terms of the label and label_text). ## Steps to reproduce the bug ```python import datasets dataset = datasets.load_dataset("mteb/mtop_intent", "en") print(dataset['train'][0]) dataset = datasets.load_dataset("mteb/mtop_domain", "en") print(dataset['train'][0]) ``` ## Expected results ``` Reusing dataset mtop (/home/nouamane/.cache/huggingface/datasets/mteb___mtop_intent/en/0.0.0/f930e32a294fed424f70263d8802390e350fff17862266e5fc156175c07d9c35) 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3/3 [00:00<00:00, 920.14it/s] {'id': 3232343436343136, 'text': 'Has Angelika Kratzer video messaged me?', 'label': 1, 'label_text': 'GET_MESSAGE'} Reusing dataset mtop (/home/nouamane/.cache/huggingface/datasets/mteb___mtop_domain/en/0.0.0/f930e32a294fed424f70263d8802390e350fff17862266e5fc156175c07d9c35) 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3/3 [00:00<00:00, 1307.59it/s] {'id': 3232343436343136, 'text': 'Has Angelika Kratzer video messaged me?', 'label': 0, 'label_text': 'messaging'} ``` ## Actual results ``` Reusing dataset mtop (/home/nouamane/.cache/huggingface/datasets/mteb___mtop/en/0.0.0/f930e32a294fed424f70263d8802390e350fff17862266e5fc156175c07d9c35) 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3/3 [00:00<00:00, 920.14it/s] {'id': 3232343436343136, 'text': 'Has Angelika Kratzer video messaged me?', 'label': 1, 'label_text': 'GET_MESSAGE'} Reusing dataset mtop (/home/nouamane/.cache/huggingface/datasets/mteb___mtop/en/0.0.0/f930e32a294fed424f70263d8802390e350fff17862266e5fc156175c07d9c35) 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3/3 [00:00<00:00, 1307.59it/s] {'id': 3232343436343136, 'text': 'Has Angelika Kratzer video messaged me?', 'label': 1, 'label_text': 'GET_MESSAGE'} ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.2.1 - Platform: macOS-12.1-arm64-arm-64bit - Python version: 3.9.12 - PyArrow version: 8.0.0 - Pandas version: 1.4.2
2022-06-02T08:18:37Z
https://github.com/huggingface/datasets/issues/4381
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4381/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4380/comments
https://api.github.com/repos/huggingface/datasets/issues/4380/timeline
2022-05-20T16:33:04Z
null
null
PR_kwDODunzps44MUz0
closed
[]
false
4,380
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Pin dill
https://api.github.com/repos/huggingface/datasets/issues/4380/events
null
https://api.github.com/repos/huggingface/datasets/issues/4380/labels{/name}
2022-05-20T13:54:19Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4380.diff", "html_url": "https://github.com/huggingface/datasets/pull/4380", "merged_at": "2022-05-20T16:33:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/4380.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4380" }
1,243,183,054
[]
https://api.github.com/repos/huggingface/datasets/issues/4380
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Hotfix #4379. CC: @sgugger
2022-06-13T10:03:52Z
https://github.com/huggingface/datasets/pull/4380
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4380/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4379/comments
https://api.github.com/repos/huggingface/datasets/issues/4379/timeline
2022-05-20T17:06:27Z
null
completed
I_kwDODunzps5KGVuu
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
4,379
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Latest dill release raises exception
https://api.github.com/repos/huggingface/datasets/issues/4379/events
null
https://api.github.com/repos/huggingface/datasets/issues/4379/labels{/name}
2022-05-20T13:48:36Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,243,175,854
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4379
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Describe the bug As reported by @sgugger, latest dill release is breaking things with Datasets. ``` ______________ ExamplesTests.test_run_speech_recognition_seq2seq _______________ self = <multiprocess.pool.ApplyResult object at 0x7fa5981a1cd0>, timeout = None def get(self, timeout=None): self.wait(timeout) if not self.ready(): raise TimeoutError if self._success: return self._value else: > raise self._value E TypeError: '>' not supported between instances of 'NoneType' and 'float' ```
2022-05-21T15:53:26Z
https://github.com/huggingface/datasets/issues/4379
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4379/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4378/comments
https://api.github.com/repos/huggingface/datasets/issues/4378/timeline
2022-05-24T13:10:27Z
null
null
PR_kwDODunzps44Lf2R
closed
[]
false
4,378
{ "avatar_url": "https://avatars.githubusercontent.com/u/121934?v=4", "events_url": "https://api.github.com/users/leondz/events{/privacy}", "followers_url": "https://api.github.com/users/leondz/followers", "following_url": "https://api.github.com/users/leondz/following{/other_user}", "gists_url": "https://api.github.com/users/leondz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/leondz", "id": 121934, "login": "leondz", "node_id": "MDQ6VXNlcjEyMTkzNA==", "organizations_url": "https://api.github.com/users/leondz/orgs", "received_events_url": "https://api.github.com/users/leondz/received_events", "repos_url": "https://api.github.com/users/leondz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/leondz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/leondz/subscriptions", "type": "User", "url": "https://api.github.com/users/leondz" }
Tidy up license metadata for google_wellformed_query, newspop, sick
https://api.github.com/repos/huggingface/datasets/issues/4378/events
null
https://api.github.com/repos/huggingface/datasets/issues/4378/labels{/name}
2022-05-20T10:16:12Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4378.diff", "html_url": "https://github.com/huggingface/datasets/pull/4378", "merged_at": "2022-05-24T13:10:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/4378.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4378" }
1,242,935,373
[]
https://api.github.com/repos/huggingface/datasets/issues/4378
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Amend three licenses on datasets to fit naming convention (lower case, cc licenses include sub-version number). I think that's it - everything else on datasets looks great & super-searchable now!
2022-05-24T13:50:23Z
https://github.com/huggingface/datasets/pull/4378
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4378/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4377/comments
https://api.github.com/repos/huggingface/datasets/issues/4377/timeline
2022-05-20T09:26:32Z
null
null
PR_kwDODunzps44K4OY
closed
[]
false
4,377
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix checksum and bug in irc_disentangle dataset
https://api.github.com/repos/huggingface/datasets/issues/4377/events
null
https://api.github.com/repos/huggingface/datasets/issues/4377/labels{/name}
2022-05-20T07:29:28Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4377.diff", "html_url": "https://github.com/huggingface/datasets/pull/4377", "merged_at": "2022-05-20T09:26:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/4377.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4377" }
1,242,746,186
[]
https://api.github.com/repos/huggingface/datasets/issues/4377
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
There was a bug in filepath segment: - wrong: `jkkummerfeld-irc-disentanglement-fd379e9` - right: `jkkummerfeld-irc-disentanglement-35f0a40` Also there was a bug in the checksum of the downloaded file. This PR fixes these issues. Fix partially #4376.
2022-05-20T09:34:36Z
https://github.com/huggingface/datasets/pull/4377
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4377/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4376/comments
https://api.github.com/repos/huggingface/datasets/issues/4376/timeline
2022-06-02T08:20:00Z
null
completed
I_kwDODunzps5KCr6g
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
4,376
{ "avatar_url": "https://avatars.githubusercontent.com/u/25671683?v=4", "events_url": "https://api.github.com/users/labouz/events{/privacy}", "followers_url": "https://api.github.com/users/labouz/followers", "following_url": "https://api.github.com/users/labouz/following{/other_user}", "gists_url": "https://api.github.com/users/labouz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/labouz", "id": 25671683, "login": "labouz", "node_id": "MDQ6VXNlcjI1NjcxNjgz", "organizations_url": "https://api.github.com/users/labouz/orgs", "received_events_url": "https://api.github.com/users/labouz/received_events", "repos_url": "https://api.github.com/users/labouz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/labouz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/labouz/subscriptions", "type": "User", "url": "https://api.github.com/users/labouz" }
irc_disentagle viewer error
https://api.github.com/repos/huggingface/datasets/issues/4376/events
null
https://api.github.com/repos/huggingface/datasets/issues/4376/labels{/name}
2022-05-19T19:15:16Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
1,242,218,144
[]
https://api.github.com/repos/huggingface/datasets/issues/4376
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
the dataviewer shows this message for "ubuntu" - "train", "test", and "validation" splits: ``` Server error Status code: 400 Exception: ValueError Message: Cannot seek streaming HTTP file ``` it appears to give the same message for the "channel_two" data as well. I get a Checksums error when using `load_data()` with this dataset. Even with the `download_mode` and `ignore_verifications` options set. i referenced the issue here: https://github.com/huggingface/datasets/issues/3807
2023-01-12T16:56:13Z
https://github.com/huggingface/datasets/issues/4376
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4376/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4375/comments
https://api.github.com/repos/huggingface/datasets/issues/4375/timeline
2022-06-10T20:47:27Z
null
null
PR_kwDODunzps44IMCS
closed
[]
false
4,375
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Support DataLoader with num_workers > 0 in streaming mode
https://api.github.com/repos/huggingface/datasets/issues/4375/events
null
https://api.github.com/repos/huggingface/datasets/issues/4375/labels{/name}
2022-05-19T15:00:31Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4375.diff", "html_url": "https://github.com/huggingface/datasets/pull/4375", "merged_at": "2022-06-10T20:47:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/4375.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4375" }
1,241,921,147
[]
https://api.github.com/repos/huggingface/datasets/issues/4375
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
### Issue It's currently not possible to properly stream a dataset using multiple `torch.utils.data.DataLoader` workers: - the `TorchIterableDataset` can't be pickled and passed to the subprocesses: https://github.com/huggingface/datasets/issues/3950 - streaming extension is failing: https://github.com/huggingface/datasets/issues/3951 - `fsspec` doesn't work out of the box in subprocesses ### Solution in this PR I fixed these to enable passing an `IterableDataset` to a `torch.utils.data.DataLoader` with `num_workers > 0`. I also had to shard the `IterableDataset` to give each worker a shard, otherwise data would be duplicated. This is implemented in `TorchIterableDataset.__iter__` and uses the new `IterableDataset._iter_shard(shard_idx)` method I also had to do a few changes the patching that enable streaming in dataset scripts: - the patches are now always applied - not just for streaming mode. They're applied when a builder is instantiated - I improved it to also check for renamed modules or attributes (ex: pandas vs pd) - I grouped all the patches of pathlib.Path into a class `xPath`, so that `Path` outside of dataset scripts stay unchanged - otherwise I didn't change the content of the extended Path methods for streaming - I fixed a bug with the `pd.read_csv` patch, opening the file in "rb" mode was missing and causing some datasets to not work in streaming mode, and compression inference was missing ### A few details regarding `fsspec` in multiprocessing From https://github.com/fsspec/filesystem_spec/pull/963#issuecomment-1131709948 : > Non-async instances might be safe in the forked child, if they hold no open files/sockets etc.; I'm not sure any implementations pass this test! > If any async instance has been created, the newly forked processes must: > 1. discard references to locks, threads and event loops and make new ones > 2. not use any async fsspec instances from the parent process > 3. clear all class instance caches Therefore in a DataLoader's worker, I clear the reference to the loop and thread (1). We should be fine for 2 and 3 already since we don't use fsspec class instances from the parent process. Fix https://github.com/huggingface/datasets/issues/3950 Fix https://github.com/huggingface/datasets/issues/3951 TODO: - [x] fix tests
2022-07-04T16:05:14Z
https://github.com/huggingface/datasets/pull/4375
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4375/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4374/comments
https://api.github.com/repos/huggingface/datasets/issues/4374/timeline
2023-07-25T15:07:16Z
null
completed
I_kwDODunzps5KBUm3
closed
[]
null
4,374
{ "avatar_url": "https://avatars.githubusercontent.com/u/32235549?v=4", "events_url": "https://api.github.com/users/StephennFernandes/events{/privacy}", "followers_url": "https://api.github.com/users/StephennFernandes/followers", "following_url": "https://api.github.com/users/StephennFernandes/following{/other_user}", "gists_url": "https://api.github.com/users/StephennFernandes/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/StephennFernandes", "id": 32235549, "login": "StephennFernandes", "node_id": "MDQ6VXNlcjMyMjM1NTQ5", "organizations_url": "https://api.github.com/users/StephennFernandes/orgs", "received_events_url": "https://api.github.com/users/StephennFernandes/received_events", "repos_url": "https://api.github.com/users/StephennFernandes/repos", "site_admin": false, "starred_url": "https://api.github.com/users/StephennFernandes/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/StephennFernandes/subscriptions", "type": "User", "url": "https://api.github.com/users/StephennFernandes" }
extremely slow processing when using a custom dataset
https://api.github.com/repos/huggingface/datasets/issues/4374/events
null
https://api.github.com/repos/huggingface/datasets/issues/4374/labels{/name}
2022-05-19T14:18:05Z
null
false
null
null
1,241,860,535
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "d876e3", "default": true, "description": "Further information is requested", "id": 1935892912, "name": "question", "node_id": "MDU6TGFiZWwxOTM1ODkyOTEy", "url": "https://api.github.com/repos/huggingface/datasets/labels/question" } ]
https://api.github.com/repos/huggingface/datasets/issues/4374
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## processing a custom dataset loaded as .txt file is extremely slow, compared to a dataset of similar volume from the hub I have a large .txt file of 22 GB which i load into HF dataset `lang_dataset = datasets.load_dataset("text", data_files="hi.txt")` further i use a pre-processing function to clean the dataset `lang_dataset["train"] = lang_dataset["train"].map( remove_non_indic_sentences, num_proc=12, batched=True, remove_columns=lang_dataset['train'].column_names), batch_size=64)` the following processing takes astronomical time to process, while hoging all the ram. similar dataset of same size that's available in the huggingface hub works completely fine. which runs the same processing function and has the same amount of data. `lang_dataset = datasets.load_dataset("oscar-corpus/OSCAR-2109", "hi", use_auth_token=True)` the hours predicted to preprocess are as follows: huggingface hub dataset: 6.5 hrs custom loaded dataset: 7000 hrs note: both the datasets are almost actually same, just provided by different sources with has +/- some samples, only one is hosted on the HF hub and the other is downloaded in a text format. ## Steps to reproduce the bug ``` import datasets import psutil import sys import glob from fastcore.utils import listify import re import gc def remove_non_indic_sentences(example): tmp_ls = [] eng_regex = r'[. a-zA-Z0-9ÖÄÅâÀΓ₯ _.,!"\'\/$]*' for e in listify(example['text']): matches = re.findall(eng_regex, e) for match in (str(match).strip() for match in matches if match not in [""," ", " ", ",", " ,", ", ", " , "]): if len(list(match.split(" "))) > 2: e = re.sub(match," ",e,count=1) tmp_ls.append(e) gc.collect() example['clean_text'] = tmp_ls return example lang_dataset = datasets.load_dataset("text", data_files="hi.txt") lang_dataset["train"] = lang_dataset["train"].map( remove_non_indic_sentences, num_proc=12, batched=True, remove_columns=lang_dataset['train'].column_names), batch_size=64) ## same thing work much faster when loading similar dataset from hub lang_dataset = datasets.load_dataset("oscar-corpus/OSCAR-2109", "hi", split="train", use_auth_token=True) lang_dataset["train"] = lang_dataset["train"].map( remove_non_indic_sentences, num_proc=12, batched=True, remove_columns=lang_dataset['train'].column_names), batch_size=64) ``` ## Actual results similar dataset of same size that's available in the huggingface hub works completely fine. which runs the same processing function and has the same amount of data. `lang_dataset = datasets.load_dataset("oscar-corpus/OSCAR-2109", "hi", use_auth_token=True) **the hours predicted to preprocess are as follows:** huggingface hub dataset: 6.5 hrs custom loaded dataset: 7000 hrs **i even tried the following:** - sharding the large 22gb text files into smaller files and loading - saving the file to disk and then loading - using lesser num_proc - using smaller batch size - processing without batches ie : without `batched=True` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.2.2.dev0 - Platform: Ubuntu 20.04 LTS - Python version: 3.9.7 - PyArrow version:8.0.0
2023-07-25T15:07:17Z
https://github.com/huggingface/datasets/issues/4374
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4374/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4373/comments
https://api.github.com/repos/huggingface/datasets/issues/4373/timeline
2022-05-20T15:16:05Z
null
null
PR_kwDODunzps44HsaY
closed
[]
false
4,373
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Remove links in docs to old dataset viewer
https://api.github.com/repos/huggingface/datasets/issues/4373/events
null
https://api.github.com/repos/huggingface/datasets/issues/4373/labels{/name}
2022-05-19T13:24:39Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4373.diff", "html_url": "https://github.com/huggingface/datasets/pull/4373", "merged_at": "2022-05-20T15:16:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/4373.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4373" }
1,241,769,310
[]
https://api.github.com/repos/huggingface/datasets/issues/4373
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Remove the links in the docs to the no longer maintained dataset viewer.
2022-05-20T15:24:28Z
https://github.com/huggingface/datasets/pull/4373
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4373/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4372/comments
https://api.github.com/repos/huggingface/datasets/issues/4372/timeline
2022-05-20T15:15:30Z
null
null
PR_kwDODunzps44HeYC
closed
[]
false
4,372
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Check if dataset features match before push in `DatasetDict.push_to_hub`
https://api.github.com/repos/huggingface/datasets/issues/4372/events
null
https://api.github.com/repos/huggingface/datasets/issues/4372/labels{/name}
2022-05-19T12:32:30Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4372.diff", "html_url": "https://github.com/huggingface/datasets/pull/4372", "merged_at": "2022-05-20T15:15:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/4372.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4372" }
1,241,703,826
[]
https://api.github.com/repos/huggingface/datasets/issues/4372
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Fix #4211
2022-05-20T15:23:36Z
https://github.com/huggingface/datasets/pull/4372
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4372/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4371/comments
https://api.github.com/repos/huggingface/datasets/issues/4371/timeline
2022-05-20T09:43:10Z
null
null
PR_kwDODunzps44GzSZ
closed
[]
false
4,371
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Add missing language tags for udhr dataset
https://api.github.com/repos/huggingface/datasets/issues/4371/events
null
https://api.github.com/repos/huggingface/datasets/issues/4371/labels{/name}
2022-05-19T09:34:10Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4371.diff", "html_url": "https://github.com/huggingface/datasets/pull/4371", "merged_at": "2022-05-20T09:43:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/4371.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4371" }
1,241,500,906
[]
https://api.github.com/repos/huggingface/datasets/issues/4371
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Related to #4362.
2022-06-08T12:03:24Z
https://github.com/huggingface/datasets/pull/4371
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4371/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4369/comments
https://api.github.com/repos/huggingface/datasets/issues/4369/timeline
2022-05-19T08:10:51Z
null
null
PR_kwDODunzps44CpCe
closed
[]
false
4,369
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Add redirect to dataset script in the repo structure page
https://api.github.com/repos/huggingface/datasets/issues/4369/events
null
https://api.github.com/repos/huggingface/datasets/issues/4369/labels{/name}
2022-05-18T17:05:33Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4369.diff", "html_url": "https://github.com/huggingface/datasets/pull/4369", "merged_at": "2022-05-19T08:10:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/4369.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4369" }
1,240,245,642
[]
https://api.github.com/repos/huggingface/datasets/issues/4369
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Following https://github.com/huggingface/hub-docs/pull/146 I added a redirection to the dataset scripts documentation in the repository structure page.
2022-05-19T08:19:01Z
https://github.com/huggingface/datasets/pull/4369
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4369/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4368/comments
https://api.github.com/repos/huggingface/datasets/issues/4368/timeline
2022-07-26T20:18:42Z
null
null
PR_kwDODunzps44CDFk
closed
[]
false
4,368
{ "avatar_url": "https://avatars.githubusercontent.com/u/4257308?v=4", "events_url": "https://api.github.com/users/seirasto/events{/privacy}", "followers_url": "https://api.github.com/users/seirasto/followers", "following_url": "https://api.github.com/users/seirasto/following{/other_user}", "gists_url": "https://api.github.com/users/seirasto/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/seirasto", "id": 4257308, "login": "seirasto", "node_id": "MDQ6VXNlcjQyNTczMDg=", "organizations_url": "https://api.github.com/users/seirasto/orgs", "received_events_url": "https://api.github.com/users/seirasto/received_events", "repos_url": "https://api.github.com/users/seirasto/repos", "site_admin": false, "starred_url": "https://api.github.com/users/seirasto/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/seirasto/subscriptions", "type": "User", "url": "https://api.github.com/users/seirasto" }
Add long answer candidates to natural questions dataset
https://api.github.com/repos/huggingface/datasets/issues/4368/events
null
https://api.github.com/repos/huggingface/datasets/issues/4368/labels{/name}
2022-05-18T14:35:42Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4368.diff", "html_url": "https://github.com/huggingface/datasets/pull/4368", "merged_at": "2022-07-26T20:18:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/4368.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4368" }
1,240,064,860
[]
https://api.github.com/repos/huggingface/datasets/issues/4368
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This is a modification of the Natural Questions dataset to include missing information specifically related to long answer candidates. (See here: https://github.com/google-research-datasets/natural-questions#long-answer-candidates). This information is important to ensure consistent comparison with prior work. It does not disturb the rest of the format . @lhoestq @albertvillanova
2022-07-26T20:30:41Z
https://github.com/huggingface/datasets/pull/4368
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4368/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4367/comments
https://api.github.com/repos/huggingface/datasets/issues/4367/timeline
2022-05-20T09:27:19Z
null
null
PR_kwDODunzps44B340
closed
[]
false
4,367
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Remove config names as yaml keys
https://api.github.com/repos/huggingface/datasets/issues/4367/events
null
https://api.github.com/repos/huggingface/datasets/issues/4367/labels{/name}
2022-05-18T13:59:24Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4367.diff", "html_url": "https://github.com/huggingface/datasets/pull/4367", "merged_at": "2022-05-20T09:27:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/4367.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4367" }
1,240,011,602
[]
https://api.github.com/repos/huggingface/datasets/issues/4367
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Many datasets have dots in their config names. However it causes issues with the YAML tags of the dataset cards since we can't have dots in YAML keys. I fix this, I removed the tags separations per config name completely, and have a single flat YAML for all configurations. Dataset search doesn't use this info anyway. I removed all the config names used as YAML keys, and I moved them in under a new `config:` key. This is related to https://github.com/huggingface/datasets/pull/2362 (internal https://github.com/huggingface/moon-landing/issues/946). Also removing the dots in the YAML keys would allow us to do as in https://github.com/huggingface/datasets/pull/4302 which removes a hack that replaces all the dots by underscores in the YAML tags. I also added a test in the CI that checks that all the YAML tags to make sure that: - they can be parsed using a YAML parser - they contain only valid YAML tags like languages or task_ids
2022-05-20T09:35:26Z
https://github.com/huggingface/datasets/pull/4367
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4367/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4366/comments
https://api.github.com/repos/huggingface/datasets/issues/4366/timeline
2022-05-18T16:36:21Z
null
completed
I_kwDODunzps5J4cpV
closed
[]
null
4,366
{ "avatar_url": "https://avatars.githubusercontent.com/u/99231535?v=4", "events_url": "https://api.github.com/users/jffgitt/events{/privacy}", "followers_url": "https://api.github.com/users/jffgitt/followers", "following_url": "https://api.github.com/users/jffgitt/following{/other_user}", "gists_url": "https://api.github.com/users/jffgitt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jffgitt", "id": 99231535, "login": "jffgitt", "node_id": "U_kgDOBeonLw", "organizations_url": "https://api.github.com/users/jffgitt/orgs", "received_events_url": "https://api.github.com/users/jffgitt/received_events", "repos_url": "https://api.github.com/users/jffgitt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jffgitt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jffgitt/subscriptions", "type": "User", "url": "https://api.github.com/users/jffgitt" }
TypeError: __init__() missing 1 required positional argument: 'scheme'
https://api.github.com/repos/huggingface/datasets/issues/4366/events
null
https://api.github.com/repos/huggingface/datasets/issues/4366/labels{/name}
2022-05-18T07:17:29Z
null
false
null
null
1,239,534,165
[ { "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists", "id": 1935892865, "name": "duplicate", "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate" } ]
https://api.github.com/repos/huggingface/datasets/issues/4366
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
"name" : "node-1", "cluster_name" : "elasticsearch", "cluster_uuid" : "", "version" : { "number" : "7.5.0", "build_flavor" : "default", "build_type" : "tar", "build_hash" : "", "build_date" : "2019-11-26T01:06:52.518245Z", "build_snapshot" : false, "lucene_version" : "8.3.0", "minimum_wire_compatibility_version" : "6.8.0", "minimum_index_compatibility_version" : "6.0.0-beta1" when I run the order: nohup python3 custom_service.pyc > service.log 2>&1& the log: nohup: εΏ½η•₯θΎ“ε…₯ Traceback (most recent call last): File "/home/xfz/p3_custom_test/custom_service.py", line 55, in <module> File "/home/xfz/p3_custom_test/custom_service.py", line 48, in doInitialize File "custom_impl.py", line 286, in custom_setup File "custom_impl.py", line 127, in create_es_index File "/usr/local/lib/python3.7/site-packages/elasticsearch/_sync/client/__init__.py", line 345, in __init__ ssl_show_warn=ssl_show_warn, File "/usr/local/lib/python3.7/site-packages/elasticsearch/_sync/client/utils.py", line 105, in client_node_configs node_configs = hosts_to_node_configs(hosts) File "/usr/local/lib/python3.7/site-packages/elasticsearch/_sync/client/utils.py", line 154, in hosts_to_node_configs node_configs.append(host_mapping_to_node_config(host)) File "/usr/local/lib/python3.7/site-packages/elasticsearch/_sync/client/utils.py", line 221, in host_mapping_to_node_config return NodeConfig(**options) # type: ignore TypeError: __init__() missing 1 required positional argument: 'scheme' [1]+ ι€€ε‡Ί 1 nohup python3 custom_service.pyc > service.log 2>&1 custom_service_pyc can't running
2022-05-18T16:36:22Z
https://github.com/huggingface/datasets/issues/4366
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4366/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4365/comments
https://api.github.com/repos/huggingface/datasets/issues/4365/timeline
2022-05-18T13:59:41Z
null
null
PR_kwDODunzps43-4fC
closed
[]
true
4,365
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Remove dots in config names
https://api.github.com/repos/huggingface/datasets/issues/4365/events
null
https://api.github.com/repos/huggingface/datasets/issues/4365/labels{/name}
2022-05-17T20:12:57Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4365.diff", "html_url": "https://github.com/huggingface/datasets/pull/4365", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4365.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4365" }
1,239,109,943
[]
https://api.github.com/repos/huggingface/datasets/issues/4365
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
20+ datasets have dots in their config names. However it causes issues with the YAML tags of the dataset cards since we can't have dots in YAML keys. This is related to https://github.com/huggingface/datasets/pull/2362 (internal https://github.com/huggingface/moon-landing/issues/946). Also removing the dots in the config names would allow us to merge https://github.com/huggingface/datasets/pull/4302 which removes a hack that replaces all the dots by underscores in the YAML tags. I also added a test in the CI that checks that all the YAML tags to make sure that: - they can be parsed using a YAML parser - they contain only valid YAML tags like `languages` or `task_ids` - they contain valid config names (no invalid characters `<>:/\|?*.`)
2023-09-24T10:02:53Z
https://github.com/huggingface/datasets/pull/4365
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4365/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4364/comments
https://api.github.com/repos/huggingface/datasets/issues/4364/timeline
2022-05-31T12:16:32Z
null
null
PR_kwDODunzps43-bmq
closed
[]
false
4,364
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Support complex feature types as `features` in packaged loaders
https://api.github.com/repos/huggingface/datasets/issues/4364/events
null
https://api.github.com/repos/huggingface/datasets/issues/4364/labels{/name}
2022-05-17T17:53:23Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4364.diff", "html_url": "https://github.com/huggingface/datasets/pull/4364", "merged_at": "2022-05-31T12:16:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/4364.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4364" }
1,238,976,106
[]
https://api.github.com/repos/huggingface/datasets/issues/4364
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This PR adds `table_cast` to the packaged loaders to fix casting to the `Image`/`Audio`, `ArrayND` and `ClassLabel` types. If these types are not present in the `builder.config.features` dictionary, the built-in `pa.Table.cast` is used for better performance. Additionally, this PR adds `cast_storage` to `ClassLabel` to support the string to int conversion in `table_cast` and ensure that integer labels are in a valid range. Fix https://github.com/huggingface/datasets/issues/4210 This PR is also a solution for these (popular) discussions: https://discuss.huggingface.co/t/converting-string-label-to-int/2816 and https://discuss.huggingface.co/t/class-labels-for-custom-datasets/15130/2 TODO: * [x] tests
2022-05-31T12:26:23Z
https://github.com/huggingface/datasets/pull/4364
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 1, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4364/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4363/comments
https://api.github.com/repos/huggingface/datasets/issues/4363/timeline
2022-06-08T09:26:56Z
null
completed
I_kwDODunzps5J2BP0
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" } ]
null
4,363
{ "avatar_url": "https://avatars.githubusercontent.com/u/7584674?v=4", "events_url": "https://api.github.com/users/roholazandie/events{/privacy}", "followers_url": "https://api.github.com/users/roholazandie/followers", "following_url": "https://api.github.com/users/roholazandie/following{/other_user}", "gists_url": "https://api.github.com/users/roholazandie/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/roholazandie", "id": 7584674, "login": "roholazandie", "node_id": "MDQ6VXNlcjc1ODQ2NzQ=", "organizations_url": "https://api.github.com/users/roholazandie/orgs", "received_events_url": "https://api.github.com/users/roholazandie/received_events", "repos_url": "https://api.github.com/users/roholazandie/repos", "site_admin": false, "starred_url": "https://api.github.com/users/roholazandie/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/roholazandie/subscriptions", "type": "User", "url": "https://api.github.com/users/roholazandie" }
The dataset preview is not available for this split.
https://api.github.com/repos/huggingface/datasets/issues/4363/events
null
https://api.github.com/repos/huggingface/datasets/issues/4363/labels{/name}
2022-05-17T16:34:43Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
null
1,238,897,652
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
https://api.github.com/repos/huggingface/datasets/issues/4363
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
I have uploaded the corpus developed by our lab in the speech domain to huggingface [datasets](https://huggingface.co/datasets/Roh/ryanspeech). You can read about the companion paper accepted in interspeech 2021 [here](https://arxiv.org/abs/2106.08468). The dataset works fine but I can't make the dataset preview work. It gives me the following error that I don't understand. Can you help me to begin debugging it? ``` Status code: 400 Exception: AttributeError Message: 'NoneType' object has no attribute 'split' ```
2022-06-08T12:32:10Z
https://github.com/huggingface/datasets/issues/4363
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4363/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/4362/comments
https://api.github.com/repos/huggingface/datasets/issues/4362/timeline
2022-06-08T19:11:21Z
null
null
PR_kwDODunzps439bkf
closed
[]
false
4,362
{ "avatar_url": "https://avatars.githubusercontent.com/u/121934?v=4", "events_url": "https://api.github.com/users/leondz/events{/privacy}", "followers_url": "https://api.github.com/users/leondz/followers", "following_url": "https://api.github.com/users/leondz/following{/other_user}", "gists_url": "https://api.github.com/users/leondz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/leondz", "id": 121934, "login": "leondz", "node_id": "MDQ6VXNlcjEyMTkzNA==", "organizations_url": "https://api.github.com/users/leondz/orgs", "received_events_url": "https://api.github.com/users/leondz/received_events", "repos_url": "https://api.github.com/users/leondz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/leondz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/leondz/subscriptions", "type": "User", "url": "https://api.github.com/users/leondz" }
Update dataset_infos for UDHN/udhr dataset
https://api.github.com/repos/huggingface/datasets/issues/4362/events
null
https://api.github.com/repos/huggingface/datasets/issues/4362/labels{/name}
2022-05-17T13:52:59Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/4362.diff", "html_url": "https://github.com/huggingface/datasets/pull/4362", "merged_at": "2022-06-08T19:11:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/4362.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4362" }
1,238,680,112
[]
https://api.github.com/repos/huggingface/datasets/issues/4362
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Checksum update to `udhr` for issue #4361
2022-06-08T19:20:11Z
https://github.com/huggingface/datasets/pull/4362
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4362/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/4361/comments
https://api.github.com/repos/huggingface/datasets/issues/4361/timeline
2022-06-08T19:11:21Z
null
completed
I_kwDODunzps5J1KI7
closed
[]
null
4,361
{ "avatar_url": "https://avatars.githubusercontent.com/u/121934?v=4", "events_url": "https://api.github.com/users/leondz/events{/privacy}", "followers_url": "https://api.github.com/users/leondz/followers", "following_url": "https://api.github.com/users/leondz/following{/other_user}", "gists_url": "https://api.github.com/users/leondz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/leondz", "id": 121934, "login": "leondz", "node_id": "MDQ6VXNlcjEyMTkzNA==", "organizations_url": "https://api.github.com/users/leondz/orgs", "received_events_url": "https://api.github.com/users/leondz/received_events", "repos_url": "https://api.github.com/users/leondz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/leondz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/leondz/subscriptions", "type": "User", "url": "https://api.github.com/users/leondz" }
`udhr` doesn't load, dataset checksum mismatch
https://api.github.com/repos/huggingface/datasets/issues/4361/events
null
https://api.github.com/repos/huggingface/datasets/issues/4361/labels{/name}
2022-05-17T13:47:09Z
null
false
null
null
1,238,671,931
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/4361
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug Loading `udhr` fails due to a checksum mismatch for some source files. Looks like both of the source files on unicode.org have changed: size + checksum in datasets repo: ``` (hfdev) leon@blade:~/datasets/datasets/udhr$ jq .default.download_checksums < dataset_infos.json { "https://unicode.org/udhr/assemblies/udhr_xml.zip": { "num_bytes": 2273633, "checksum": "0565fa62c2ff155b84123198bcc967edd8c5eb9679eadc01e6fb44a5cf730fee" }, "https://unicode.org/udhr/assemblies/udhr_txt.zip": { "num_bytes": 2107471, "checksum": "087b474a070dd4096ae3028f9ee0b30dcdcb030cc85a1ca02e143be46327e5e5" } } ``` size + checksum regenerated from current source files: ``` (hfdev) leon@blade:~/datasets/datasets/udhr$ rm dataset_infos.json (hfdev) leon@blade:~/datasets/datasets/udhr$ datasets-cli test --save_infos udhr.py Using custom data configuration default Testing builder 'default' (1/1) Downloading and preparing dataset udhn/default (download: 4.18 MiB, generated: 6.15 MiB, post-processed: Unknown size, total: 10.33 MiB) to /home/leon/.cache/huggingface/datasets/udhn/default/0.0.0/ad74b91fa2b3c386e5751b0c52bdfda76d334f76731142fd432d4acc2e2fde66... Dataset udhn downloaded and prepared to /home/leon/.cache/huggingface/datasets/udhn/default/0.0.0/ad74b91fa2b3c386e5751b0c52bdfda76d334f76731142fd432d4acc2e2fde66. Subsequent calls will reuse this data. 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:00<00:00, 686.69it/s] Dataset Infos file saved at dataset_infos.json Test successful. (hfdev) leon@blade:~/datasets/datasets/udhr$ jq .default.download_checksums < dataset_infos.json { "https://unicode.org/udhr/assemblies/udhr_xml.zip": { "num_bytes": 2389690, "checksum": "a3350912790196c6e1b26bfd1c8a50e8575f5cf185922ecd9bd15713d7d21438" }, "https://unicode.org/udhr/assemblies/udhr_txt.zip": { "num_bytes": 2215441, "checksum": "cb87ecb25b56f34e4fd6f22b323000524fd9c06ae2a29f122b048789cf17e9fe" } } (hfdev) leon@blade:~/datasets/datasets/udhr$ ``` --- is unicode.org a sustainable hosting solution for this dataset? ## Steps to reproduce the bug ```python from datasets import load_dataset udhr = load_dataset("udhr") ``` ## Expected results That a Dataset object containing the UDHR data will be returned. ## Actual results ``` >>> d = load_dataset('udhr') Using custom data configuration default Downloading and preparing dataset udhn/default (download: 4.18 MiB, generated: 6.15 MiB, post-processed: Unknown size, total: 10.33 MiB) to /home/leon/.cache/huggingface/datasets/udhn/default/0.0.0/ad74b91fa2b3c386e5751b0c52bdfda76d334f76731142fd432d4acc2e2fde66... Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/leon/.local/lib/python3.9/site-packages/datasets/load.py", line 1731, in load_dataset builder_instance.download_and_prepare( File "/home/leon/.local/lib/python3.9/site-packages/datasets/builder.py", line 613, in download_and_prepare self._download_and_prepare( File "/home/leon/.local/lib/python3.9/site-packages/datasets/builder.py", line 1117, in _download_and_prepare super()._download_and_prepare(dl_manager, verify_infos, check_duplicate_keys=verify_infos) File "/home/leon/.local/lib/python3.9/site-packages/datasets/builder.py", line 684, in _download_and_prepare verify_checksums( File "/home/leon/.local/lib/python3.9/site-packages/datasets/utils/info_utils.py", line 40, in verify_checksums raise NonMatchingChecksumError(error_msg + str(bad_urls)) datasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://unicode.org/udhr/assemblies/udhr_xml.zip', 'https://unicode.org/udhr/assemblies/udhr_txt.zip'] >>> ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.2.1 commit/4110fb6034f79c5fb470cf1043ff52180e9c63b7 - Platform: Linux Ubuntu 20.04 - Python version: 3.9.12 - PyArrow version: 8.0.0
2022-06-08T19:11:21Z
https://github.com/huggingface/datasets/issues/4361
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4361/reactions" }
false