comments_url
stringlengths
70
70
timeline_url
stringlengths
70
70
closed_at
stringlengths
20
20
performed_via_github_app
null
state_reason
stringclasses
3 values
node_id
stringlengths
18
32
state
stringclasses
2 values
assignees
listlengths
0
4
draft
bool
2 classes
number
int64
1.61k
6.73k
user
dict
title
stringlengths
1
290
events_url
stringlengths
68
68
milestone
dict
labels_url
stringlengths
75
75
created_at
stringlengths
20
20
active_lock_reason
null
locked
bool
1 class
assignee
dict
pull_request
dict
id
int64
771M
2.18B
labels
listlengths
0
4
url
stringlengths
61
61
comments
listlengths
0
30
repository_url
stringclasses
1 value
author_association
stringclasses
3 values
body
stringlengths
0
228k
updated_at
stringlengths
20
20
html_url
stringlengths
49
51
reactions
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/2736/comments
https://api.github.com/repos/huggingface/datasets/issues/2736/timeline
null
null
null
MDU6SXNzdWU5NTY4OTUxOTk=
open
[]
null
2,736
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Add Microsoft Building Footprints dataset
https://api.github.com/repos/huggingface/datasets/issues/2736/events
null
https://api.github.com/repos/huggingface/datasets/issues/2736/labels{/name}
2021-07-30T16:17:08Z
null
false
null
null
956,895,199
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" }, { "color": "bfdadc", "default": false, "description": "Vision datasets", "id": 3608941089, "name": "vision", "node_id": "LA_kwDODunzps7XHBIh", "url": "https://api.github.com/repos/huggingface/datasets/labels/vision" } ]
https://api.github.com/repos/huggingface/datasets/issues/2736
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Adding a Dataset - **Name:** Microsoft Building Footprints - **Description:** With the goal to increase the coverage of building footprint data available as open data for OpenStreetMap and humanitarian efforts, we have released millions of building footprints as open data available to download free of charge. - **Paper:** *link to the dataset paper if available* - **Data:** https://www.microsoft.com/en-us/maps/building-footprints - **Motivation:** this can be a useful dataset for researchers working on climate change adaptation, urban studies, geography, etc. Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md). Reported by: @sashavor
2021-12-08T12:09:03Z
https://github.com/huggingface/datasets/issues/2736
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2736/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2735/comments
https://api.github.com/repos/huggingface/datasets/issues/2735/timeline
null
null
null
MDU6SXNzdWU5NTY4ODkzNjU=
open
[]
null
2,735
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Add Open Buildings dataset
https://api.github.com/repos/huggingface/datasets/issues/2735/events
null
https://api.github.com/repos/huggingface/datasets/issues/2735/labels{/name}
2021-07-30T16:08:39Z
null
false
null
null
956,889,365
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
https://api.github.com/repos/huggingface/datasets/issues/2735
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Adding a Dataset - **Name:** Open Buildings - **Description:** A dataset of building footprints to support social good applications. Building footprints are useful for a range of important applications, from population estimation, urban planning and humanitarian response, to environmental and climate science. This large-scale open dataset contains the outlines of buildings derived from high-resolution satellite imagery in order to support these types of uses. The project being based in Ghana, the current focus is on the continent of Africa. See: "Mapping Africa's Buildings with Satellite Imagery" https://ai.googleblog.com/2021/07/mapping-africas-buildings-with.html - **Paper:** https://arxiv.org/abs/2107.12283 - **Data:** https://sites.research.google/open-buildings/ - **Motivation:** *what are some good reasons to have this dataset* Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md). Reported by: @osanseviero
2021-07-31T05:01:25Z
https://github.com/huggingface/datasets/issues/2735
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2735/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2734/comments
https://api.github.com/repos/huggingface/datasets/issues/2734/timeline
2021-07-30T15:47:58Z
null
null
MDExOlB1bGxSZXF1ZXN0NzAwMzc4NjI4
closed
[]
false
2,734
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update BibTeX entry
https://api.github.com/repos/huggingface/datasets/issues/2734/events
null
https://api.github.com/repos/huggingface/datasets/issues/2734/labels{/name}
2021-07-30T15:22:51Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2734.diff", "html_url": "https://github.com/huggingface/datasets/pull/2734", "merged_at": "2021-07-30T15:47:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/2734.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2734" }
956,844,874
[]
https://api.github.com/repos/huggingface/datasets/issues/2734
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Update BibTeX entry.
2021-07-30T15:47:58Z
https://github.com/huggingface/datasets/pull/2734
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2734/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2733/comments
https://api.github.com/repos/huggingface/datasets/issues/2733/timeline
2021-07-30T13:24:30Z
null
null
MDExOlB1bGxSZXF1ZXN0NzAwMjc1NDMy
closed
[]
false
2,733
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Add missing parquet known extension
https://api.github.com/repos/huggingface/datasets/issues/2733/events
null
https://api.github.com/repos/huggingface/datasets/issues/2733/labels{/name}
2021-07-30T13:01:20Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2733.diff", "html_url": "https://github.com/huggingface/datasets/pull/2733", "merged_at": "2021-07-30T13:24:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/2733.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2733" }
956,725,476
[]
https://api.github.com/repos/huggingface/datasets/issues/2733
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This code was failing because the parquet extension wasn't recognized: ```python from datasets import load_dataset base_url = "https://storage.googleapis.com/huggingface-nlp/cache/datasets/wikipedia/20200501.en/1.0.0/" data_files = {"train": base_url + "wikipedia-train.parquet"} wiki = load_dataset("parquet", data_files=data_files, split="train", streaming=True) ``` It raises ```python NotImplementedError: Extraction protocol for file at https://storage.googleapis.com/huggingface-nlp/cache/datasets/wikipedia/20200501.en/1.0.0/wikipedia-train.parquet is not implemented yet ``` I added `parquet` to the list of known extensions EDIT: added pickle, conllu, xml extensions as well
2021-07-30T13:24:31Z
https://github.com/huggingface/datasets/pull/2733
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2733/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2732/comments
https://api.github.com/repos/huggingface/datasets/issues/2732/timeline
2021-07-30T15:58:14Z
null
null
MDExOlB1bGxSZXF1ZXN0NzAwMjMzMzQy
closed
[]
false
2,732
{ "avatar_url": "https://avatars.githubusercontent.com/u/5150963?v=4", "events_url": "https://api.github.com/users/yavuzKomecoglu/events{/privacy}", "followers_url": "https://api.github.com/users/yavuzKomecoglu/followers", "following_url": "https://api.github.com/users/yavuzKomecoglu/following{/other_user}", "gists_url": "https://api.github.com/users/yavuzKomecoglu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yavuzKomecoglu", "id": 5150963, "login": "yavuzKomecoglu", "node_id": "MDQ6VXNlcjUxNTA5NjM=", "organizations_url": "https://api.github.com/users/yavuzKomecoglu/orgs", "received_events_url": "https://api.github.com/users/yavuzKomecoglu/received_events", "repos_url": "https://api.github.com/users/yavuzKomecoglu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yavuzKomecoglu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yavuzKomecoglu/subscriptions", "type": "User", "url": "https://api.github.com/users/yavuzKomecoglu" }
Updated TTC4900 Dataset
https://api.github.com/repos/huggingface/datasets/issues/2732/events
null
https://api.github.com/repos/huggingface/datasets/issues/2732/labels{/name}
2021-07-30T11:52:14Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2732.diff", "html_url": "https://github.com/huggingface/datasets/pull/2732", "merged_at": "2021-07-30T15:58:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/2732.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2732" }
956,676,360
[]
https://api.github.com/repos/huggingface/datasets/issues/2732
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
- The source address of the TTC4900 dataset of [@savasy](https://github.com/savasy) has been updated for direct download. - Updated readme.
2021-07-30T16:00:51Z
https://github.com/huggingface/datasets/pull/2732
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2732/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2731/comments
https://api.github.com/repos/huggingface/datasets/issues/2731/timeline
2021-09-16T13:50:54Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk5NzQwMjg5
closed
[]
false
2,731
{ "avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4", "events_url": "https://api.github.com/users/Rocketknight1/events{/privacy}", "followers_url": "https://api.github.com/users/Rocketknight1/followers", "following_url": "https://api.github.com/users/Rocketknight1/following{/other_user}", "gists_url": "https://api.github.com/users/Rocketknight1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Rocketknight1", "id": 12866554, "login": "Rocketknight1", "node_id": "MDQ6VXNlcjEyODY2NTU0", "organizations_url": "https://api.github.com/users/Rocketknight1/orgs", "received_events_url": "https://api.github.com/users/Rocketknight1/received_events", "repos_url": "https://api.github.com/users/Rocketknight1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Rocketknight1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Rocketknight1/subscriptions", "type": "User", "url": "https://api.github.com/users/Rocketknight1" }
Adding to_tf_dataset method
https://api.github.com/repos/huggingface/datasets/issues/2731/events
null
https://api.github.com/repos/huggingface/datasets/issues/2731/labels{/name}
2021-07-29T18:10:25Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2731.diff", "html_url": "https://github.com/huggingface/datasets/pull/2731", "merged_at": "2021-09-16T13:50:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/2731.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2731" }
956,087,452
[]
https://api.github.com/repos/huggingface/datasets/issues/2731
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Oh my **god** do not merge this yet, it's just a draft. I've added a method (via a mixin) to the `arrow_dataset.Dataset` class that automatically converts our Dataset classes to TF Dataset classes ready for training. It hopefully has most of the features we want, including streaming from disk (no need to load the whole dataset in memory!), correct shuffling, variable-length batches to reduce compute, and correct support for unusual padding. It achieves that by calling the tokenizer `pad` method in the middle of a TF compute graph via a very hacky call to `tf.py_function`, which is heretical but seems to work. A number of issues need to be resolved before it's ready to merge, though: 1) Is a MixIn the right way to do this? Do other classes besides `arrow_dataset.Dataset` need this method too? 2) Needs an argument to support constant-length batches for TPU training - this is easy to add and I'll do it soon. 3) Needs the user to supply the list of columns to drop from the arrow `Dataset`. Is there some automatic way to get the columns we want, or see which columns were added by the tokenizer? 4) Assumes the label column is always present and always called "label" - this is probably not great, but I'm not sure what the 'correct' thing to do here is.
2021-09-16T13:50:54Z
https://github.com/huggingface/datasets/pull/2731
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2731/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2730/comments
https://api.github.com/repos/huggingface/datasets/issues/2730/timeline
null
null
null
MDU6SXNzdWU5NTU5ODc4MzQ=
open
[]
null
2,730
{ "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "events_url": "https://api.github.com/users/yjernite/events{/privacy}", "followers_url": "https://api.github.com/users/yjernite/followers", "following_url": "https://api.github.com/users/yjernite/following{/other_user}", "gists_url": "https://api.github.com/users/yjernite/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yjernite", "id": 10469459, "login": "yjernite", "node_id": "MDQ6VXNlcjEwNDY5NDU5", "organizations_url": "https://api.github.com/users/yjernite/orgs", "received_events_url": "https://api.github.com/users/yjernite/received_events", "repos_url": "https://api.github.com/users/yjernite/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yjernite/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yjernite/subscriptions", "type": "User", "url": "https://api.github.com/users/yjernite" }
Update CommonVoice with new release
https://api.github.com/repos/huggingface/datasets/issues/2730/events
null
https://api.github.com/repos/huggingface/datasets/issues/2730/labels{/name}
2021-07-29T15:59:59Z
null
false
null
null
955,987,834
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
https://api.github.com/repos/huggingface/datasets/issues/2730
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Adding a Dataset - **Name:** CommonVoice mid-2021 release - **Description:** more data in CommonVoice: Languages that have increased the most by percentage are Thai (almost 20x growth, from 12 hours to 250 hours), Luganda (almost 9x growth, from 8 to 80), Esperanto (7x growth, from 100 to 840), and Tamil (almost 8x, from 24 to 220). - **Paper:** https://discourse.mozilla.org/t/common-voice-2021-mid-year-dataset-release/83812 - **Data:** https://commonvoice.mozilla.org/en/datasets - **Motivation:** More data and more varied. I think we just need to add configs in the existing dataset script. Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
2021-08-07T16:19:19Z
https://github.com/huggingface/datasets/issues/2730
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2730/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2729/comments
https://api.github.com/repos/huggingface/datasets/issues/2729/timeline
2021-07-30T13:03:55Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk5NTk5MjA4
closed
[]
false
2,729
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix IndexError while loading Arabic Billion Words dataset
https://api.github.com/repos/huggingface/datasets/issues/2729/events
null
https://api.github.com/repos/huggingface/datasets/issues/2729/labels{/name}
2021-07-29T14:47:02Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2729.diff", "html_url": "https://github.com/huggingface/datasets/pull/2729", "merged_at": "2021-07-30T13:03:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/2729.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2729" }
955,920,489
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2729
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Catch `IndexError` and ignore that record. Close #2727.
2021-07-30T13:03:55Z
https://github.com/huggingface/datasets/pull/2729
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2729/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2728/comments
https://api.github.com/repos/huggingface/datasets/issues/2728/timeline
null
null
null
MDU6SXNzdWU5NTU4OTI5NzA=
open
[]
null
2,728
{ "avatar_url": "https://avatars.githubusercontent.com/u/22492839?v=4", "events_url": "https://api.github.com/users/PierreColombo/events{/privacy}", "followers_url": "https://api.github.com/users/PierreColombo/followers", "following_url": "https://api.github.com/users/PierreColombo/following{/other_user}", "gists_url": "https://api.github.com/users/PierreColombo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/PierreColombo", "id": 22492839, "login": "PierreColombo", "node_id": "MDQ6VXNlcjIyNDkyODM5", "organizations_url": "https://api.github.com/users/PierreColombo/orgs", "received_events_url": "https://api.github.com/users/PierreColombo/received_events", "repos_url": "https://api.github.com/users/PierreColombo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/PierreColombo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/PierreColombo/subscriptions", "type": "User", "url": "https://api.github.com/users/PierreColombo" }
Concurrent use of same dataset (already downloaded)
https://api.github.com/repos/huggingface/datasets/issues/2728/events
null
https://api.github.com/repos/huggingface/datasets/issues/2728/labels{/name}
2021-07-29T14:18:38Z
null
false
null
null
955,892,970
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2728
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug When launching several jobs at the same time loading the same dataset trigger some errors see (last comments). ## Steps to reproduce the bug export HF_DATASETS_CACHE=/gpfswork/rech/toto/datasets for MODEL in "bert-base-uncased" "roberta-base" "distilbert-base-cased"; do # "bert-base-uncased" "bert-large-cased" "roberta-large" "albert-base-v1" "albert-large-v1"; do for TASK_NAME in "mrpc" "rte" 'imdb' "paws" "mnli"; do export OUTPUT_DIR=${MODEL}_${TASK_NAME} sbatch --job-name=${OUTPUT_DIR} \ --gres=gpu:1 \ --no-requeue \ --cpus-per-task=10 \ --hint=nomultithread \ --time=1:00:00 \ --output=jobinfo/${OUTPUT_DIR}_%j.out \ --error=jobinfo/${OUTPUT_DIR}_%j.err \ --qos=qos_gpu-t4 \ --wrap="module purge; module load pytorch-gpu/py3/1.7.0 ; export HF_DATASETS_OFFLINE=1; export HF_DATASETS_CACHE=/gpfswork/rech/toto/datasets; python compute_measures.py --seed=$SEED --saving_path=results --batch_size=$BATCH_SIZE --task_name=$TASK_NAME --model_name=/gpfswork/rech/toto/transformers_models/$MODEL" done done ```python # Sample code to reproduce the bug dataset_train = load_dataset('imdb', split='train', download_mode="reuse_cache_if_exists") dataset_train = dataset_train.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'), batched=True).select(list(range(args.filter))) dataset_val = load_dataset('imdb', split='train', download_mode="reuse_cache_if_exists") dataset_val = dataset_val.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'), batched=True).select(list(range(args.filter, args.filter + 5000))) dataset_test = load_dataset('imdb', split='test', download_mode="reuse_cache_if_exists") dataset_test = dataset_test.map(lambda e: tokenizer(e['text'], truncation=True, padding='max_length'), batched=True) ``` ## Expected results I believe I am doing something wrong with the objects. ## Actual results Traceback (most recent call last): File "/gpfslocalsup/pub/anaconda-py3/2020.02/envs/pytorch-gpu-1.7.0/lib/python3.7/site-packages/datasets/builder.py", line 652, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/gpfslocalsup/pub/anaconda-py3/2020.02/envs/pytorch-gpu-1.7.0/lib/python3.7/site-packages/datasets/builder.py", line 983, in _prepare_split check_duplicates=True, File "/gpfslocalsup/pub/anaconda-py3/2020.02/envs/pytorch-gpu-1.7.0/lib/python3.7/site-packages/datasets/arrow_writer.py", line 192, in __init__ self.stream = pa.OSFile(self._path, "wb") File "pyarrow/io.pxi", line 829, in pyarrow.lib.OSFile.__cinit__ File "pyarrow/io.pxi", line 844, in pyarrow.lib.OSFile._open_writable File "pyarrow/error.pxi", line 122, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 97, in pyarrow.lib.check_status FileNotFoundError: [Errno 2] Failed to open local file '/gpfswork/rech/tts/unm25jp/datasets/paws/labeled_final/1.1.0/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete/paws-test.arrow'. Detail: [errno 2] No such file or directory During handling of the above exception, another exception occurred: Traceback (most recent call last): File "compute_measures.py", line 181, in <module> train_loader, val_loader, test_loader = get_dataloader(args) File "/gpfsdswork/projects/rech/toto/intRAOcular/dataset_utils.py", line 69, in get_dataloader dataset_train = load_dataset('paws', "labeled_final", split='train', download_mode="reuse_cache_if_exists") File "/gpfslocalsup/pub/anaconda-py3/2020.02/envs/pytorch-gpu-1.7.0/lib/python3.7/site-packages/datasets/load.py", line 748, in load_dataset use_auth_token=use_auth_token, File "/gpfslocalsup/pub/anaconda-py3/2020.02/envs/pytorch-gpu-1.7.0/lib/python3.7/site-packages/datasets/builder.py", line 575, in download_and_prepare dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "/gpfslocalsup/pub/anaconda-py3/2020.02/envs/pytorch-gpu-1.7.0/lib/python3.7/site-packages/datasets/builder.py", line 658, in _download_and_prepare + str(e) OSError: Cannot find data file. Original error: [Errno 2] Failed to open local file '/gpfswork/rech/toto/datasets/paws/labeled_final/1.1.0/09d8fae989bb569009a8f5b879ccf2924d3e5cd55bfe2e89e6dab1c0b50ecd34.incomplete/paws-test.arrow'. Detail: [errno 2] No such file or directory ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: datasets==1.8.0 - Platform: linux (jeanzay) - Python version: pyarrow==2.0.0 - PyArrow version: 3.7.8
2021-08-02T07:25:57Z
https://github.com/huggingface/datasets/issues/2728
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2728/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2727/comments
https://api.github.com/repos/huggingface/datasets/issues/2727/timeline
2021-07-30T13:03:55Z
null
completed
MDU6SXNzdWU5NTU4MTIxNDk=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,727
{ "avatar_url": "https://avatars.githubusercontent.com/u/9285264?v=4", "events_url": "https://api.github.com/users/M-Salti/events{/privacy}", "followers_url": "https://api.github.com/users/M-Salti/followers", "following_url": "https://api.github.com/users/M-Salti/following{/other_user}", "gists_url": "https://api.github.com/users/M-Salti/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/M-Salti", "id": 9285264, "login": "M-Salti", "node_id": "MDQ6VXNlcjkyODUyNjQ=", "organizations_url": "https://api.github.com/users/M-Salti/orgs", "received_events_url": "https://api.github.com/users/M-Salti/received_events", "repos_url": "https://api.github.com/users/M-Salti/repos", "site_admin": false, "starred_url": "https://api.github.com/users/M-Salti/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/M-Salti/subscriptions", "type": "User", "url": "https://api.github.com/users/M-Salti" }
Error in loading the Arabic Billion Words Corpus
https://api.github.com/repos/huggingface/datasets/issues/2727/events
null
https://api.github.com/repos/huggingface/datasets/issues/2727/labels{/name}
2021-07-29T12:53:09Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
955,812,149
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2727
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug I get `IndexError: list index out of range` when trying to load the `Techreen` and `Almustaqbal` configs of the dataset. ## Steps to reproduce the bug ```python load_dataset("arabic_billion_words", "Techreen") load_dataset("arabic_billion_words", "Almustaqbal") ``` ## Expected results The datasets load succefully. ## Actual results ```python _extract_tags(self, sample, tag) 139 if len(out) > 0: 140 break --> 141 return out[0] 142 143 def _clean_text(self, text): IndexError: list index out of range ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.10.2 - Platform: Ubuntu 18.04.5 LTS - Python version: 3.7.11 - PyArrow version: 3.0.0
2021-07-30T13:03:55Z
https://github.com/huggingface/datasets/issues/2727
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2727/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2726/comments
https://api.github.com/repos/huggingface/datasets/issues/2726/timeline
2021-07-29T12:00:25Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk5Mzg5MDk1
closed
[]
false
2,726
{ "avatar_url": "https://avatars.githubusercontent.com/u/30535146?v=4", "events_url": "https://api.github.com/users/shabie/events{/privacy}", "followers_url": "https://api.github.com/users/shabie/followers", "following_url": "https://api.github.com/users/shabie/following{/other_user}", "gists_url": "https://api.github.com/users/shabie/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/shabie", "id": 30535146, "login": "shabie", "node_id": "MDQ6VXNlcjMwNTM1MTQ2", "organizations_url": "https://api.github.com/users/shabie/orgs", "received_events_url": "https://api.github.com/users/shabie/received_events", "repos_url": "https://api.github.com/users/shabie/repos", "site_admin": false, "starred_url": "https://api.github.com/users/shabie/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shabie/subscriptions", "type": "User", "url": "https://api.github.com/users/shabie" }
Typo fix `tokenize_exemple`
https://api.github.com/repos/huggingface/datasets/issues/2726/events
null
https://api.github.com/repos/huggingface/datasets/issues/2726/labels{/name}
2021-07-29T10:03:37Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2726.diff", "html_url": "https://github.com/huggingface/datasets/pull/2726", "merged_at": "2021-07-29T12:00:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/2726.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2726" }
955,674,388
[]
https://api.github.com/repos/huggingface/datasets/issues/2726
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
There is a small typo in the main README.md
2021-07-29T12:00:25Z
https://github.com/huggingface/datasets/pull/2726
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2726/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2725/comments
https://api.github.com/repos/huggingface/datasets/issues/2725/timeline
2021-07-28T16:38:02Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk4ODMwNjYw
closed
[]
false
2,725
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Pass use_auth_token to request_etags
https://api.github.com/repos/huggingface/datasets/issues/2725/events
null
https://api.github.com/repos/huggingface/datasets/issues/2725/labels{/name}
2021-07-28T16:13:29Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2725.diff", "html_url": "https://github.com/huggingface/datasets/pull/2725", "merged_at": "2021-07-28T16:38:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/2725.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2725" }
955,020,776
[]
https://api.github.com/repos/huggingface/datasets/issues/2725
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix #2724.
2021-07-28T16:38:02Z
https://github.com/huggingface/datasets/pull/2725
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2725/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2724/comments
https://api.github.com/repos/huggingface/datasets/issues/2724/timeline
2021-07-28T16:38:01Z
null
completed
MDU6SXNzdWU5NTQ5MTk2MDc=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,724
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
404 Error when loading remote data files from private repo
https://api.github.com/repos/huggingface/datasets/issues/2724/events
null
https://api.github.com/repos/huggingface/datasets/issues/2724/labels{/name}
2021-07-28T14:24:23Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
954,919,607
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2724
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Describe the bug When loading remote data files from a private repo, a 404 error is raised. ## Steps to reproduce the bug ```python url = hf_hub_url("lewtun/asr-preds-test", "preds.jsonl", repo_type="dataset") dset = load_dataset("json", data_files=url, use_auth_token=True) # HTTPError: 404 Client Error: Not Found for url: https://huggingface.co/datasets/lewtun/asr-preds-test/resolve/main/preds.jsonl ``` ## Expected results Load dataset. ## Actual results 404 Error.
2021-07-29T04:58:49Z
https://github.com/huggingface/datasets/issues/2724
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2724/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2723/comments
https://api.github.com/repos/huggingface/datasets/issues/2723/timeline
2021-07-28T15:22:23Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk4Njk0NDMw
closed
[]
false
2,723
{ "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "events_url": "https://api.github.com/users/thomasw21/events{/privacy}", "followers_url": "https://api.github.com/users/thomasw21/followers", "following_url": "https://api.github.com/users/thomasw21/following{/other_user}", "gists_url": "https://api.github.com/users/thomasw21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thomasw21", "id": 24695242, "login": "thomasw21", "node_id": "MDQ6VXNlcjI0Njk1MjQy", "organizations_url": "https://api.github.com/users/thomasw21/orgs", "received_events_url": "https://api.github.com/users/thomasw21/received_events", "repos_url": "https://api.github.com/users/thomasw21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thomasw21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomasw21/subscriptions", "type": "User", "url": "https://api.github.com/users/thomasw21" }
Fix en subset by modifying dataset_info with correct validation infos
https://api.github.com/repos/huggingface/datasets/issues/2723/events
null
https://api.github.com/repos/huggingface/datasets/issues/2723/labels{/name}
2021-07-28T13:36:19Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2723.diff", "html_url": "https://github.com/huggingface/datasets/pull/2723", "merged_at": "2021-07-28T15:22:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/2723.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2723" }
954,864,104
[]
https://api.github.com/repos/huggingface/datasets/issues/2723
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
- Related to: #2682 We correct the values of `en` subset concerning the expected validation values (both `num_bytes` and `num_examples`. Instead of having: `{"name": "validation", "num_bytes": 828589180707, "num_examples": 364868892, "dataset_name": "c4"}` We replace with correct values: `{"name": "validation", "num_bytes": 825767266, "num_examples": 364608, "dataset_name": "c4"}` There are still issues with validation with other subsets, but I can't download all the files, unzip to check for the correct number of bytes. (If you have a fast way to obtain those values for other subsets, I can do this in this PR ... otherwise I can't spend those resources)
2021-07-28T15:22:23Z
https://github.com/huggingface/datasets/pull/2723
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2723/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2722/comments
https://api.github.com/repos/huggingface/datasets/issues/2722/timeline
2022-03-21T08:27:51Z
null
completed
MDU6SXNzdWU5NTQ0NDYwNTM=
closed
[]
null
2,722
{ "avatar_url": "https://avatars.githubusercontent.com/u/33200481?v=4", "events_url": "https://api.github.com/users/PosoSAgapo/events{/privacy}", "followers_url": "https://api.github.com/users/PosoSAgapo/followers", "following_url": "https://api.github.com/users/PosoSAgapo/following{/other_user}", "gists_url": "https://api.github.com/users/PosoSAgapo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/PosoSAgapo", "id": 33200481, "login": "PosoSAgapo", "node_id": "MDQ6VXNlcjMzMjAwNDgx", "organizations_url": "https://api.github.com/users/PosoSAgapo/orgs", "received_events_url": "https://api.github.com/users/PosoSAgapo/received_events", "repos_url": "https://api.github.com/users/PosoSAgapo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/PosoSAgapo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/PosoSAgapo/subscriptions", "type": "User", "url": "https://api.github.com/users/PosoSAgapo" }
Missing cache file
https://api.github.com/repos/huggingface/datasets/issues/2722/events
null
https://api.github.com/repos/huggingface/datasets/issues/2722/labels{/name}
2021-07-28T03:52:07Z
null
false
null
null
954,446,053
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2722
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Strangely missing cache file after I restart my program again. `glue_dataset = datasets.load_dataset('glue', 'sst2')` `FileNotFoundError: [Errno 2] No such file or directory: /Users/chris/.cache/huggingface/datasets/glue/sst2/1.0.0/dacbe3125aa31d7f70367a07a8a9e72a5a0bfeb5fc42e75c9db75b96d6053ad/dataset_info.json'`
2022-03-21T08:27:51Z
https://github.com/huggingface/datasets/issues/2722
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2722/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2721/comments
https://api.github.com/repos/huggingface/datasets/issues/2721/timeline
2021-07-28T08:53:18Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk4MTY0Njg3
closed
[]
false
2,721
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Deal with the bad check in test_load.py
https://api.github.com/repos/huggingface/datasets/issues/2721/events
null
https://api.github.com/repos/huggingface/datasets/issues/2721/labels{/name}
2021-07-27T20:23:23Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2721.diff", "html_url": "https://github.com/huggingface/datasets/pull/2721", "merged_at": "2021-07-28T08:53:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/2721.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2721" }
954,238,230
[]
https://api.github.com/repos/huggingface/datasets/issues/2721
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This PR removes a check that's been added in #2684. My intention with this check was to capture an URL in the error message, but instead, it captures a substring of the previous regex match in the test function. Another option would be to replace this check with: ```python m_paths = re.findall(r"\S*_dummy/_dummy.py\b", str(exc_info.value)) # on Linux this will match an URL as well as a local_path due to different os.sep, so take the last element (an URL always comes last in the list) assert len(m_paths) > 0 and is_remote_url(m_paths[-1]) # is_remote_url comes from datasets.utils.file_utils ``` @lhoestq Let me know which one of these two approaches (delete or replace) do you prefer?
2021-07-28T09:58:34Z
https://github.com/huggingface/datasets/pull/2721
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2721/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2720/comments
https://api.github.com/repos/huggingface/datasets/issues/2720/timeline
2021-07-27T18:38:16Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk3OTgxNjMx
closed
[]
false
2,720
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
fix: 🐛 fix two typos
https://api.github.com/repos/huggingface/datasets/issues/2720/events
null
https://api.github.com/repos/huggingface/datasets/issues/2720/labels{/name}
2021-07-27T15:50:17Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2720.diff", "html_url": "https://github.com/huggingface/datasets/pull/2720", "merged_at": "2021-07-27T18:38:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/2720.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2720" }
954,024,426
[]
https://api.github.com/repos/huggingface/datasets/issues/2720
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
2021-07-27T18:38:17Z
https://github.com/huggingface/datasets/pull/2720
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2720/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2719/comments
https://api.github.com/repos/huggingface/datasets/issues/2719/timeline
null
null
null
MDU6SXNzdWU5NTM5MzI0MTY=
open
[]
null
2,719
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
Use ETag in streaming mode to detect resource updates
https://api.github.com/repos/huggingface/datasets/issues/2719/events
null
https://api.github.com/repos/huggingface/datasets/issues/2719/labels{/name}
2021-07-27T14:17:09Z
null
false
null
null
953,932,416
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
https://api.github.com/repos/huggingface/datasets/issues/2719
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
**Is your feature request related to a problem? Please describe.** I want to cache data I generate from processing a dataset I've loaded in streaming mode, but I've currently no way to know if the remote data has been updated or not, thus I don't know when to invalidate my cache. **Describe the solution you'd like** Take the ETag of the data files into account and provide it (directly or through a hash) to give a signal that I can invalidate my cache. **Describe alternatives you've considered** None
2021-10-22T09:36:08Z
https://github.com/huggingface/datasets/issues/2719
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2719/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2718/comments
https://api.github.com/repos/huggingface/datasets/issues/2718/timeline
2021-09-13T17:20:52Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk3NDE0NTQy
closed
[]
false
2,718
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
New documentation structure
https://api.github.com/repos/huggingface/datasets/issues/2718/events
null
https://api.github.com/repos/huggingface/datasets/issues/2718/labels{/name}
2021-07-26T23:15:13Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2718.diff", "html_url": "https://github.com/huggingface/datasets/pull/2718", "merged_at": "2021-09-13T17:20:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/2718.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2718" }
953,360,663
[]
https://api.github.com/repos/huggingface/datasets/issues/2718
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Organize Datasets documentation into four documentation types to improve clarity and discoverability of content. **Content to add in the very short term (feel free to add anything I'm missing):** - A discussion on why Datasets uses Arrow that includes some context and background about why we use Arrow. Would also be great to talk about Datasets speed and performance here, and if you can share any benchmarking/tests you did, that would be awesome! Finally, a discussion about how memory-mapping frees the user from RAM constraints would be very helpful. - Explain why you would want to disable or override verifications when loading a dataset. - If possible, include a code sample of when the number of elements in the field of an output dictionary aren’t the same as the other fields in the output dictionary (taken from the [note](https://huggingface.co/docs/datasets/processing.html#augmenting-the-dataset) here).
2021-09-13T17:20:53Z
https://github.com/huggingface/datasets/pull/2718
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2718/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2717/comments
https://api.github.com/repos/huggingface/datasets/issues/2717/timeline
2021-07-26T16:30:06Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk3MDkzNDEx
closed
[]
false
2,717
{ "avatar_url": "https://avatars.githubusercontent.com/u/7098967?v=4", "events_url": "https://api.github.com/users/amankhandelia/events{/privacy}", "followers_url": "https://api.github.com/users/amankhandelia/followers", "following_url": "https://api.github.com/users/amankhandelia/following{/other_user}", "gists_url": "https://api.github.com/users/amankhandelia/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/amankhandelia", "id": 7098967, "login": "amankhandelia", "node_id": "MDQ6VXNlcjcwOTg5Njc=", "organizations_url": "https://api.github.com/users/amankhandelia/orgs", "received_events_url": "https://api.github.com/users/amankhandelia/received_events", "repos_url": "https://api.github.com/users/amankhandelia/repos", "site_admin": false, "starred_url": "https://api.github.com/users/amankhandelia/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amankhandelia/subscriptions", "type": "User", "url": "https://api.github.com/users/amankhandelia" }
Fix shuffle on IterableDataset that disables batching in case any functions were mapped
https://api.github.com/repos/huggingface/datasets/issues/2717/events
null
https://api.github.com/repos/huggingface/datasets/issues/2717/labels{/name}
2021-07-26T14:42:22Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2717.diff", "html_url": "https://github.com/huggingface/datasets/pull/2717", "merged_at": "2021-07-26T16:30:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/2717.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2717" }
952,979,976
[]
https://api.github.com/repos/huggingface/datasets/issues/2717
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Made a very minor change to fix the issue#2716. Added the missing argument in the constructor call. As discussed in the bug report, the change is made to prevent the `shuffle` method call from resetting the value of `batched` attribute in `MappedExamplesIterable` Fix #2716.
2021-07-26T18:04:14Z
https://github.com/huggingface/datasets/pull/2717
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2717/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2716/comments
https://api.github.com/repos/huggingface/datasets/issues/2716/timeline
2021-07-26T18:04:43Z
null
completed
MDU6SXNzdWU5NTI5MDI3Nzg=
closed
[]
null
2,716
{ "avatar_url": "https://avatars.githubusercontent.com/u/7098967?v=4", "events_url": "https://api.github.com/users/amankhandelia/events{/privacy}", "followers_url": "https://api.github.com/users/amankhandelia/followers", "following_url": "https://api.github.com/users/amankhandelia/following{/other_user}", "gists_url": "https://api.github.com/users/amankhandelia/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/amankhandelia", "id": 7098967, "login": "amankhandelia", "node_id": "MDQ6VXNlcjcwOTg5Njc=", "organizations_url": "https://api.github.com/users/amankhandelia/orgs", "received_events_url": "https://api.github.com/users/amankhandelia/received_events", "repos_url": "https://api.github.com/users/amankhandelia/repos", "site_admin": false, "starred_url": "https://api.github.com/users/amankhandelia/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amankhandelia/subscriptions", "type": "User", "url": "https://api.github.com/users/amankhandelia" }
Calling shuffle on IterableDataset will disable batching in case any functions were mapped
https://api.github.com/repos/huggingface/datasets/issues/2716/events
null
https://api.github.com/repos/huggingface/datasets/issues/2716/labels{/name}
2021-07-26T13:24:59Z
null
false
null
null
952,902,778
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2716
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
When using dataset in streaming mode, if one applies `shuffle` method on the dataset and `map` method for which `batched=True` than the batching operation will not happen, instead `batched` will be set to `False` I did RCA on the dataset codebase, the problem is emerging from [this line of code](https://github.com/huggingface/datasets/blob/d25a0bf94d9f9a9aa6cabdf5b450b9c327d19729/src/datasets/iterable_dataset.py#L197) here as it is `self.ex_iterable.shuffle_data_sources(seed), function=self.function, batch_size=self.batch_size`, as one can see it is missing batched argument, which means that the iterator fallsback to default constructor value, which in this case is `False`. To remedy the problem we can change this line to `self.ex_iterable.shuffle_data_sources(seed), function=self.function, batched=self.batched, batch_size=self.batch_size`
2021-07-26T18:04:43Z
https://github.com/huggingface/datasets/issues/2716
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2716/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2715/comments
https://api.github.com/repos/huggingface/datasets/issues/2715/timeline
2021-07-26T13:27:59Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk2OTc5MjQ1
closed
[]
false
2,715
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update PAN-X data URL in XTREME dataset
https://api.github.com/repos/huggingface/datasets/issues/2715/events
null
https://api.github.com/repos/huggingface/datasets/issues/2715/labels{/name}
2021-07-26T12:21:17Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2715.diff", "html_url": "https://github.com/huggingface/datasets/pull/2715", "merged_at": "2021-07-26T13:27:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/2715.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2715" }
952,845,229
[]
https://api.github.com/repos/huggingface/datasets/issues/2715
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Related to #2710, #2691.
2021-07-26T13:27:59Z
https://github.com/huggingface/datasets/pull/2715
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2715/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2714/comments
https://api.github.com/repos/huggingface/datasets/issues/2714/timeline
null
null
null
MDU6SXNzdWU5NTI1ODA4MjA=
open
[]
null
2,714
{ "avatar_url": "https://avatars.githubusercontent.com/u/1493902?v=4", "events_url": "https://api.github.com/users/pennyl67/events{/privacy}", "followers_url": "https://api.github.com/users/pennyl67/followers", "following_url": "https://api.github.com/users/pennyl67/following{/other_user}", "gists_url": "https://api.github.com/users/pennyl67/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/pennyl67", "id": 1493902, "login": "pennyl67", "node_id": "MDQ6VXNlcjE0OTM5MDI=", "organizations_url": "https://api.github.com/users/pennyl67/orgs", "received_events_url": "https://api.github.com/users/pennyl67/received_events", "repos_url": "https://api.github.com/users/pennyl67/repos", "site_admin": false, "starred_url": "https://api.github.com/users/pennyl67/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pennyl67/subscriptions", "type": "User", "url": "https://api.github.com/users/pennyl67" }
add more precise information for size
https://api.github.com/repos/huggingface/datasets/issues/2714/events
null
https://api.github.com/repos/huggingface/datasets/issues/2714/labels{/name}
2021-07-26T07:11:03Z
null
false
null
null
952,580,820
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2714
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
For the import into ELG, we would like a more precise description of the size of the dataset, instead of the current size categories. The size can be expressed in bytes, or any other preferred size unit. As suggested in the slack channel, perhaps this could be computed with a regex for existing datasets.
2021-07-26T09:16:25Z
https://github.com/huggingface/datasets/issues/2714
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2714/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2713/comments
https://api.github.com/repos/huggingface/datasets/issues/2713/timeline
2021-07-26T09:30:55Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk2Njk3MzU0
closed
[]
false
2,713
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Enumerate all ner_tags values in WNUT 17 dataset
https://api.github.com/repos/huggingface/datasets/issues/2713/events
null
https://api.github.com/repos/huggingface/datasets/issues/2713/labels{/name}
2021-07-26T05:22:16Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2713.diff", "html_url": "https://github.com/huggingface/datasets/pull/2713", "merged_at": "2021-07-26T09:30:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/2713.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2713" }
952,515,256
[]
https://api.github.com/repos/huggingface/datasets/issues/2713
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This PR does: - Enumerate all ner_tags in dataset card Data Fields section - Add all metadata tags to dataset card Close #2709.
2021-07-26T09:30:55Z
https://github.com/huggingface/datasets/pull/2713
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2713/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2710/comments
https://api.github.com/repos/huggingface/datasets/issues/2710/timeline
2021-07-26T09:34:23Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk2MDYyNjAy
closed
[]
false
2,710
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update WikiANN data URL
https://api.github.com/repos/huggingface/datasets/issues/2710/events
null
https://api.github.com/repos/huggingface/datasets/issues/2710/labels{/name}
2021-07-23T16:29:21Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2710.diff", "html_url": "https://github.com/huggingface/datasets/pull/2710", "merged_at": "2021-07-26T09:34:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/2710.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2710" }
951,723,326
[]
https://api.github.com/repos/huggingface/datasets/issues/2710
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
WikiANN data source URL is no longer accessible: 404 error from Dropbox. We have decided to host it at Hugging Face. This PR updates the data source URL, the metadata JSON file and the dataset card. Close #2691.
2021-07-26T09:34:23Z
https://github.com/huggingface/datasets/pull/2710
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2710/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2709/comments
https://api.github.com/repos/huggingface/datasets/issues/2709/timeline
2021-07-26T09:30:55Z
null
completed
MDU6SXNzdWU5NTE1MzQ3NTc=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,709
{ "avatar_url": "https://avatars.githubusercontent.com/u/31095360?v=4", "events_url": "https://api.github.com/users/maxpel/events{/privacy}", "followers_url": "https://api.github.com/users/maxpel/followers", "following_url": "https://api.github.com/users/maxpel/following{/other_user}", "gists_url": "https://api.github.com/users/maxpel/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/maxpel", "id": 31095360, "login": "maxpel", "node_id": "MDQ6VXNlcjMxMDk1MzYw", "organizations_url": "https://api.github.com/users/maxpel/orgs", "received_events_url": "https://api.github.com/users/maxpel/received_events", "repos_url": "https://api.github.com/users/maxpel/repos", "site_admin": false, "starred_url": "https://api.github.com/users/maxpel/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/maxpel/subscriptions", "type": "User", "url": "https://api.github.com/users/maxpel" }
Missing documentation for wnut_17 (ner_tags)
https://api.github.com/repos/huggingface/datasets/issues/2709/events
null
https://api.github.com/repos/huggingface/datasets/issues/2709/labels{/name}
2021-07-23T12:25:32Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
951,534,757
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2709
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
On the info page of the wnut_17 data set (https://huggingface.co/datasets/wnut_17), the model output of ner-tags is only documented for these 5 cases: `ner_tags: a list of classification labels, with possible values including O (0), B-corporation (1), I-corporation (2), B-creative-work (3), I-creative-work (4).` I trained a model with the data and it gives me 13 classes: ``` "id2label": { "0": 0, "1": 1, "2": 2, "3": 3, "4": 4, "5": 5, "6": 6, "7": 7, "8": 8, "9": 9, "10": 10, "11": 11, "12": 12 } "label2id": { "0": 0, "1": 1, "10": 10, "11": 11, "12": 12, "2": 2, "3": 3, "4": 4, "5": 5, "6": 6, "7": 7, "8": 8, "9": 9 } ``` The paper (https://www.aclweb.org/anthology/W17-4418.pdf) explains those 6 categories, but the ordering does not match: ``` 1. person 2. location (including GPE, facility) 3. corporation 4. product (tangible goods, or well-defined services) 5. creative-work (song, movie, book and so on) 6. group (subsuming music band, sports team, and non-corporate organisations) ``` I would be very helpful for me, if somebody could clarify the model ouputs and explain the "B-" and "I-" prefixes to me. Really great work with that and the other packages, I couldn't believe that training the model with that data was basically a one-liner!
2021-07-26T09:30:55Z
https://github.com/huggingface/datasets/issues/2709
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2709/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2708/comments
https://api.github.com/repos/huggingface/datasets/issues/2708/timeline
2021-07-23T13:30:07Z
null
completed
MDU6SXNzdWU5NTEwOTI2NjA=
closed
[]
null
2,708
{ "avatar_url": "https://avatars.githubusercontent.com/u/2441454?v=4", "events_url": "https://api.github.com/users/danyaljj/events{/privacy}", "followers_url": "https://api.github.com/users/danyaljj/followers", "following_url": "https://api.github.com/users/danyaljj/following{/other_user}", "gists_url": "https://api.github.com/users/danyaljj/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/danyaljj", "id": 2441454, "login": "danyaljj", "node_id": "MDQ6VXNlcjI0NDE0NTQ=", "organizations_url": "https://api.github.com/users/danyaljj/orgs", "received_events_url": "https://api.github.com/users/danyaljj/received_events", "repos_url": "https://api.github.com/users/danyaljj/repos", "site_admin": false, "starred_url": "https://api.github.com/users/danyaljj/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/danyaljj/subscriptions", "type": "User", "url": "https://api.github.com/users/danyaljj" }
QASC: incomplete training set
https://api.github.com/repos/huggingface/datasets/issues/2708/events
null
https://api.github.com/repos/huggingface/datasets/issues/2708/labels{/name}
2021-07-22T21:59:44Z
null
false
null
null
951,092,660
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2708
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug The training instances are not loaded properly. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("qasc", script_version='1.10.2') def load_instances(split): instances = dataset[split] print(f"split: {split} - size: {len(instances)}") for x in instances: print(json.dumps(x)) load_instances('test') load_instances('validation') load_instances('train') ``` ## results For test and validation, we can see the examples in the output (which is good!): ``` split: test - size: 920 {"answerKey": "", "choices": {"label": ["A", "B", "C", "D", "E", "F", "G", "H"], "text": ["Anthax", "under water", "uterus", "wombs", "two", "moles", "live", "embryo"]}, "combinedfact": "", "fact1": "", "fact2": "", "formatted_question": "What type of birth do therian mammals have? (A) Anthax (B) under water (C) uterus (D) wombs (E) two (F) moles (G) live (H) embryo", "id": "3C44YUNSI1OBFBB8D36GODNOZN9DPA", "question": "What type of birth do therian mammals have?"} {"answerKey": "", "choices": {"label": ["A", "B", "C", "D", "E", "F", "G", "H"], "text": ["Corvidae", "arthropods", "birds", "backbones", "keratin", "Jurassic", "front paws", "Parakeets."]}, "combinedfact": "", "fact1": "", "fact2": "", "formatted_question": "By what time had mouse-sized viviparous mammals evolved? (A) Corvidae (B) arthropods (C) birds (D) backbones (E) keratin (F) Jurassic (G) front paws (H) Parakeets.", "id": "3B1NLC6UGZVERVLZFT7OUYQLD1SGPZ", "question": "By what time had mouse-sized viviparous mammals evolved?"} {"answerKey": "", "choices": {"label": ["A", "B", "C", "D", "E", "F", "G", "H"], "text": ["Reduced friction", "causes infection", "vital to a good life", "prevents water loss", "camouflage from consumers", "Protection against predators", "spur the growth of the plant", "a smooth surface"]}, "combinedfact": "", "fact1": "", "fact2": "", "formatted_question": "What does a plant's skin do? (A) Reduced friction (B) causes infection (C) vital to a good life (D) prevents water loss (E) camouflage from consumers (F) Protection against predators (G) spur the growth of the plant (H) a smooth surface", "id": "3QRYMNZ7FYGITFVSJET3PS0F4S0NT9", "question": "What does a plant's skin do?"} ... ``` However, only a few instances are loaded for the training split, which is not correct. ## Environment info - `datasets` version: '1.10.2' - Platform: MaxOS - Python version:3.7 - PyArrow version: 3.0.0
2021-07-23T13:30:07Z
https://github.com/huggingface/datasets/issues/2708
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2708/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2707/comments
https://api.github.com/repos/huggingface/datasets/issues/2707/timeline
2021-07-26T14:29:07Z
null
completed
MDU6SXNzdWU5NTA4MTI5NDU=
closed
[]
null
2,707
{ "avatar_url": "https://avatars.githubusercontent.com/u/26467159?v=4", "events_url": "https://api.github.com/users/dwil2444/events{/privacy}", "followers_url": "https://api.github.com/users/dwil2444/followers", "following_url": "https://api.github.com/users/dwil2444/following{/other_user}", "gists_url": "https://api.github.com/users/dwil2444/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dwil2444", "id": 26467159, "login": "dwil2444", "node_id": "MDQ6VXNlcjI2NDY3MTU5", "organizations_url": "https://api.github.com/users/dwil2444/orgs", "received_events_url": "https://api.github.com/users/dwil2444/received_events", "repos_url": "https://api.github.com/users/dwil2444/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dwil2444/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dwil2444/subscriptions", "type": "User", "url": "https://api.github.com/users/dwil2444" }
404 Not Found Error when loading LAMA dataset
https://api.github.com/repos/huggingface/datasets/issues/2707/events
null
https://api.github.com/repos/huggingface/datasets/issues/2707/labels{/name}
2021-07-22T15:52:33Z
null
false
null
null
950,812,945
[]
https://api.github.com/repos/huggingface/datasets/issues/2707
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
The [LAMA](https://huggingface.co/datasets/viewer/?dataset=lama) probing dataset is not available for download: Steps to Reproduce: 1. `from datasets import load_dataset` 2. `dataset = load_dataset('lama', 'trex')`. Results: `FileNotFoundError: Couldn't find file locally at lama/lama.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.2/datasets/lama/lama.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/lama/lama.py`
2021-07-26T14:29:07Z
https://github.com/huggingface/datasets/issues/2707
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2707/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2706/comments
https://api.github.com/repos/huggingface/datasets/issues/2706/timeline
2021-07-22T12:43:00Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk1MTI3ODgz
closed
[]
false
2,706
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update BibTeX entry
https://api.github.com/repos/huggingface/datasets/issues/2706/events
null
https://api.github.com/repos/huggingface/datasets/issues/2706/labels{/name}
2021-07-22T12:29:29Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2706.diff", "html_url": "https://github.com/huggingface/datasets/pull/2706", "merged_at": "2021-07-22T12:43:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/2706.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2706" }
950,606,561
[]
https://api.github.com/repos/huggingface/datasets/issues/2706
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Update BibTeX entry.
2021-07-22T12:43:00Z
https://github.com/huggingface/datasets/pull/2706
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2706/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2705/comments
https://api.github.com/repos/huggingface/datasets/issues/2705/timeline
2021-07-23T08:07:32Z
null
completed
MDU6SXNzdWU5NTA0ODg1ODM=
closed
[]
null
2,705
{ "avatar_url": "https://avatars.githubusercontent.com/u/39296659?v=4", "events_url": "https://api.github.com/users/ronbutan/events{/privacy}", "followers_url": "https://api.github.com/users/ronbutan/followers", "following_url": "https://api.github.com/users/ronbutan/following{/other_user}", "gists_url": "https://api.github.com/users/ronbutan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ronbutan", "id": 39296659, "login": "ronbutan", "node_id": "MDQ6VXNlcjM5Mjk2NjU5", "organizations_url": "https://api.github.com/users/ronbutan/orgs", "received_events_url": "https://api.github.com/users/ronbutan/received_events", "repos_url": "https://api.github.com/users/ronbutan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ronbutan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ronbutan/subscriptions", "type": "User", "url": "https://api.github.com/users/ronbutan" }
404 not found error on loading WIKIANN dataset
https://api.github.com/repos/huggingface/datasets/issues/2705/events
null
https://api.github.com/repos/huggingface/datasets/issues/2705/labels{/name}
2021-07-22T09:55:50Z
null
false
null
null
950,488,583
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2705
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Unable to retreive wikiann English dataset ## Steps to reproduce the bug ```python from datasets import list_datasets, load_dataset, list_metrics, load_metric WIKIANN = load_dataset("wikiann","en") ``` ## Expected results Colab notebook should display successful download status ## Actual results FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/12h3qqog6q4bjve/panx_dataset.tar?dl=1 ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.10.1 - Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.11 - PyArrow version: 3.0.0
2021-07-23T08:07:32Z
https://github.com/huggingface/datasets/issues/2705
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2705/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2704/comments
https://api.github.com/repos/huggingface/datasets/issues/2704/timeline
2021-07-22T10:02:40Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk1MDIzMTEz
closed
[]
false
2,704
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix pick default config name message
https://api.github.com/repos/huggingface/datasets/issues/2704/events
null
https://api.github.com/repos/huggingface/datasets/issues/2704/labels{/name}
2021-07-22T09:49:43Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2704.diff", "html_url": "https://github.com/huggingface/datasets/pull/2704", "merged_at": "2021-07-22T10:02:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/2704.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2704" }
950,483,980
[]
https://api.github.com/repos/huggingface/datasets/issues/2704
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
The error message to tell which config name to load is not displayed. This is because in the code it was considering the config kwargs to be non-empty, which is a special case for custom configs created on the fly. It appears after this change: https://github.com/huggingface/datasets/pull/2659 I fixed that by making the config kwargs empty by default, even if default parameters are passed Fix https://github.com/huggingface/datasets/issues/2703
2021-07-22T10:02:41Z
https://github.com/huggingface/datasets/pull/2704
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2704/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2703/comments
https://api.github.com/repos/huggingface/datasets/issues/2703/timeline
2021-07-22T10:02:40Z
null
completed
MDU6SXNzdWU5NTA0ODIyODQ=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
2,703
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Bad message when config name is missing
https://api.github.com/repos/huggingface/datasets/issues/2703/events
null
https://api.github.com/repos/huggingface/datasets/issues/2703/labels{/name}
2021-07-22T09:47:23Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
950,482,284
[]
https://api.github.com/repos/huggingface/datasets/issues/2703
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
When loading a dataset that have several configurations, we expect to see an error message if the user doesn't specify a config name. However in `datasets` 1.10.0 and 1.10.1 it doesn't show the right message: ```python import datasets datasets.load_dataset("glue") ``` raises ```python AttributeError: 'BuilderConfig' object has no attribute 'text_features' ``` instead of ```python ValueError: Config name is missing. Please pick one among the available configs: ['cola', 'sst2', 'mrpc', 'qqp', 'stsb', 'mnli', 'mnli_mismatched', 'mnli_matched', 'qnli', 'rte', 'wnli', 'ax'] Example of usage: `load_dataset('glue', 'cola')` ```
2021-07-22T10:02:40Z
https://github.com/huggingface/datasets/issues/2703
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2703/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2702/comments
https://api.github.com/repos/huggingface/datasets/issues/2702/timeline
2021-07-22T09:17:38Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk0OTkyOTc1
closed
[]
false
2,702
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update BibTeX entry
https://api.github.com/repos/huggingface/datasets/issues/2702/events
null
https://api.github.com/repos/huggingface/datasets/issues/2702/labels{/name}
2021-07-22T09:04:39Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2702.diff", "html_url": "https://github.com/huggingface/datasets/pull/2702", "merged_at": "2021-07-22T09:17:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/2702.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2702" }
950,448,159
[]
https://api.github.com/repos/huggingface/datasets/issues/2702
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Update BibTeX entry.
2021-07-22T09:17:39Z
https://github.com/huggingface/datasets/pull/2702
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2702/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2701/comments
https://api.github.com/repos/huggingface/datasets/issues/2701/timeline
2021-07-22T09:33:31Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk0OTcxMzM3
closed
[]
false
2,701
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix download_mode docstrings
https://api.github.com/repos/huggingface/datasets/issues/2701/events
null
https://api.github.com/repos/huggingface/datasets/issues/2701/labels{/name}
2021-07-22T08:30:25Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2701.diff", "html_url": "https://github.com/huggingface/datasets/pull/2701", "merged_at": "2021-07-22T09:33:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/2701.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2701" }
950,422,403
[ { "color": "0075ca", "default": true, "description": "Improvements or additions to documentation", "id": 1935892861, "name": "documentation", "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation" } ]
https://api.github.com/repos/huggingface/datasets/issues/2701
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix `download_mode` docstrings.
2021-07-22T09:33:31Z
https://github.com/huggingface/datasets/pull/2701
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2701/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2700/comments
https://api.github.com/repos/huggingface/datasets/issues/2700/timeline
2021-07-22T07:09:07Z
null
completed
MDU6SXNzdWU5NTAyNzYzMjU=
closed
[]
null
2,700
{ "avatar_url": "https://avatars.githubusercontent.com/u/5582286?v=4", "events_url": "https://api.github.com/users/kswamy15/events{/privacy}", "followers_url": "https://api.github.com/users/kswamy15/followers", "following_url": "https://api.github.com/users/kswamy15/following{/other_user}", "gists_url": "https://api.github.com/users/kswamy15/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kswamy15", "id": 5582286, "login": "kswamy15", "node_id": "MDQ6VXNlcjU1ODIyODY=", "organizations_url": "https://api.github.com/users/kswamy15/orgs", "received_events_url": "https://api.github.com/users/kswamy15/received_events", "repos_url": "https://api.github.com/users/kswamy15/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kswamy15/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kswamy15/subscriptions", "type": "User", "url": "https://api.github.com/users/kswamy15" }
from datasets import Dataset is failing
https://api.github.com/repos/huggingface/datasets/issues/2700/events
null
https://api.github.com/repos/huggingface/datasets/issues/2700/labels{/name}
2021-07-22T03:51:23Z
null
false
null
null
950,276,325
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2700
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug A clear and concise description of what the bug is. ## Steps to reproduce the bug ```python # Sample code to reproduce the bug from datasets import Dataset ``` ## Expected results A clear and concise description of the expected results. ## Actual results Specify the actual results or traceback. /usr/local/lib/python3.7/dist-packages/datasets/utils/file_utils.py in <module>() 25 import posixpath 26 import requests ---> 27 from tqdm.contrib.concurrent import thread_map 28 29 from .. import __version__, config, utils ModuleNotFoundError: No module named 'tqdm.contrib.concurrent' --------------------------------------------------------------------------- NOTE: If your import is failing due to a missing package, you can manually install dependencies using either !pip or !apt. To view examples of installing some common dependencies, click the "Open Examples" button below. --------------------------------------------------------------------------- ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: latest version as of 07/21/2021 - Platform: Google Colab - Python version: 3.7 - PyArrow version:
2021-07-22T07:23:45Z
https://github.com/huggingface/datasets/issues/2700
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2700/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2699/comments
https://api.github.com/repos/huggingface/datasets/issues/2699/timeline
null
null
null
MDU6SXNzdWU5NTAyMjEyMjY=
open
[]
null
2,699
{ "avatar_url": "https://avatars.githubusercontent.com/u/4436747?v=4", "events_url": "https://api.github.com/users/eyaler/events{/privacy}", "followers_url": "https://api.github.com/users/eyaler/followers", "following_url": "https://api.github.com/users/eyaler/following{/other_user}", "gists_url": "https://api.github.com/users/eyaler/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/eyaler", "id": 4436747, "login": "eyaler", "node_id": "MDQ6VXNlcjQ0MzY3NDc=", "organizations_url": "https://api.github.com/users/eyaler/orgs", "received_events_url": "https://api.github.com/users/eyaler/received_events", "repos_url": "https://api.github.com/users/eyaler/repos", "site_admin": false, "starred_url": "https://api.github.com/users/eyaler/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eyaler/subscriptions", "type": "User", "url": "https://api.github.com/users/eyaler" }
cannot combine splits merging and streaming?
https://api.github.com/repos/huggingface/datasets/issues/2699/events
null
https://api.github.com/repos/huggingface/datasets/issues/2699/labels{/name}
2021-07-22T01:13:25Z
null
false
null
null
950,221,226
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2699
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
this does not work: `dataset = datasets.load_dataset('mc4','iw',split='train+validation',streaming=True)` with error: `ValueError: Bad split: train+validation. Available splits: ['train', 'validation']` these work: `dataset = datasets.load_dataset('mc4','iw',split='train+validation')` `dataset = datasets.load_dataset('mc4','iw',split='train',streaming=True)` `dataset = datasets.load_dataset('mc4','iw',split='validation',streaming=True)` i could not find a reference to this in the documentation and the error message is confusing. also would be nice to allow streaming for the merged splits
2021-07-22T08:27:47Z
https://github.com/huggingface/datasets/issues/2699
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2699/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2698/comments
https://api.github.com/repos/huggingface/datasets/issues/2698/timeline
2021-07-26T13:25:26Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk0NzUxMzMw
closed
[]
false
2,698
{ "avatar_url": "https://avatars.githubusercontent.com/u/1177582?v=4", "events_url": "https://api.github.com/users/pcuenca/events{/privacy}", "followers_url": "https://api.github.com/users/pcuenca/followers", "following_url": "https://api.github.com/users/pcuenca/following{/other_user}", "gists_url": "https://api.github.com/users/pcuenca/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/pcuenca", "id": 1177582, "login": "pcuenca", "node_id": "MDQ6VXNlcjExNzc1ODI=", "organizations_url": "https://api.github.com/users/pcuenca/orgs", "received_events_url": "https://api.github.com/users/pcuenca/received_events", "repos_url": "https://api.github.com/users/pcuenca/repos", "site_admin": false, "starred_url": "https://api.github.com/users/pcuenca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pcuenca/subscriptions", "type": "User", "url": "https://api.github.com/users/pcuenca" }
Ignore empty batch when writing
https://api.github.com/repos/huggingface/datasets/issues/2698/events
null
https://api.github.com/repos/huggingface/datasets/issues/2698/labels{/name}
2021-07-21T22:35:30Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2698.diff", "html_url": "https://github.com/huggingface/datasets/pull/2698", "merged_at": "2021-07-26T13:25:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/2698.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2698" }
950,159,867
[]
https://api.github.com/repos/huggingface/datasets/issues/2698
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This prevents an schema update with unknown column types, as reported in #2644. This is my first attempt at fixing the issue. I tested the following: - First batch returned by a batched map operation is empty. - An intermediate batch is empty. - `python -m unittest tests.test_arrow_writer` passes. However, `arrow_writer` looks like a pretty generic interface, I'm not sure if there are other uses I may have overlooked. Let me know if that's the case, or if a better approach would be preferable.
2021-07-26T14:56:03Z
https://github.com/huggingface/datasets/pull/2698
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2698/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2697/comments
https://api.github.com/repos/huggingface/datasets/issues/2697/timeline
2021-07-22T07:09:07Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk0NjMyODg0
closed
[]
false
2,697
{ "avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4", "events_url": "https://api.github.com/users/nateraw/events{/privacy}", "followers_url": "https://api.github.com/users/nateraw/followers", "following_url": "https://api.github.com/users/nateraw/following{/other_user}", "gists_url": "https://api.github.com/users/nateraw/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nateraw", "id": 32437151, "login": "nateraw", "node_id": "MDQ6VXNlcjMyNDM3MTUx", "organizations_url": "https://api.github.com/users/nateraw/orgs", "received_events_url": "https://api.github.com/users/nateraw/received_events", "repos_url": "https://api.github.com/users/nateraw/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nateraw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nateraw/subscriptions", "type": "User", "url": "https://api.github.com/users/nateraw" }
Fix import on Colab
https://api.github.com/repos/huggingface/datasets/issues/2697/events
null
https://api.github.com/repos/huggingface/datasets/issues/2697/labels{/name}
2021-07-21T19:03:38Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2697.diff", "html_url": "https://github.com/huggingface/datasets/pull/2697", "merged_at": "2021-07-22T07:09:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/2697.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2697" }
950,021,623
[]
https://api.github.com/repos/huggingface/datasets/issues/2697
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Fix #2695, fix #2700.
2021-07-22T07:09:08Z
https://github.com/huggingface/datasets/pull/2697
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2697/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2696/comments
https://api.github.com/repos/huggingface/datasets/issues/2696/timeline
2021-07-26T09:38:37Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk0NTMwODg3
closed
[]
false
2,696
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Add support for disable_progress_bar on Windows
https://api.github.com/repos/huggingface/datasets/issues/2696/events
null
https://api.github.com/repos/huggingface/datasets/issues/2696/labels{/name}
2021-07-21T16:34:53Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2696.diff", "html_url": "https://github.com/huggingface/datasets/pull/2696", "merged_at": "2021-07-26T09:38:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/2696.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2696" }
949,901,726
[]
https://api.github.com/repos/huggingface/datasets/issues/2696
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This PR is a continuation of #2667 and adds support for `utils.disable_progress_bar()` on Windows when using multiprocessing. This [answer](https://stackoverflow.com/a/6596695/14095927) on SO explains it nicely why the current approach (with calling `utils.is_progress_bar_enabled()` inside `Dataset._map_single`) would not work on Windows.
2021-07-26T13:31:14Z
https://github.com/huggingface/datasets/pull/2696
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2696/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2695/comments
https://api.github.com/repos/huggingface/datasets/issues/2695/timeline
2021-07-22T07:09:07Z
null
completed
MDU6SXNzdWU5NDk4NjQ4MjM=
closed
[]
null
2,695
{ "avatar_url": "https://avatars.githubusercontent.com/u/43239645?v=4", "events_url": "https://api.github.com/users/bayartsogt-ya/events{/privacy}", "followers_url": "https://api.github.com/users/bayartsogt-ya/followers", "following_url": "https://api.github.com/users/bayartsogt-ya/following{/other_user}", "gists_url": "https://api.github.com/users/bayartsogt-ya/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bayartsogt-ya", "id": 43239645, "login": "bayartsogt-ya", "node_id": "MDQ6VXNlcjQzMjM5NjQ1", "organizations_url": "https://api.github.com/users/bayartsogt-ya/orgs", "received_events_url": "https://api.github.com/users/bayartsogt-ya/received_events", "repos_url": "https://api.github.com/users/bayartsogt-ya/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bayartsogt-ya/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bayartsogt-ya/subscriptions", "type": "User", "url": "https://api.github.com/users/bayartsogt-ya" }
Cannot import load_dataset on Colab
https://api.github.com/repos/huggingface/datasets/issues/2695/events
null
https://api.github.com/repos/huggingface/datasets/issues/2695/labels{/name}
2021-07-21T15:52:51Z
null
false
null
null
949,864,823
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2695
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Got tqdm concurrent module not found error during importing load_dataset from datasets. ## Steps to reproduce the bug Here [colab notebook](https://colab.research.google.com/drive/1pErWWnVP4P4mVHjSFUtkePd8Na_Qirg4?usp=sharing) to reproduce the error On colab: ```python !pip install datasets from datasets import load_dataset ``` ## Expected results Works without error ## Actual results Specify the actual results or traceback. ``` ModuleNotFoundError Traceback (most recent call last) <ipython-input-2-8cc7de4c69eb> in <module>() ----> 1 from datasets import load_dataset, load_metric, Metric, MetricInfo, Features, Value 2 from sklearn.metrics import mean_squared_error /usr/local/lib/python3.7/dist-packages/datasets/__init__.py in <module>() 31 ) 32 ---> 33 from .arrow_dataset import Dataset, concatenate_datasets 34 from .arrow_reader import ArrowReader, ReadInstruction 35 from .arrow_writer import ArrowWriter /usr/local/lib/python3.7/dist-packages/datasets/arrow_dataset.py in <module>() 40 from tqdm.auto import tqdm 41 ---> 42 from datasets.tasks.text_classification import TextClassification 43 44 from . import config, utils /usr/local/lib/python3.7/dist-packages/datasets/tasks/__init__.py in <module>() 1 from typing import Optional 2 ----> 3 from ..utils.logging import get_logger 4 from .automatic_speech_recognition import AutomaticSpeechRecognition 5 from .base import TaskTemplate /usr/local/lib/python3.7/dist-packages/datasets/utils/__init__.py in <module>() 19 20 from . import logging ---> 21 from .download_manager import DownloadManager, GenerateMode 22 from .file_utils import DownloadConfig, cached_path, hf_bucket_url, is_remote_url, temp_seed 23 from .mock_download_manager import MockDownloadManager /usr/local/lib/python3.7/dist-packages/datasets/utils/download_manager.py in <module>() 24 25 from .. import config ---> 26 from .file_utils import ( 27 DownloadConfig, 28 cached_path, /usr/local/lib/python3.7/dist-packages/datasets/utils/file_utils.py in <module>() 25 import posixpath 26 import requests ---> 27 from tqdm.contrib.concurrent import thread_map 28 29 from .. import __version__, config, utils ModuleNotFoundError: No module named 'tqdm.contrib.concurrent' ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.10.0 - Platform: Colab - Python version: 3.7.11 - PyArrow version: 3.0.0
2021-07-22T07:26:25Z
https://github.com/huggingface/datasets/issues/2695
{ "+1": 3, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/2695/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2694/comments
https://api.github.com/repos/huggingface/datasets/issues/2694/timeline
2021-07-22T10:41:47Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk0NDg0NTcy
closed
[]
false
2,694
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
fix: 🐛 change string format to allow copy/paste to work in bash
https://api.github.com/repos/huggingface/datasets/issues/2694/events
null
https://api.github.com/repos/huggingface/datasets/issues/2694/labels{/name}
2021-07-21T15:30:40Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2694.diff", "html_url": "https://github.com/huggingface/datasets/pull/2694", "merged_at": "2021-07-22T10:41:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/2694.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2694" }
949,844,722
[]
https://api.github.com/repos/huggingface/datasets/issues/2694
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Before: copy/paste resulted in an error because the square bracket characters `[]` are special characters in bash
2021-07-22T10:41:47Z
https://github.com/huggingface/datasets/pull/2694
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2694/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2693/comments
https://api.github.com/repos/huggingface/datasets/issues/2693/timeline
2021-07-21T14:53:51Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk0NDQ1ODAz
closed
[]
false
2,693
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix OSCAR Esperanto
https://api.github.com/repos/huggingface/datasets/issues/2693/events
null
https://api.github.com/repos/huggingface/datasets/issues/2693/labels{/name}
2021-07-21T14:43:50Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2693.diff", "html_url": "https://github.com/huggingface/datasets/pull/2693", "merged_at": "2021-07-21T14:53:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/2693.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2693" }
949,797,014
[]
https://api.github.com/repos/huggingface/datasets/issues/2693
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
The Esperanto part (original) of OSCAR has the wrong number of examples: ```python from datasets import load_dataset raw_datasets = load_dataset("oscar", "unshuffled_original_eo") ``` raises ```python NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=314188336, num_examples=121171, dataset_name='oscar'), 'recorded': SplitInfo(name='train', num_bytes=314064514, num_examples=121168, dataset_name='oscar')}] ``` I updated the number of expected examples in dataset_infos.json cc @sgugger
2021-07-21T14:53:52Z
https://github.com/huggingface/datasets/pull/2693
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 1, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2693/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2692/comments
https://api.github.com/repos/huggingface/datasets/issues/2692/timeline
2021-07-21T15:31:40Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk0NDE4MDg1
closed
[]
false
2,692
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update BibTeX entry
https://api.github.com/repos/huggingface/datasets/issues/2692/events
null
https://api.github.com/repos/huggingface/datasets/issues/2692/labels{/name}
2021-07-21T14:23:35Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2692.diff", "html_url": "https://github.com/huggingface/datasets/pull/2692", "merged_at": "2021-07-21T15:31:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/2692.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2692" }
949,765,484
[]
https://api.github.com/repos/huggingface/datasets/issues/2692
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Update BibTeX entry
2021-07-21T15:31:41Z
https://github.com/huggingface/datasets/pull/2692
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2692/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2691/comments
https://api.github.com/repos/huggingface/datasets/issues/2691/timeline
2021-07-26T09:34:22Z
null
completed
MDU6SXNzdWU5NDk3NTgzNzk=
closed
[]
null
2,691
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
xtreme / pan-x cannot be downloaded
https://api.github.com/repos/huggingface/datasets/issues/2691/events
null
https://api.github.com/repos/huggingface/datasets/issues/2691/labels{/name}
2021-07-21T14:18:05Z
null
false
null
null
949,758,379
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2691
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug Dataset xtreme / pan-x cannot be loaded Seems related to https://github.com/huggingface/datasets/pull/2326 ## Steps to reproduce the bug ```python dataset = load_dataset("xtreme", "PAN-X.fr") ``` ## Expected results Load the dataset ## Actual results ``` FileNotFoundError: Couldn't find file at https://www.dropbox.com/s/12h3qqog6q4bjve/panx_dataset.tar?dl=1 ``` ## Environment info - `datasets` version: 1.9.0 - Platform: macOS-11.4-x86_64-i386-64bit - Python version: 3.8.11 - PyArrow version: 4.0.1
2021-07-26T09:34:22Z
https://github.com/huggingface/datasets/issues/2691
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2691/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2690/comments
https://api.github.com/repos/huggingface/datasets/issues/2690/timeline
2021-07-27T18:40:54Z
null
null
MDExOlB1bGxSZXF1ZXN0Njk0MjU5MDc1
closed
[]
false
2,690
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
Docs details
https://api.github.com/repos/huggingface/datasets/issues/2690/events
null
https://api.github.com/repos/huggingface/datasets/issues/2690/labels{/name}
2021-07-21T10:43:14Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2690.diff", "html_url": "https://github.com/huggingface/datasets/pull/2690", "merged_at": "2021-07-27T18:40:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/2690.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2690" }
949,574,500
[]
https://api.github.com/repos/huggingface/datasets/issues/2690
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Some comments here: - the code samples assume the expected libraries have already been installed. Maybe add a section at start, or add it to every code sample. Something like `pip install datasets transformers torch 'datasets[streaming]'` (maybe just link to https://huggingface.co/docs/datasets/installation.html + a one-liner that installs all the requirements / alternatively a requirements.txt file) - "If you’d like to play with the examples, you must install it from source." in https://huggingface.co/docs/datasets/installation.html: it's not clear to me what this means (what are these "examples"?) - in https://huggingface.co/docs/datasets/loading_datasets.html: "or AWS bucket if it’s not already stored in the library". It's the only place in the doc (aside from the docstring https://huggingface.co/docs/datasets/package_reference/loading_methods.html?highlight=aws bucket#datasets.list_datasets) where the "AWS bucket" is mentioned. It's not easy to understand what this means. Maybe explain more, and link to https://s3.amazonaws.com/datasets.huggingface.co and/or https://huggingface.co/docs/datasets/filesystems.html. - example in https://huggingface.co/docs/datasets/loading_datasets.html#manually-downloading-files is obsoleted by https://github.com/huggingface/datasets/pull/2326. Also: see https://github.com/huggingface/datasets/issues/2691 for a bug on this specific dataset. - in https://huggingface.co/docs/datasets/loading_datasets.html#manually-downloading-files the doc says "After you’ve downloaded the files, you can point to the folder hosting them locally with the data_dir argument as follows:", but the following example does not show how to use `data_dir` - in https://huggingface.co/docs/datasets/loading_datasets.html#csv-files, it would be nice to have an URL to the csv loader reference (but I'm not sure there is one in the API reference). This comment applies in many places in the doc: I would want the API reference to contain doc for all the code/functions/classes... and I would want a lot more links inside the doc pointing to the API entries. - in the API reference (docstrings) I would prefer "SOURCE" to link to github instead of a copy of the code inside the docs site (eg. https://github.com/huggingface/datasets/blob/master/src/datasets/load.py#L711 instead of https://huggingface.co/docs/datasets/_modules/datasets/load.html#load_dataset) - it seems like not all the API is exposed in the doc. For example, there is no doc for [`disable_progress_bar`](https://github.com/huggingface/datasets/search?q=disable_progress_bar), see https://huggingface.co/docs/datasets/search.html?q=disable_progress_bar, even if the code contains docstrings. Does it mean that the function is not officially supported? (otherwise, maybe it also deserves a mention in https://huggingface.co/docs/datasets/package_reference/logging_methods.html) - in https://huggingface.co/docs/datasets/loading_datasets.html?highlight=most%20efficient%20format%20have%20json%20files%20consisting%20multiple%20json%20objects#json-files, "The most efficient format is to have JSON files consisting of multiple JSON objects, one per line, representing individual data rows:", maybe link to https://en.wikipedia.org/wiki/JSON_streaming#Line-delimited_JSON and give it a name ("line-delimited JSON"? "JSON Lines" as in https://huggingface.co/docs/datasets/processing.html#exporting-a-dataset-to-csv-json-parquet-or-to-python-objects ?) - in https://huggingface.co/docs/datasets/loading_datasets.html, for the local files sections, it would be nice to provide sample csv / json / text files to download, so that it's easier for the reader to try to load them (instead: they won't try) - the doc explains how to shard a dataset, but does not explain why and when a dataset should be sharded (I have no idea... for [parallelizing](https://huggingface.co/docs/datasets/processing.html#multiprocessing)?). It does neither give an idea of the number of shards a dataset typically should have and why. - the code example in https://huggingface.co/docs/datasets/processing.html#mapping-in-a-distributed-setting does not work, because `training_args` has not been defined before in the doc.
2021-07-27T18:40:54Z
https://github.com/huggingface/datasets/pull/2690
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2690/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2689/comments
https://api.github.com/repos/huggingface/datasets/issues/2689/timeline
2021-07-21T13:11:04Z
null
completed
MDU6SXNzdWU5NDk0NDcxMDQ=
closed
[]
null
2,689
{ "avatar_url": "https://avatars.githubusercontent.com/u/25532159?v=4", "events_url": "https://api.github.com/users/PaulLerner/events{/privacy}", "followers_url": "https://api.github.com/users/PaulLerner/followers", "following_url": "https://api.github.com/users/PaulLerner/following{/other_user}", "gists_url": "https://api.github.com/users/PaulLerner/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/PaulLerner", "id": 25532159, "login": "PaulLerner", "node_id": "MDQ6VXNlcjI1NTMyMTU5", "organizations_url": "https://api.github.com/users/PaulLerner/orgs", "received_events_url": "https://api.github.com/users/PaulLerner/received_events", "repos_url": "https://api.github.com/users/PaulLerner/repos", "site_admin": false, "starred_url": "https://api.github.com/users/PaulLerner/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/PaulLerner/subscriptions", "type": "User", "url": "https://api.github.com/users/PaulLerner" }
cannot save the dataset to disk after rename_column
https://api.github.com/repos/huggingface/datasets/issues/2689/events
null
https://api.github.com/repos/huggingface/datasets/issues/2689/labels{/name}
2021-07-21T08:13:40Z
null
false
null
null
949,447,104
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2689
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug If you use `rename_column` and do no other modification, you will be unable to save the dataset using `save_to_disk` ## Steps to reproduce the bug ```python # Sample code to reproduce the bug In [1]: from datasets import Dataset, load_from_disk In [5]: dataset=Dataset.from_dict({'foo': [0]}) In [7]: dataset.save_to_disk('foo') In [8]: dataset=load_from_disk('foo') In [10]: dataset=dataset.rename_column('foo', 'bar') In [11]: dataset.save_to_disk('foo') --------------------------------------------------------------------------- PermissionError Traceback (most recent call last) <ipython-input-11-a3bc0d4fc339> in <module> ----> 1 dataset.save_to_disk('foo') /mnt/beegfs/projects/meerqat/anaconda3/envs/meerqat/lib/python3.7/site-packages/datasets/arrow_dataset.py in save_to_disk(self, dataset_path , fs) 597 if Path(dataset_path, config.DATASET_ARROW_FILENAME) in cache_files_paths: 598 raise PermissionError( --> 599 f"Tried to overwrite {Path(dataset_path, config.DATASET_ARROW_FILENAME)} but a dataset can't overwrite itself." 600 ) 601 if Path(dataset_path, config.DATASET_INDICES_FILENAME) in cache_files_paths: PermissionError: Tried to overwrite foo/dataset.arrow but a dataset can't overwrite itself. ``` N. B. I created the dataset from dict to enable easy reproduction but the same happens if you load an existing dataset (e.g. starting from `In [8]`) ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.8.0 - Platform: Linux-3.10.0-1160.11.1.el7.x86_64-x86_64-with-centos-7.9.2009-Core - Python version: 3.7.10 - PyArrow version: 3.0.0
2023-11-02T14:54:00Z
https://github.com/huggingface/datasets/issues/2689
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2689/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2688/comments
https://api.github.com/repos/huggingface/datasets/issues/2688/timeline
2021-07-21T16:34:53Z
null
completed
MDU6SXNzdWU5NDkxODIwNzQ=
closed
[]
null
2,688
{ "avatar_url": "https://avatars.githubusercontent.com/u/4436747?v=4", "events_url": "https://api.github.com/users/eyaler/events{/privacy}", "followers_url": "https://api.github.com/users/eyaler/followers", "following_url": "https://api.github.com/users/eyaler/following{/other_user}", "gists_url": "https://api.github.com/users/eyaler/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/eyaler", "id": 4436747, "login": "eyaler", "node_id": "MDQ6VXNlcjQ0MzY3NDc=", "organizations_url": "https://api.github.com/users/eyaler/orgs", "received_events_url": "https://api.github.com/users/eyaler/received_events", "repos_url": "https://api.github.com/users/eyaler/repos", "site_admin": false, "starred_url": "https://api.github.com/users/eyaler/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eyaler/subscriptions", "type": "User", "url": "https://api.github.com/users/eyaler" }
hebrew language codes he and iw should be treated as aliases
https://api.github.com/repos/huggingface/datasets/issues/2688/events
null
https://api.github.com/repos/huggingface/datasets/issues/2688/labels{/name}
2021-07-20T23:13:52Z
null
false
null
null
949,182,074
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2688
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
https://huggingface.co/datasets/mc4 not listed when searching for hebrew datasets (he) as it uses the older language code iw, preventing discoverability.
2021-07-21T16:34:53Z
https://github.com/huggingface/datasets/issues/2688
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2688/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2687/comments
https://api.github.com/repos/huggingface/datasets/issues/2687/timeline
2021-07-21T13:04:55Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkzNjY1NDI2
closed
[]
false
2,687
{ "avatar_url": "https://avatars.githubusercontent.com/u/44175589?v=4", "events_url": "https://api.github.com/users/slowwavesleep/events{/privacy}", "followers_url": "https://api.github.com/users/slowwavesleep/followers", "following_url": "https://api.github.com/users/slowwavesleep/following{/other_user}", "gists_url": "https://api.github.com/users/slowwavesleep/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/slowwavesleep", "id": 44175589, "login": "slowwavesleep", "node_id": "MDQ6VXNlcjQ0MTc1NTg5", "organizations_url": "https://api.github.com/users/slowwavesleep/orgs", "received_events_url": "https://api.github.com/users/slowwavesleep/received_events", "repos_url": "https://api.github.com/users/slowwavesleep/repos", "site_admin": false, "starred_url": "https://api.github.com/users/slowwavesleep/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/slowwavesleep/subscriptions", "type": "User", "url": "https://api.github.com/users/slowwavesleep" }
Minor documentation fix
https://api.github.com/repos/huggingface/datasets/issues/2687/events
null
https://api.github.com/repos/huggingface/datasets/issues/2687/labels{/name}
2021-07-20T17:43:23Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2687.diff", "html_url": "https://github.com/huggingface/datasets/pull/2687", "merged_at": "2021-07-21T13:04:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/2687.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2687" }
948,890,481
[]
https://api.github.com/repos/huggingface/datasets/issues/2687
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Currently, [Writing a dataset loading script](https://huggingface.co/docs/datasets/add_dataset.html) page has a small error. A link to `matinf` dataset in [_Dataset scripts of reference_](https://huggingface.co/docs/datasets/add_dataset.html#dataset-scripts-of-reference) section actually leads to `xsquad`, instead. This PR fixes that.
2021-07-21T13:04:55Z
https://github.com/huggingface/datasets/pull/2687
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2687/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2686/comments
https://api.github.com/repos/huggingface/datasets/issues/2686/timeline
2021-07-20T16:27:15Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkzNTk4OTE3
closed
[]
false
2,686
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix bad config ids that name cache directories
https://api.github.com/repos/huggingface/datasets/issues/2686/events
null
https://api.github.com/repos/huggingface/datasets/issues/2686/labels{/name}
2021-07-20T16:00:45Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2686.diff", "html_url": "https://github.com/huggingface/datasets/pull/2686", "merged_at": "2021-07-20T16:27:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/2686.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2686" }
948,811,669
[]
https://api.github.com/repos/huggingface/datasets/issues/2686
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
`data_dir=None` was considered a dataset config parameter, hence creating a special config_id for all dataset being loaded. Since the config_id is used to name the cache directories, this leaded to datasets being regenerated for users. I fixed this by ignoring the value of `data_dir` when it's `None` when computing the config_id. I also added a test to make sure the cache directories are not unexpectedly renamed in the future. Fix https://github.com/huggingface/datasets/issues/2683
2021-07-20T16:27:15Z
https://github.com/huggingface/datasets/pull/2686
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2686/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2685/comments
https://api.github.com/repos/huggingface/datasets/issues/2685/timeline
2021-07-21T13:11:58Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkzNTgxNTk2
closed
[]
false
2,685
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix Blog Authorship Corpus dataset
https://api.github.com/repos/huggingface/datasets/issues/2685/events
null
https://api.github.com/repos/huggingface/datasets/issues/2685/labels{/name}
2021-07-20T15:44:50Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2685.diff", "html_url": "https://github.com/huggingface/datasets/pull/2685", "merged_at": "2021-07-21T13:11:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/2685.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2685" }
948,791,572
[]
https://api.github.com/repos/huggingface/datasets/issues/2685
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This PR: - Update the JSON metadata file, which previously was raising a `NonMatchingSplitsSizesError` - Fix the codec of the data files (`latin_1` instead of `utf-8`), which previously was raising ` UnicodeDecodeError` for some files Close #2679.
2021-07-21T13:11:58Z
https://github.com/huggingface/datasets/pull/2685
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2685/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2684/comments
https://api.github.com/repos/huggingface/datasets/issues/2684/timeline
2021-07-22T14:01:10Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkzNTY0MDY4
closed
[]
false
2,684
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Print absolute local paths in load_dataset error messages
https://api.github.com/repos/huggingface/datasets/issues/2684/events
null
https://api.github.com/repos/huggingface/datasets/issues/2684/labels{/name}
2021-07-20T15:28:28Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2684.diff", "html_url": "https://github.com/huggingface/datasets/pull/2684", "merged_at": "2021-07-22T14:01:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/2684.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2684" }
948,771,753
[]
https://api.github.com/repos/huggingface/datasets/issues/2684
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Use absolute local paths in the error messages of `load_dataset` as per @stas00's suggestion in https://github.com/huggingface/datasets/pull/2500#issuecomment-874891223
2021-07-22T20:48:19Z
https://github.com/huggingface/datasets/pull/2684
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2684/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2683/comments
https://api.github.com/repos/huggingface/datasets/issues/2683/timeline
2021-07-20T16:27:15Z
null
completed
MDU6SXNzdWU5NDg3MjEzNzk=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
2,683
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Cache directories changed due to recent changes in how config kwargs are handled
https://api.github.com/repos/huggingface/datasets/issues/2683/events
null
https://api.github.com/repos/huggingface/datasets/issues/2683/labels{/name}
2021-07-20T14:37:57Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
948,721,379
[]
https://api.github.com/repos/huggingface/datasets/issues/2683
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Since #2659 I can see weird cache directory names with hashes in the config id, even though no additional config kwargs are passed. For example: ```python from datasets import load_dataset_builder c4_builder = load_dataset_builder("c4", "en") print(c4_builder.cache_dir) # /Users/quentinlhoest/.cache/huggingface/datasets/c4/en-174d3b7155eb68db/0.0.0/... # instead of # /Users/quentinlhoest/.cache/huggingface/datasets/c4/en/0.0.0/... ``` This issue could be annoying since it would simply ignore old cache directories for users, and regenerate datasets cc @stas00 this is what you experienced a few days ago
2021-07-20T16:27:15Z
https://github.com/huggingface/datasets/issues/2683
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2683/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2682/comments
https://api.github.com/repos/huggingface/datasets/issues/2682/timeline
2021-07-20T14:38:10Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkzNTE2NjU2
closed
[]
false
2,682
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix c4 expected files
https://api.github.com/repos/huggingface/datasets/issues/2682/events
null
https://api.github.com/repos/huggingface/datasets/issues/2682/labels{/name}
2021-07-20T14:29:31Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2682.diff", "html_url": "https://github.com/huggingface/datasets/pull/2682", "merged_at": "2021-07-20T14:38:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/2682.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2682" }
948,713,137
[]
https://api.github.com/repos/huggingface/datasets/issues/2682
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Some files were not registered in the list of expected files to download Fix https://github.com/huggingface/datasets/issues/2677
2021-07-20T14:38:11Z
https://github.com/huggingface/datasets/pull/2682
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2682/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2681/comments
https://api.github.com/repos/huggingface/datasets/issues/2681/timeline
2021-07-20T15:44:17Z
null
completed
MDU6SXNzdWU5NDg3MDg2NDU=
closed
[]
null
2,681
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
5 duplicate datasets
https://api.github.com/repos/huggingface/datasets/issues/2681/events
null
https://api.github.com/repos/huggingface/datasets/issues/2681/labels{/name}
2021-07-20T14:25:00Z
null
false
null
null
948,708,645
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2681
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug In 5 cases, I could find a dataset on Paperswithcode which references two Hugging Face datasets as dataset loaders. They are: - https://paperswithcode.com/dataset/multinli -> https://huggingface.co/datasets/multi_nli and https://huggingface.co/datasets/multi_nli_mismatch <img width="838" alt="Capture d’écran 2021-07-20 à 16 33 58" src="https://user-images.githubusercontent.com/1676121/126342757-4625522a-f788-41a3-bd1f-2a8b9817bbf5.png"> - https://paperswithcode.com/dataset/squad -> https://huggingface.co/datasets/squad and https://huggingface.co/datasets/squad_v2 - https://paperswithcode.com/dataset/narrativeqa -> https://huggingface.co/datasets/narrativeqa and https://huggingface.co/datasets/narrativeqa_manual - https://paperswithcode.com/dataset/hate-speech-and-offensive-language -> https://huggingface.co/datasets/hate_offensive and https://huggingface.co/datasets/hate_speech_offensive - https://paperswithcode.com/dataset/newsph-nli -> https://huggingface.co/datasets/newsph and https://huggingface.co/datasets/newsph_nli Possible solutions: - don't fix (it works) - for each pair of duplicate datasets, remove one, and create an alias to the other. ## Steps to reproduce the bug Visit the Paperswithcode links, and look at the "Dataset Loaders" section ## Expected results There should only be one reference to a Hugging Face dataset loader ## Actual results Two Hugging Face dataset loaders
2021-07-20T15:44:17Z
https://github.com/huggingface/datasets/issues/2681
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2681/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2680/comments
https://api.github.com/repos/huggingface/datasets/issues/2680/timeline
2021-07-20T14:04:10Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkzNDYyNzY3
closed
[]
false
2,680
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
feat: 🎸 add paperswithcode id for qasper dataset
https://api.github.com/repos/huggingface/datasets/issues/2680/events
null
https://api.github.com/repos/huggingface/datasets/issues/2680/labels{/name}
2021-07-20T13:22:29Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2680.diff", "html_url": "https://github.com/huggingface/datasets/pull/2680", "merged_at": "2021-07-20T14:04:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/2680.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2680" }
948,649,716
[]
https://api.github.com/repos/huggingface/datasets/issues/2680
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
The reverse reference exists on paperswithcode: https://paperswithcode.com/dataset/qasper
2021-07-20T14:04:10Z
https://github.com/huggingface/datasets/pull/2680
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2680/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2679/comments
https://api.github.com/repos/huggingface/datasets/issues/2679/timeline
2021-07-21T13:11:58Z
null
completed
MDU6SXNzdWU5NDg1MDY2Mzg=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,679
{ "avatar_url": "https://avatars.githubusercontent.com/u/38069449?v=4", "events_url": "https://api.github.com/users/izaskr/events{/privacy}", "followers_url": "https://api.github.com/users/izaskr/followers", "following_url": "https://api.github.com/users/izaskr/following{/other_user}", "gists_url": "https://api.github.com/users/izaskr/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/izaskr", "id": 38069449, "login": "izaskr", "node_id": "MDQ6VXNlcjM4MDY5NDQ5", "organizations_url": "https://api.github.com/users/izaskr/orgs", "received_events_url": "https://api.github.com/users/izaskr/received_events", "repos_url": "https://api.github.com/users/izaskr/repos", "site_admin": false, "starred_url": "https://api.github.com/users/izaskr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/izaskr/subscriptions", "type": "User", "url": "https://api.github.com/users/izaskr" }
Cannot load the blog_authorship_corpus due to codec errors
https://api.github.com/repos/huggingface/datasets/issues/2679/events
null
https://api.github.com/repos/huggingface/datasets/issues/2679/labels{/name}
2021-07-20T10:13:20Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
948,506,638
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2679
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug A codec error is raised while loading the blog_authorship_corpus. ## Steps to reproduce the bug ``` from datasets import load_dataset raw_datasets = load_dataset("blog_authorship_corpus") ``` ## Expected results Loading the dataset without errors. ## Actual results An error similar to the one below was raised for (what seems like) every XML file. /home/izaskr/.cache/huggingface/datasets/downloads/extracted/7cf52524f6517e168604b41c6719292e8f97abbe8f731e638b13423f4212359a/blogs/788358.male.24.Arts.Libra.xml cannot be loaded. Error message: 'utf-8' codec can't decode byte 0xe7 in position 7551: invalid continuation byte Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/izaskr/anaconda3/envs/local_vae_older/lib/python3.8/site-packages/datasets/load.py", line 856, in load_dataset builder_instance.download_and_prepare( File "/home/izaskr/anaconda3/envs/local_vae_older/lib/python3.8/site-packages/datasets/builder.py", line 583, in download_and_prepare self._download_and_prepare( File "/home/izaskr/anaconda3/envs/local_vae_older/lib/python3.8/site-packages/datasets/builder.py", line 671, in _download_and_prepare verify_splits(self.info.splits, split_dict) File "/home/izaskr/anaconda3/envs/local_vae_older/lib/python3.8/site-packages/datasets/utils/info_utils.py", line 74, in verify_splits raise NonMatchingSplitsSizesError(str(bad_splits)) datasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dataset_name='blog_authorship_corpus')}, {'expected': SplitInfo(name='validation', num_bytes=37500394, num_examples=31277, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='validation', num_bytes=32553710, num_examples=28521, dataset_name='blog_authorship_corpus')}] ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.9.0 - Platform: Linux-4.15.0-132-generic-x86_64-with-glibc2.10 - Python version: 3.8.8 - PyArrow version: 4.0.1
2021-07-21T17:02:21Z
https://github.com/huggingface/datasets/issues/2679
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2679/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2678/comments
https://api.github.com/repos/huggingface/datasets/issues/2678/timeline
2021-07-21T13:03:02Z
null
completed
MDU6SXNzdWU5NDg0NzEyMjI=
closed
[]
null
2,678
{ "avatar_url": "https://avatars.githubusercontent.com/u/47216475?v=4", "events_url": "https://api.github.com/users/prikmm/events{/privacy}", "followers_url": "https://api.github.com/users/prikmm/followers", "following_url": "https://api.github.com/users/prikmm/following{/other_user}", "gists_url": "https://api.github.com/users/prikmm/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/prikmm", "id": 47216475, "login": "prikmm", "node_id": "MDQ6VXNlcjQ3MjE2NDc1", "organizations_url": "https://api.github.com/users/prikmm/orgs", "received_events_url": "https://api.github.com/users/prikmm/received_events", "repos_url": "https://api.github.com/users/prikmm/repos", "site_admin": false, "starred_url": "https://api.github.com/users/prikmm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/prikmm/subscriptions", "type": "User", "url": "https://api.github.com/users/prikmm" }
Import Error in Kaggle notebook
https://api.github.com/repos/huggingface/datasets/issues/2678/events
null
https://api.github.com/repos/huggingface/datasets/issues/2678/labels{/name}
2021-07-20T09:28:38Z
null
false
null
null
948,471,222
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2678
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Not able to import datasets library in kaggle notebooks ## Steps to reproduce the bug ```python !pip install datasets import datasets ``` ## Expected results No such error ## Actual results ``` ImportError Traceback (most recent call last) <ipython-input-9-652e886d387f> in <module> ----> 1 import datasets /opt/conda/lib/python3.7/site-packages/datasets/__init__.py in <module> 31 ) 32 ---> 33 from .arrow_dataset import Dataset, concatenate_datasets 34 from .arrow_reader import ArrowReader, ReadInstruction 35 from .arrow_writer import ArrowWriter /opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py in <module> 36 import pandas as pd 37 import pyarrow as pa ---> 38 import pyarrow.compute as pc 39 from multiprocess import Pool, RLock 40 from tqdm.auto import tqdm /opt/conda/lib/python3.7/site-packages/pyarrow/compute.py in <module> 16 # under the License. 17 ---> 18 from pyarrow._compute import ( # noqa 19 Function, 20 FunctionOptions, ImportError: /opt/conda/lib/python3.7/site-packages/pyarrow/_compute.cpython-37m-x86_64-linux-gnu.so: undefined symbol: _ZNK5arrow7compute15KernelSignature8ToStringEv ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.9.0 - Platform: Kaggle - Python version: 3.7.10 - PyArrow version: 4.0.1
2021-07-21T13:59:26Z
https://github.com/huggingface/datasets/issues/2678
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2678/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2677/comments
https://api.github.com/repos/huggingface/datasets/issues/2677/timeline
2021-07-20T14:38:10Z
null
completed
MDU6SXNzdWU5NDg0Mjk3ODg=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
2,677
{ "avatar_url": "https://avatars.githubusercontent.com/u/36672861?v=4", "events_url": "https://api.github.com/users/Aktsvigun/events{/privacy}", "followers_url": "https://api.github.com/users/Aktsvigun/followers", "following_url": "https://api.github.com/users/Aktsvigun/following{/other_user}", "gists_url": "https://api.github.com/users/Aktsvigun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Aktsvigun", "id": 36672861, "login": "Aktsvigun", "node_id": "MDQ6VXNlcjM2NjcyODYx", "organizations_url": "https://api.github.com/users/Aktsvigun/orgs", "received_events_url": "https://api.github.com/users/Aktsvigun/received_events", "repos_url": "https://api.github.com/users/Aktsvigun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Aktsvigun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Aktsvigun/subscriptions", "type": "User", "url": "https://api.github.com/users/Aktsvigun" }
Error when downloading C4
https://api.github.com/repos/huggingface/datasets/issues/2677/events
null
https://api.github.com/repos/huggingface/datasets/issues/2677/labels{/name}
2021-07-20T08:37:30Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
948,429,788
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2677
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Hi, I am trying to download `en` corpus from C4 dataset. However, I get an error caused by validation files download (see image). My code is very primitive: `datasets.load_dataset('c4', 'en')` Is this a bug or do I have some configurations missing on my server? Thanks! <img width="1014" alt="Снимок экрана 2021-07-20 в 11 37 17" src="https://user-images.githubusercontent.com/36672861/126289448-6e0db402-5f3f-485a-bf74-eb6e0271fc25.png">
2021-07-20T14:41:31Z
https://github.com/huggingface/datasets/issues/2677
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2677/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2676/comments
https://api.github.com/repos/huggingface/datasets/issues/2676/timeline
2021-07-19T17:51:38Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkyNjc2NTg5
closed
[]
false
2,676
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Increase json reader block_size automatically
https://api.github.com/repos/huggingface/datasets/issues/2676/events
null
https://api.github.com/repos/huggingface/datasets/issues/2676/labels{/name}
2021-07-19T14:51:14Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2676.diff", "html_url": "https://github.com/huggingface/datasets/pull/2676", "merged_at": "2021-07-19T17:51:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/2676.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2676" }
947,734,909
[]
https://api.github.com/repos/huggingface/datasets/issues/2676
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Currently some files can't be read with the default parameters of the JSON lines reader. For example this one: https://huggingface.co/datasets/thomwolf/codeparrot/resolve/main/file-000000000006.json.gz raises a pyarrow error: ```python ArrowInvalid: straddling object straddles two block boundaries (try to increase block size?) ``` The block size that is used is the default one by pyarrow (related to this [jira issue](https://issues.apache.org/jira/browse/ARROW-9612)). To fix this issue I changed the block_size to increase automatically if there is a straddling issue when parsing a batch of json lines. By default the value is `chunksize // 32` in order to leverage multithreading, and it doubles every time a straddling issue occurs. The block_size is then reset for each file. cc @thomwolf @albertvillanova
2021-07-19T17:51:39Z
https://github.com/huggingface/datasets/pull/2676
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2676/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2675/comments
https://api.github.com/repos/huggingface/datasets/issues/2675/timeline
2021-07-19T19:33:25Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkyNjEwNTA1
closed
[]
false
2,675
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Parallelize ETag requests
https://api.github.com/repos/huggingface/datasets/issues/2675/events
null
https://api.github.com/repos/huggingface/datasets/issues/2675/labels{/name}
2021-07-19T13:30:42Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2675.diff", "html_url": "https://github.com/huggingface/datasets/pull/2675", "merged_at": "2021-07-19T19:33:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/2675.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2675" }
947,657,732
[]
https://api.github.com/repos/huggingface/datasets/issues/2675
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Since https://github.com/huggingface/datasets/pull/2628 we use the ETag or the remote data files to compute the directory in the cache where a dataset is saved. This is useful in order to reload the dataset from the cache only if the remote files haven't changed. In this I made the ETag requests parallel using multithreading. There is also a tqdm progress bar that shows up if there are more than 16 data files.
2021-07-19T19:33:25Z
https://github.com/huggingface/datasets/pull/2675
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2675/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2674/comments
https://api.github.com/repos/huggingface/datasets/issues/2674/timeline
2021-07-19T08:07:03Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkyMzMzODU3
closed
[]
false
2,674
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix sacrebleu parameter name
https://api.github.com/repos/huggingface/datasets/issues/2674/events
null
https://api.github.com/repos/huggingface/datasets/issues/2674/labels{/name}
2021-07-19T07:07:26Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2674.diff", "html_url": "https://github.com/huggingface/datasets/pull/2674", "merged_at": "2021-07-19T08:07:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/2674.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2674" }
947,338,202
[]
https://api.github.com/repos/huggingface/datasets/issues/2674
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
DONE: - Fix parameter name: `smooth` to `smooth_method`. - Improve kwargs description. - Align docs on using a metric. - Add example of passing additional arguments in using metrics. Related to #2669.
2021-07-19T08:07:03Z
https://github.com/huggingface/datasets/pull/2674
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2674/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2673/comments
https://api.github.com/repos/huggingface/datasets/issues/2673/timeline
2021-07-19T07:08:03Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkyMzAxMTgw
closed
[]
false
2,673
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix potential DuplicatedKeysError in SQuAD
https://api.github.com/repos/huggingface/datasets/issues/2673/events
null
https://api.github.com/repos/huggingface/datasets/issues/2673/labels{/name}
2021-07-19T06:08:00Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2673.diff", "html_url": "https://github.com/huggingface/datasets/pull/2673", "merged_at": "2021-07-19T07:08:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/2673.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2673" }
947,300,008
[]
https://api.github.com/repos/huggingface/datasets/issues/2673
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
DONE: - Fix potential DiplicatedKeysError by ensuring keys are unique. - Align examples in the docs with SQuAD code. We should promote as a good practice, that the keys should be programmatically generated as unique, instead of read from data (which might be not unique).
2021-07-19T07:08:03Z
https://github.com/huggingface/datasets/pull/2673
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2673/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2672/comments
https://api.github.com/repos/huggingface/datasets/issues/2672/timeline
2021-07-19T06:28:56Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkyMjk2NDQ4
closed
[]
false
2,672
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix potential DuplicatedKeysError in LibriSpeech
https://api.github.com/repos/huggingface/datasets/issues/2672/events
null
https://api.github.com/repos/huggingface/datasets/issues/2672/labels{/name}
2021-07-19T06:00:49Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2672.diff", "html_url": "https://github.com/huggingface/datasets/pull/2672", "merged_at": "2021-07-19T06:28:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/2672.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2672" }
947,294,605
[]
https://api.github.com/repos/huggingface/datasets/issues/2672
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
DONE: - Fix unnecessary path join. - Fix potential DiplicatedKeysError by ensuring keys are unique. We should promote as a good practice, that the keys should be programmatically generated as unique, instead of read from data (which might be not unique).
2021-07-19T06:28:57Z
https://github.com/huggingface/datasets/pull/2672
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2672/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2671/comments
https://api.github.com/repos/huggingface/datasets/issues/2671/timeline
2021-07-19T06:45:50Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkyMjc5MTM0
closed
[]
false
2,671
{ "avatar_url": "https://avatars.githubusercontent.com/u/32900185?v=4", "events_url": "https://api.github.com/users/aslihanuysall/events{/privacy}", "followers_url": "https://api.github.com/users/aslihanuysall/followers", "following_url": "https://api.github.com/users/aslihanuysall/following{/other_user}", "gists_url": "https://api.github.com/users/aslihanuysall/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/aslihanuysall", "id": 32900185, "login": "aslihanuysall", "node_id": "MDQ6VXNlcjMyOTAwMTg1", "organizations_url": "https://api.github.com/users/aslihanuysall/orgs", "received_events_url": "https://api.github.com/users/aslihanuysall/received_events", "repos_url": "https://api.github.com/users/aslihanuysall/repos", "site_admin": false, "starred_url": "https://api.github.com/users/aslihanuysall/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/aslihanuysall/subscriptions", "type": "User", "url": "https://api.github.com/users/aslihanuysall" }
Mesinesp development and training data sets have been added.
https://api.github.com/repos/huggingface/datasets/issues/2671/events
null
https://api.github.com/repos/huggingface/datasets/issues/2671/labels{/name}
2021-07-19T05:14:38Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2671.diff", "html_url": "https://github.com/huggingface/datasets/pull/2671", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2671.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2671" }
947,273,875
[]
https://api.github.com/repos/huggingface/datasets/issues/2671
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
https://zenodo.org/search?page=1&size=20&q=mesinesp, Mesinesp has Medical Semantic Indexed records in Spanish. Indexing is done using DeCS codes, a sort of Spanish equivalent to MeSH terms. The Mesinesp (Spanish BioASQ track, see https://temu.bsc.es/mesinesp) development set has a total of 750 records. The Mesinesp (Spanish BioASQ track, see https://temu.bsc.es/mesinesp) training set has a total of 369,368 records.
2021-07-19T07:32:28Z
https://github.com/huggingface/datasets/pull/2671
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2671/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2670/comments
https://api.github.com/repos/huggingface/datasets/issues/2670/timeline
null
null
null
MDU6SXNzdWU5NDcxMjA3MDk=
open
[]
null
2,670
{ "avatar_url": "https://avatars.githubusercontent.com/u/5583410?v=4", "events_url": "https://api.github.com/users/ggdupont/events{/privacy}", "followers_url": "https://api.github.com/users/ggdupont/followers", "following_url": "https://api.github.com/users/ggdupont/following{/other_user}", "gists_url": "https://api.github.com/users/ggdupont/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ggdupont", "id": 5583410, "login": "ggdupont", "node_id": "MDQ6VXNlcjU1ODM0MTA=", "organizations_url": "https://api.github.com/users/ggdupont/orgs", "received_events_url": "https://api.github.com/users/ggdupont/received_events", "repos_url": "https://api.github.com/users/ggdupont/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ggdupont/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ggdupont/subscriptions", "type": "User", "url": "https://api.github.com/users/ggdupont" }
Using sharding to parallelize indexing
https://api.github.com/repos/huggingface/datasets/issues/2670/events
null
https://api.github.com/repos/huggingface/datasets/issues/2670/labels{/name}
2021-07-18T21:26:26Z
null
false
null
null
947,120,709
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2670
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
**Is your feature request related to a problem? Please describe.** Creating an elasticsearch index on large dataset could be quite long and cannot be parallelized on shard (the index creation is colliding) **Describe the solution you'd like** When working on dataset shards, if an index already exists, its mapping should be checked and if compatible, the indexing process should continue with the shard data. Additionally, at the end of the process, the `_indexes` dict should be send back to the original dataset object (from which the shards have been created) to allow to use the index for later filtering on the whole dataset. **Describe alternatives you've considered** Each dataset shard could created independent partial indices. then on the whole dataset level, indices should be all referred in `_indexes` dict and be used in querying through `get_nearest_examples()`. The drawback is that the scores will be computed independently on the partial indices leading to inconsistent values for most scoring based on corpus level statistics (tf/idf, BM25). **Additional context** The objectives is to parallelize the index creation to speed-up the process (ie surcharging the ES server which is fine to handle large load) while later enabling search on the whole dataset.
2021-10-07T13:33:25Z
https://github.com/huggingface/datasets/issues/2670
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 4, "url": "https://api.github.com/repos/huggingface/datasets/issues/2670/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2669/comments
https://api.github.com/repos/huggingface/datasets/issues/2669/timeline
2021-07-18T11:19:04Z
null
completed
MDU6SXNzdWU5NDY5ODI5OTg=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,669
{ "avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4", "events_url": "https://api.github.com/users/BramVanroy/events{/privacy}", "followers_url": "https://api.github.com/users/BramVanroy/followers", "following_url": "https://api.github.com/users/BramVanroy/following{/other_user}", "gists_url": "https://api.github.com/users/BramVanroy/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/BramVanroy", "id": 2779410, "login": "BramVanroy", "node_id": "MDQ6VXNlcjI3Nzk0MTA=", "organizations_url": "https://api.github.com/users/BramVanroy/orgs", "received_events_url": "https://api.github.com/users/BramVanroy/received_events", "repos_url": "https://api.github.com/users/BramVanroy/repos", "site_admin": false, "starred_url": "https://api.github.com/users/BramVanroy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BramVanroy/subscriptions", "type": "User", "url": "https://api.github.com/users/BramVanroy" }
Metric kwargs are not passed to underlying external metric f1_score
https://api.github.com/repos/huggingface/datasets/issues/2669/events
null
https://api.github.com/repos/huggingface/datasets/issues/2669/labels{/name}
2021-07-18T08:32:31Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
946,982,998
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2669
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug When I want to use F1 score with average="min", this keyword argument does not seem to be passed through to the underlying sklearn metric. This is evident because [sklearn](https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html) throws an error telling me so. ## Steps to reproduce the bug ```python import datasets f1 = datasets.load_metric("f1", keep_in_memory=True, average="min") f1.add_batch(predictions=[0,2,3], references=[1, 2, 3]) f1.compute() ``` ## Expected results No error, because `average="min"` should be passed correctly to f1_score in sklearn. ## Actual results ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "C:\Users\bramv\.virtualenvs\pipeline-TpEsXVex\lib\site-packages\datasets\metric.py", line 402, in compute output = self._compute(predictions=predictions, references=references, **kwargs) File "C:\Users\bramv\.cache\huggingface\modules\datasets_modules\metrics\f1\82177930a325d4c28342bba0f116d73f6d92fb0c44cd67be32a07c1262b61cfe\f1.py", line 97, in _compute "f1": f1_score( File "C:\Users\bramv\.virtualenvs\pipeline-TpEsXVex\lib\site-packages\sklearn\utils\validation.py", line 63, in inner_f return f(*args, **kwargs) File "C:\Users\bramv\.virtualenvs\pipeline-TpEsXVex\lib\site-packages\sklearn\metrics\_classification.py", line 1071, in f1_score return fbeta_score(y_true, y_pred, beta=1, labels=labels, File "C:\Users\bramv\.virtualenvs\pipeline-TpEsXVex\lib\site-packages\sklearn\utils\validation.py", line 63, in inner_f return f(*args, **kwargs) File "C:\Users\bramv\.virtualenvs\pipeline-TpEsXVex\lib\site-packages\sklearn\metrics\_classification.py", line 1195, in fbeta_score _, _, f, _ = precision_recall_fscore_support(y_true, y_pred, File "C:\Users\bramv\.virtualenvs\pipeline-TpEsXVex\lib\site-packages\sklearn\utils\validation.py", line 63, in inner_f return f(*args, **kwargs) File "C:\Users\bramv\.virtualenvs\pipeline-TpEsXVex\lib\site-packages\sklearn\metrics\_classification.py", line 1464, in precision_recall_fscore_support labels = _check_set_wise_labels(y_true, y_pred, average, labels, File "C:\Users\bramv\.virtualenvs\pipeline-TpEsXVex\lib\site-packages\sklearn\metrics\_classification.py", line 1294, in _check_set_wise_labels raise ValueError("Target is %s but average='binary'. Please " ValueError: Target is multiclass but average='binary'. Please choose another average setting, one of [None, 'micro', 'macro', 'weighted']. ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.9.0 - Platform: Windows-10-10.0.19041-SP0 - Python version: 3.9.2 - PyArrow version: 4.0.1
2021-07-18T18:36:05Z
https://github.com/huggingface/datasets/issues/2669
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2669/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2668/comments
https://api.github.com/repos/huggingface/datasets/issues/2668/timeline
2021-07-29T11:50:31Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkxOTY1MTY1
closed
[]
false
2,668
{ "avatar_url": "https://avatars.githubusercontent.com/u/44175589?v=4", "events_url": "https://api.github.com/users/slowwavesleep/events{/privacy}", "followers_url": "https://api.github.com/users/slowwavesleep/followers", "following_url": "https://api.github.com/users/slowwavesleep/following{/other_user}", "gists_url": "https://api.github.com/users/slowwavesleep/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/slowwavesleep", "id": 44175589, "login": "slowwavesleep", "node_id": "MDQ6VXNlcjQ0MTc1NTg5", "organizations_url": "https://api.github.com/users/slowwavesleep/orgs", "received_events_url": "https://api.github.com/users/slowwavesleep/received_events", "repos_url": "https://api.github.com/users/slowwavesleep/repos", "site_admin": false, "starred_url": "https://api.github.com/users/slowwavesleep/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/slowwavesleep/subscriptions", "type": "User", "url": "https://api.github.com/users/slowwavesleep" }
Add Russian SuperGLUE
https://api.github.com/repos/huggingface/datasets/issues/2668/events
null
https://api.github.com/repos/huggingface/datasets/issues/2668/labels{/name}
2021-07-17T17:41:28Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2668.diff", "html_url": "https://github.com/huggingface/datasets/pull/2668", "merged_at": "2021-07-29T11:50:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/2668.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2668" }
946,867,622
[]
https://api.github.com/repos/huggingface/datasets/issues/2668
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Hi, This adds the [Russian SuperGLUE](https://russiansuperglue.com/) dataset. For the most part I reused the code for the original SuperGLUE, although there are some relatively minor differences in the structure that I accounted for.
2021-07-29T11:50:31Z
https://github.com/huggingface/datasets/pull/2668
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2668/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2667/comments
https://api.github.com/repos/huggingface/datasets/issues/2667/timeline
2021-07-19T17:32:00Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkxOTYwNzc3
closed
[]
false
2,667
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Use tqdm from tqdm_utils
https://api.github.com/repos/huggingface/datasets/issues/2667/events
null
https://api.github.com/repos/huggingface/datasets/issues/2667/labels{/name}
2021-07-17T17:06:35Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2667.diff", "html_url": "https://github.com/huggingface/datasets/pull/2667", "merged_at": "2021-07-19T17:32:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/2667.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2667" }
946,861,908
[]
https://api.github.com/repos/huggingface/datasets/issues/2667
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This PR replaces `tqdm` from the `tqdm` lib with `tqdm` from `datasets.utils.tqdm_utils`. With this change, it's possible to disable progress bars just by calling `disable_progress_bar`. Note this doesn't work on Windows when using multiprocessing due to how global variables are shared between processes. Currently, there is no easy way to disable progress bars in a multiprocess setting on Windows (patching logging with `datasets.utils.logging.get_verbosity = lambda: datasets.utils.logging.NOTSET` doesn't seem to work as well), so adding support for this is a future goal. Additionally, this PR adds a unit ("ba" for batches) to the bar printed by `Dataset.to_json` (this change is motivated by https://github.com/huggingface/datasets/issues/2657).
2021-07-19T17:39:10Z
https://github.com/huggingface/datasets/pull/2667
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2667/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2666/comments
https://api.github.com/repos/huggingface/datasets/issues/2666/timeline
2022-10-03T09:37:35Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkxOTMzMDM1
closed
[]
true
2,666
{ "avatar_url": "https://avatars.githubusercontent.com/u/69807323?v=4", "events_url": "https://api.github.com/users/arampacha/events{/privacy}", "followers_url": "https://api.github.com/users/arampacha/followers", "following_url": "https://api.github.com/users/arampacha/following{/other_user}", "gists_url": "https://api.github.com/users/arampacha/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/arampacha", "id": 69807323, "login": "arampacha", "node_id": "MDQ6VXNlcjY5ODA3MzIz", "organizations_url": "https://api.github.com/users/arampacha/orgs", "received_events_url": "https://api.github.com/users/arampacha/received_events", "repos_url": "https://api.github.com/users/arampacha/repos", "site_admin": false, "starred_url": "https://api.github.com/users/arampacha/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/arampacha/subscriptions", "type": "User", "url": "https://api.github.com/users/arampacha" }
Adds CodeClippy dataset [WIP]
https://api.github.com/repos/huggingface/datasets/issues/2666/events
null
https://api.github.com/repos/huggingface/datasets/issues/2666/labels{/name}
2021-07-17T13:32:04Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2666.diff", "html_url": "https://github.com/huggingface/datasets/pull/2666", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2666.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2666" }
946,825,140
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
https://api.github.com/repos/huggingface/datasets/issues/2666
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
CodeClippy is an opensource code dataset scrapped from github during flax-jax-community-week https://the-eye.eu/public/AI/training_data/code_clippy_data/
2023-07-26T23:06:01Z
https://github.com/huggingface/datasets/pull/2666
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2666/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2665/comments
https://api.github.com/repos/huggingface/datasets/issues/2665/timeline
2022-10-03T09:38:10Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkxOTMwNjky
closed
[]
true
2,665
{ "avatar_url": "https://avatars.githubusercontent.com/u/69807323?v=4", "events_url": "https://api.github.com/users/arampacha/events{/privacy}", "followers_url": "https://api.github.com/users/arampacha/followers", "following_url": "https://api.github.com/users/arampacha/following{/other_user}", "gists_url": "https://api.github.com/users/arampacha/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/arampacha", "id": 69807323, "login": "arampacha", "node_id": "MDQ6VXNlcjY5ODA3MzIz", "organizations_url": "https://api.github.com/users/arampacha/orgs", "received_events_url": "https://api.github.com/users/arampacha/received_events", "repos_url": "https://api.github.com/users/arampacha/repos", "site_admin": false, "starred_url": "https://api.github.com/users/arampacha/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/arampacha/subscriptions", "type": "User", "url": "https://api.github.com/users/arampacha" }
Adds APPS dataset to the hub [WIP]
https://api.github.com/repos/huggingface/datasets/issues/2665/events
null
https://api.github.com/repos/huggingface/datasets/issues/2665/labels{/name}
2021-07-17T13:13:17Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2665.diff", "html_url": "https://github.com/huggingface/datasets/pull/2665", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2665.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2665" }
946,822,036
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
https://api.github.com/repos/huggingface/datasets/issues/2665
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
A loading script for [APPS dataset](https://github.com/hendrycks/apps)
2022-10-03T09:38:10Z
https://github.com/huggingface/datasets/pull/2665
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2665/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2663/comments
https://api.github.com/repos/huggingface/datasets/issues/2663/timeline
2021-09-13T13:56:37Z
null
completed
MDU6SXNzdWU5NDY1NTIyNzM=
closed
[]
null
2,663
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stas00", "id": 10676103, "login": "stas00", "node_id": "MDQ6VXNlcjEwNjc2MTAz", "organizations_url": "https://api.github.com/users/stas00/orgs", "received_events_url": "https://api.github.com/users/stas00/received_events", "repos_url": "https://api.github.com/users/stas00/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "type": "User", "url": "https://api.github.com/users/stas00" }
[`to_json`] add multi-proc sharding support
https://api.github.com/repos/huggingface/datasets/issues/2663/events
null
https://api.github.com/repos/huggingface/datasets/issues/2663/labels{/name}
2021-07-16T19:41:50Z
null
false
null
null
946,552,273
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2663
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
As discussed on slack it appears that `to_json` is quite slow on huge datasets like OSCAR. I implemented sharded saving, which is much much faster - but the tqdm bars all overwrite each other, so it's hard to make sense of the progress, so if possible ideally this multi-proc support could be implemented internally in `to_json` via `num_proc` argument. I guess `num_proc` will be the number of shards? I think the user will need to use this feature wisely, since too many processes writing to say normal style HD is likely to be slower than one process. I'm not sure whether the user should be responsible to concatenate the shards at the end or `datasets`, either way works for my needs. The code I was using: ``` from multiprocessing import cpu_count, Process, Queue [...] filtered_dataset = concat_dataset.map(filter_short_documents, batched=True, batch_size=256, num_proc=cpu_count()) DATASET_NAME = "oscar" SHARDS = 10 def process_shard(idx): print(f"Sharding {idx}") ds_shard = filtered_dataset.shard(SHARDS, idx, contiguous=True) # ds_shard = ds_shard.shuffle() # remove contiguous=True above if shuffling print(f"Saving {DATASET_NAME}-{idx}.jsonl") ds_shard.to_json(f"{DATASET_NAME}-{idx}.jsonl", orient="records", lines=True, force_ascii=False) queue = Queue() processes = [Process(target=process_shard, args=(idx,)) for idx in range(SHARDS)] for p in processes: p.start() for p in processes: p.join() ``` Thank you! @lhoestq
2021-09-13T13:56:37Z
https://github.com/huggingface/datasets/issues/2663
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2663/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2662/comments
https://api.github.com/repos/huggingface/datasets/issues/2662/timeline
2021-08-25T14:18:08Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkxNjM5MjU5
closed
[]
false
2,662
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Load Dataset from the Hub (NO DATASET SCRIPT)
https://api.github.com/repos/huggingface/datasets/issues/2662/events
null
https://api.github.com/repos/huggingface/datasets/issues/2662/labels{/name}
2021-07-16T17:21:58Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2662.diff", "html_url": "https://github.com/huggingface/datasets/pull/2662", "merged_at": "2021-08-25T14:18:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/2662.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2662" }
946,470,815
[]
https://api.github.com/repos/huggingface/datasets/issues/2662
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Load the data from any Dataset repository on the Hub This PR adds support for loading datasets from any dataset repository on the hub, without requiring any dataset script. As a user it's now possible to create a repo and upload some csv/json/text/parquet files, and then be able to load the data in one line. Here is an example with the `allenai/c4` repository that contains a lot of compressed json lines files: ```python from datasets import load_dataset data_files = {"train": "en/c4-train.*.json.gz"} c4 = load_dataset("allenai/c4", data_files=data_files, split="train", streaming=True) print(c4.n_shards) # 1024 print(next(iter(c4))) # {'text': 'Beginners BBQ Class Takin...'} ``` By default it loads all the files, but as shown in the example you can choose the ones you want with unix style patterns. Of course it's still possible to use dataset scripts since they offer the most flexibility. ## Implementation details It uses `huggingface_hub` to list the files in a dataset repository. If you provide a path to a local directory instead of a repository name, it works the same way but it uses `glob`. Depending on the data files available, or passed in the `data_files` parameter, one of the available builders will be used among the csv, json, text and parquet builders. Because of this, it's not possible to load both csv and json files at once. In this case you have to load them separately and then concatenate the two datasets for example. ## TODO - [x] tests - [x] docs - [x] when huggingface_hub gets a new release, update the CI and the setup.py Close https://github.com/huggingface/datasets/issues/2629
2021-08-25T14:53:01Z
https://github.com/huggingface/datasets/pull/2662
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 5, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 5, "url": "https://api.github.com/repos/huggingface/datasets/issues/2662/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2661/comments
https://api.github.com/repos/huggingface/datasets/issues/2661/timeline
2021-08-04T17:03:53Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkxNjE5MzAz
closed
[]
false
2,661
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Add SD task for SUPERB
https://api.github.com/repos/huggingface/datasets/issues/2661/events
null
https://api.github.com/repos/huggingface/datasets/issues/2661/labels{/name}
2021-07-16T16:43:21Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2661.diff", "html_url": "https://github.com/huggingface/datasets/pull/2661", "merged_at": "2021-08-04T17:03:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/2661.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2661" }
946,446,967
[]
https://api.github.com/repos/huggingface/datasets/issues/2661
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Include the SD (Speaker Diarization) task as described in the [SUPERB paper](https://arxiv.org/abs/2105.01051) and `s3prl` [instructions](https://github.com/s3prl/s3prl/tree/master/s3prl/downstream#sd-speaker-diarization). TODO: - [x] Generate the LibriMix corpus - [x] Prepare the corpus for diarization - [x] Upload these files to the superb-data repo - [x] Transcribe the corresponding s3prl processing of these files into our superb loading script - [x] README: tags + description sections - ~~Add DER metric~~ (we leave the DER metric for a follow-up PR) Related to #2619. Close #2653. cc: @lewtun
2021-08-04T17:03:53Z
https://github.com/huggingface/datasets/pull/2661
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2661/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2660/comments
https://api.github.com/repos/huggingface/datasets/issues/2660/timeline
2021-09-06T14:12:23Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkxNTA4NzE0
closed
[]
false
2,660
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Move checks from _map_single to map
https://api.github.com/repos/huggingface/datasets/issues/2660/events
null
https://api.github.com/repos/huggingface/datasets/issues/2660/labels{/name}
2021-07-16T13:53:33Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2660.diff", "html_url": "https://github.com/huggingface/datasets/pull/2660", "merged_at": "2021-09-06T14:12:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/2660.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2660" }
946,316,180
[]
https://api.github.com/repos/huggingface/datasets/issues/2660
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
The goal of this PR is to remove duplicated checks in the `map` logic to execute them only once whenever possible (`fn_kwargs`, `input_columns`, ...). Additionally, this PR improves the consistency (to align it with `input_columns`) of the `remove_columns` check by adding support for a single string value, which is then wrapped into a list.
2021-09-06T14:12:23Z
https://github.com/huggingface/datasets/pull/2660
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2660/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2659/comments
https://api.github.com/repos/huggingface/datasets/issues/2659/timeline
2021-07-16T12:46:07Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkxMzcwNzU3
closed
[]
false
2,659
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Allow dataset config kwargs to be None
https://api.github.com/repos/huggingface/datasets/issues/2659/events
null
https://api.github.com/repos/huggingface/datasets/issues/2659/labels{/name}
2021-07-16T10:25:38Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2659.diff", "html_url": "https://github.com/huggingface/datasets/pull/2659", "merged_at": "2021-07-16T12:46:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/2659.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2659" }
946,155,407
[]
https://api.github.com/repos/huggingface/datasets/issues/2659
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Close https://github.com/huggingface/datasets/issues/2658 The dataset config kwargs that were set to None we simply ignored. This was an issue when None has some meaning for certain parameters of certain builders, like the `sep` parameter of the "csv" builder that allows to infer to separator. cc @SBrandeis
2021-07-16T12:46:07Z
https://github.com/huggingface/datasets/pull/2659
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2659/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2658/comments
https://api.github.com/repos/huggingface/datasets/issues/2658/timeline
2021-07-16T12:46:06Z
null
completed
MDU6SXNzdWU5NDYxMzk1MzI=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
2,658
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Can't pass `sep=None` to load_dataset("csv", ...) to infer the separator via pandas.read_csv
https://api.github.com/repos/huggingface/datasets/issues/2658/events
null
https://api.github.com/repos/huggingface/datasets/issues/2658/labels{/name}
2021-07-16T10:05:44Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
946,139,532
[]
https://api.github.com/repos/huggingface/datasets/issues/2658
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
When doing `load_dataset("csv", sep=None)`, the `sep` passed to `pd.read_csv` is still the default `sep=","` instead, which makes it impossible to make the csv loader infer the separator. Related to https://github.com/huggingface/datasets/pull/2656 cc @SBrandeis
2021-07-16T12:46:06Z
https://github.com/huggingface/datasets/issues/2658
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2658/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2657/comments
https://api.github.com/repos/huggingface/datasets/issues/2657/timeline
null
null
null
MDU6SXNzdWU5NDU4MjI4Mjk=
open
[]
null
2,657
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stas00", "id": 10676103, "login": "stas00", "node_id": "MDQ6VXNlcjEwNjc2MTAz", "organizations_url": "https://api.github.com/users/stas00/orgs", "received_events_url": "https://api.github.com/users/stas00/received_events", "repos_url": "https://api.github.com/users/stas00/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "type": "User", "url": "https://api.github.com/users/stas00" }
`to_json` reporting enhancements
https://api.github.com/repos/huggingface/datasets/issues/2657/events
null
https://api.github.com/repos/huggingface/datasets/issues/2657/labels{/name}
2021-07-15T23:32:18Z
null
false
null
null
945,822,829
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2657
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
While using `to_json` 2 things came to mind that would have made the experience easier on the user: 1. Could we have a `desc` arg for the tqdm use and a fallback to just `to_json` so that it'd be clear to the user what's happening? Surely, one can just print the description before calling json, but I thought perhaps it'd help to have it self-identify like you did for other progress bars recently. 2. It took me a while to make sense of the reported numbers: ``` 22%|██▏ | 1536/7076 [12:30:57<44:09:42, 28.70s/it] ``` So iteration here happens to be 10K samples, and the total is 70M records. But the user does't know that, so the progress bar is perfect, but the numbers it reports are meaningless until one discovers that 1it=10K samples. And one still has to convert these in the head - so it's not quick. Not exactly sure what's the best way to approach this, perhaps it can be part of `desc`? or report M or K, so it'd be built-in if it were to print, e.g.: ``` 22%|██▏ | 15360K/70760K [12:30:57<44:09:42, 28.70s/it] ``` or ``` 22%|██▏ | 15.36M/70.76M [12:30:57<44:09:42, 28.70s/it] ``` (while of course remaining friendly to small datasets) I forget if tqdm lets you add a magnitude identifier to the running count. Thank you!
2021-07-15T23:33:53Z
https://github.com/huggingface/datasets/issues/2657
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2657/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2656/comments
https://api.github.com/repos/huggingface/datasets/issues/2656/timeline
2021-07-16T10:23:26Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkwNzUzNjA3
closed
[]
false
2,656
{ "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "events_url": "https://api.github.com/users/SBrandeis/events{/privacy}", "followers_url": "https://api.github.com/users/SBrandeis/followers", "following_url": "https://api.github.com/users/SBrandeis/following{/other_user}", "gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SBrandeis", "id": 33657802, "login": "SBrandeis", "node_id": "MDQ6VXNlcjMzNjU3ODAy", "organizations_url": "https://api.github.com/users/SBrandeis/orgs", "received_events_url": "https://api.github.com/users/SBrandeis/received_events", "repos_url": "https://api.github.com/users/SBrandeis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions", "type": "User", "url": "https://api.github.com/users/SBrandeis" }
Change `from_csv` default arguments
https://api.github.com/repos/huggingface/datasets/issues/2656/events
null
https://api.github.com/repos/huggingface/datasets/issues/2656/labels{/name}
2021-07-15T14:09:06Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2656.diff", "html_url": "https://github.com/huggingface/datasets/pull/2656", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2656.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2656" }
945,421,790
[]
https://api.github.com/repos/huggingface/datasets/issues/2656
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Passing `sep=None` to pandas's `read_csv` lets pandas guess the CSV file's separator This PR allows users to use this pandas's feature by passing `sep=None` to `Dataset.from_csv`: ```python Dataset.from_csv( ..., sep=None ) ```
2023-09-24T09:56:44Z
https://github.com/huggingface/datasets/pull/2656
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2656/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2655/comments
https://api.github.com/repos/huggingface/datasets/issues/2655/timeline
2024-01-09T07:46:28Z
null
completed
MDU6SXNzdWU5NDUzODI3MjM=
closed
[]
null
2,655
{ "avatar_url": "https://avatars.githubusercontent.com/u/8976546?v=4", "events_url": "https://api.github.com/users/Dref360/events{/privacy}", "followers_url": "https://api.github.com/users/Dref360/followers", "following_url": "https://api.github.com/users/Dref360/following{/other_user}", "gists_url": "https://api.github.com/users/Dref360/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Dref360", "id": 8976546, "login": "Dref360", "node_id": "MDQ6VXNlcjg5NzY1NDY=", "organizations_url": "https://api.github.com/users/Dref360/orgs", "received_events_url": "https://api.github.com/users/Dref360/received_events", "repos_url": "https://api.github.com/users/Dref360/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Dref360/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Dref360/subscriptions", "type": "User", "url": "https://api.github.com/users/Dref360" }
Allow the selection of multiple columns at once
https://api.github.com/repos/huggingface/datasets/issues/2655/events
null
https://api.github.com/repos/huggingface/datasets/issues/2655/labels{/name}
2021-07-15T13:30:45Z
null
false
null
null
945,382,723
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2655
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
**Is your feature request related to a problem? Please describe.** Similar to pandas, it would be great if we could select multiple columns at once. **Describe the solution you'd like** ```python my_dataset = ... # Has columns ['idx', 'sentence', 'label'] idx, label = my_dataset[['idx', 'label']] ``` **Describe alternatives you've considered** we can do `[dataset[col] for col in ('idx', 'label')]` **Additional context** This is of course very minor.
2024-01-09T15:11:27Z
https://github.com/huggingface/datasets/issues/2655
{ "+1": 8, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 8, "url": "https://api.github.com/repos/huggingface/datasets/issues/2655/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2654/comments
https://api.github.com/repos/huggingface/datasets/issues/2654/timeline
null
null
null
MDU6SXNzdWU5NDUxNjcyMzE=
open
[ { "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" } ]
null
2,654
{ "avatar_url": "https://avatars.githubusercontent.com/u/32632186?v=4", "events_url": "https://api.github.com/users/philschmid/events{/privacy}", "followers_url": "https://api.github.com/users/philschmid/followers", "following_url": "https://api.github.com/users/philschmid/following{/other_user}", "gists_url": "https://api.github.com/users/philschmid/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/philschmid", "id": 32632186, "login": "philschmid", "node_id": "MDQ6VXNlcjMyNjMyMTg2", "organizations_url": "https://api.github.com/users/philschmid/orgs", "received_events_url": "https://api.github.com/users/philschmid/received_events", "repos_url": "https://api.github.com/users/philschmid/repos", "site_admin": false, "starred_url": "https://api.github.com/users/philschmid/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/philschmid/subscriptions", "type": "User", "url": "https://api.github.com/users/philschmid" }
Give a user feedback if the dataset he loads is streamable or not
https://api.github.com/repos/huggingface/datasets/issues/2654/events
null
https://api.github.com/repos/huggingface/datasets/issues/2654/labels{/name}
2021-07-15T09:07:27Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
null
945,167,231
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2654
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
**Is your feature request related to a problem? Please describe.** I would love to know if a `dataset` is with the current implementation streamable or not. **Describe the solution you'd like** We could show a warning when a dataset is loaded with `load_dataset('...',streaming=True)` when its lot streamable, e.g. if it is an archive. **Describe alternatives you've considered** Add a new metadata tag for "streaming"
2021-08-02T11:03:21Z
https://github.com/huggingface/datasets/issues/2654
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2654/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2653/comments
https://api.github.com/repos/huggingface/datasets/issues/2653/timeline
2021-08-04T17:03:52Z
null
completed
MDU6SXNzdWU5NDUxMDIzMjE=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,653
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Add SD task for SUPERB
https://api.github.com/repos/huggingface/datasets/issues/2653/events
{ "closed_at": "2021-09-02T05:34:03Z", "closed_issues": 2, "created_at": "2021-07-09T05:49:00Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-07-30T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/7", "id": 6931350, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/7/labels", "node_id": "MDk6TWlsZXN0b25lNjkzMTM1MA==", "number": 7, "open_issues": 0, "state": "closed", "title": "1.11", "updated_at": "2021-09-02T05:34:03Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/7" }
https://api.github.com/repos/huggingface/datasets/issues/2653/labels{/name}
2021-07-15T07:51:40Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
945,102,321
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
https://api.github.com/repos/huggingface/datasets/issues/2653
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Include the SD (Speaker Diarization) task as described in the [SUPERB paper](https://arxiv.org/abs/2105.01051) and `s3prl` [instructions](https://github.com/s3prl/s3prl/tree/master/s3prl/downstream#sd-speaker-diarization). Steps: - [x] Generate the LibriMix corpus - [x] Prepare the corpus for diarization - [x] Upload these files to the superb-data repo - [x] Transcribe the corresponding s3prl processing of these files into our superb loading script - [ ] README: tags + description sections Related to #2619. cc: @lewtun
2021-08-04T17:03:52Z
https://github.com/huggingface/datasets/issues/2653
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2653/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2652/comments
https://api.github.com/repos/huggingface/datasets/issues/2652/timeline
2021-07-15T09:57:31Z
null
null
MDExOlB1bGxSZXF1ZXN0NjkwMjg0MTI4
closed
[]
false
2,652
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Fix logging docstring
https://api.github.com/repos/huggingface/datasets/issues/2652/events
null
https://api.github.com/repos/huggingface/datasets/issues/2652/labels{/name}
2021-07-14T23:19:58Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2652.diff", "html_url": "https://github.com/huggingface/datasets/pull/2652", "merged_at": "2021-07-15T09:57:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/2652.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2652" }
944,865,924
[]
https://api.github.com/repos/huggingface/datasets/issues/2652
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Remove "no tqdm bars" from the docstring in the logging module to align it with the changes introduced in #2534.
2021-07-18T11:41:06Z
https://github.com/huggingface/datasets/pull/2652
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2652/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2651/comments
https://api.github.com/repos/huggingface/datasets/issues/2651/timeline
2021-07-15T03:41:35Z
null
completed
MDU6SXNzdWU5NDQ3OTY5NjE=
closed
[]
null
2,651
{ "avatar_url": "https://avatars.githubusercontent.com/u/1147443?v=4", "events_url": "https://api.github.com/users/Isa-rentacs/events{/privacy}", "followers_url": "https://api.github.com/users/Isa-rentacs/followers", "following_url": "https://api.github.com/users/Isa-rentacs/following{/other_user}", "gists_url": "https://api.github.com/users/Isa-rentacs/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Isa-rentacs", "id": 1147443, "login": "Isa-rentacs", "node_id": "MDQ6VXNlcjExNDc0NDM=", "organizations_url": "https://api.github.com/users/Isa-rentacs/orgs", "received_events_url": "https://api.github.com/users/Isa-rentacs/received_events", "repos_url": "https://api.github.com/users/Isa-rentacs/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Isa-rentacs/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Isa-rentacs/subscriptions", "type": "User", "url": "https://api.github.com/users/Isa-rentacs" }
Setting log level higher than warning does not suppress progress bar
https://api.github.com/repos/huggingface/datasets/issues/2651/events
null
https://api.github.com/repos/huggingface/datasets/issues/2651/labels{/name}
2021-07-14T21:06:51Z
null
false
null
null
944,796,961
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2651
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug I would like to disable progress bars for `.map` method (and other methods like `.filter` and `load_dataset` as well). According to #1627 one can suppress it by setting log level higher than `warning`, however doing so doesn't suppress it with version 1.9.0. I also tried to set `DATASETS_VERBOSITY` environment variable to `error` or `critical` but it also didn't work. ## Steps to reproduce the bug ```python import datasets from datasets.utils.logging import set_verbosity_error set_verbosity_error() def dummy_map(batch): return batch common_voice_train = datasets.load_dataset("common_voice", "de", split="train") common_voice_test = datasets.load_dataset("common_voice", "de", split="test") common_voice_train.map(dummy_map) ``` ## Expected results - The progress bar for `.map` call won't be shown ## Actual results - The progress bar for `.map` is still shown ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.9.0 - Platform: Linux-5.4.0-1045-aws-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.5 - PyArrow version: 4.0.1
2022-07-08T14:51:57Z
https://github.com/huggingface/datasets/issues/2651
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2651/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2650/comments
https://api.github.com/repos/huggingface/datasets/issues/2650/timeline
2023-11-28T19:11:40Z
null
completed
MDU6SXNzdWU5NDQ2NzI1NjU=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "events_url": "https://api.github.com/users/TevenLeScao/events{/privacy}", "followers_url": "https://api.github.com/users/TevenLeScao/followers", "following_url": "https://api.github.com/users/TevenLeScao/following{/other_user}", "gists_url": "https://api.github.com/users/TevenLeScao/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/TevenLeScao", "id": 26709476, "login": "TevenLeScao", "node_id": "MDQ6VXNlcjI2NzA5NDc2", "organizations_url": "https://api.github.com/users/TevenLeScao/orgs", "received_events_url": "https://api.github.com/users/TevenLeScao/received_events", "repos_url": "https://api.github.com/users/TevenLeScao/repos", "site_admin": false, "starred_url": "https://api.github.com/users/TevenLeScao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TevenLeScao/subscriptions", "type": "User", "url": "https://api.github.com/users/TevenLeScao" } ]
null
2,650
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stas00", "id": 10676103, "login": "stas00", "node_id": "MDQ6VXNlcjEwNjc2MTAz", "organizations_url": "https://api.github.com/users/stas00/orgs", "received_events_url": "https://api.github.com/users/stas00/received_events", "repos_url": "https://api.github.com/users/stas00/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "type": "User", "url": "https://api.github.com/users/stas00" }
[load_dataset] shard and parallelize the process
https://api.github.com/repos/huggingface/datasets/issues/2650/events
null
https://api.github.com/repos/huggingface/datasets/issues/2650/labels{/name}
2021-07-14T18:04:58Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "events_url": "https://api.github.com/users/TevenLeScao/events{/privacy}", "followers_url": "https://api.github.com/users/TevenLeScao/followers", "following_url": "https://api.github.com/users/TevenLeScao/following{/other_user}", "gists_url": "https://api.github.com/users/TevenLeScao/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/TevenLeScao", "id": 26709476, "login": "TevenLeScao", "node_id": "MDQ6VXNlcjI2NzA5NDc2", "organizations_url": "https://api.github.com/users/TevenLeScao/orgs", "received_events_url": "https://api.github.com/users/TevenLeScao/received_events", "repos_url": "https://api.github.com/users/TevenLeScao/repos", "site_admin": false, "starred_url": "https://api.github.com/users/TevenLeScao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TevenLeScao/subscriptions", "type": "User", "url": "https://api.github.com/users/TevenLeScao" }
null
944,672,565
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2650
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
- Some huge datasets take forever to build the first time. (e.g. oscar/en) as it's done in a single cpu core. - If the build crashes, everything done up to that point gets lost Request: Shard the build over multiple arrow files, which would enable: - much faster build by parallelizing the build process - if the process crashed, the completed arrow files don't need to be re-built again Thank you! @lhoestq
2023-11-28T19:11:41Z
https://github.com/huggingface/datasets/issues/2650
{ "+1": 5, "-1": 0, "confused": 0, "eyes": 3, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 2, "total_count": 10, "url": "https://api.github.com/repos/huggingface/datasets/issues/2650/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2649/comments
https://api.github.com/repos/huggingface/datasets/issues/2649/timeline
null
null
null
MDU6SXNzdWU5NDQ2NTEyMjk=
open
[]
null
2,649
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stas00", "id": 10676103, "login": "stas00", "node_id": "MDQ6VXNlcjEwNjc2MTAz", "organizations_url": "https://api.github.com/users/stas00/orgs", "received_events_url": "https://api.github.com/users/stas00/received_events", "repos_url": "https://api.github.com/users/stas00/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "type": "User", "url": "https://api.github.com/users/stas00" }
adding progress bar / ETA for `load_dataset`
https://api.github.com/repos/huggingface/datasets/issues/2649/events
null
https://api.github.com/repos/huggingface/datasets/issues/2649/labels{/name}
2021-07-14T17:34:39Z
null
false
null
null
944,651,229
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2649
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Please consider: ``` Downloading and preparing dataset oscar/unshuffled_deduplicated_en (download: 462.40 GiB, generated: 1.18 TiB, post-processed: Unknown size, total: 1.63 TiB) to cache/oscar/unshuffled_deduplicated_en/1.0.0/84838bd49d2295f62008383b05620571535451d84545037bb94d6f3501651df2... HF google storage unreachable. Downloading and preparing it from source ``` and no indication whatsoever of whether things work well or when it'll be done. It's important to have an estimated completion time for when doing slurm jobs since some instances have a cap on run-time. I think for this particular job it sat for 30min in total silence and then after 30min it started generating: ``` 897850 examples [07:24, 10286.71 examples/s] ``` which is already great! Request: 1. ETA - knowing how many hours to allocate for a slurm job 2. progress bar - helps to know things are working and aren't stuck and where we are at. Thank you! @lhoestq
2023-03-27T10:32:49Z
https://github.com/huggingface/datasets/issues/2649
{ "+1": 5, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 5, "url": "https://api.github.com/repos/huggingface/datasets/issues/2649/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2648/comments
https://api.github.com/repos/huggingface/datasets/issues/2648/timeline
null
null
null
MDU6SXNzdWU5NDQ0ODQ1MjI=
open
[ { "avatar_url": "https://avatars.githubusercontent.com/u/26653468?v=4", "events_url": "https://api.github.com/users/bhadreshpsavani/events{/privacy}", "followers_url": "https://api.github.com/users/bhadreshpsavani/followers", "following_url": "https://api.github.com/users/bhadreshpsavani/following{/other_user}", "gists_url": "https://api.github.com/users/bhadreshpsavani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bhadreshpsavani", "id": 26653468, "login": "bhadreshpsavani", "node_id": "MDQ6VXNlcjI2NjUzNDY4", "organizations_url": "https://api.github.com/users/bhadreshpsavani/orgs", "received_events_url": "https://api.github.com/users/bhadreshpsavani/received_events", "repos_url": "https://api.github.com/users/bhadreshpsavani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bhadreshpsavani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhadreshpsavani/subscriptions", "type": "User", "url": "https://api.github.com/users/bhadreshpsavani" } ]
null
2,648
{ "avatar_url": "https://avatars.githubusercontent.com/u/26653468?v=4", "events_url": "https://api.github.com/users/bhadreshpsavani/events{/privacy}", "followers_url": "https://api.github.com/users/bhadreshpsavani/followers", "following_url": "https://api.github.com/users/bhadreshpsavani/following{/other_user}", "gists_url": "https://api.github.com/users/bhadreshpsavani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bhadreshpsavani", "id": 26653468, "login": "bhadreshpsavani", "node_id": "MDQ6VXNlcjI2NjUzNDY4", "organizations_url": "https://api.github.com/users/bhadreshpsavani/orgs", "received_events_url": "https://api.github.com/users/bhadreshpsavani/received_events", "repos_url": "https://api.github.com/users/bhadreshpsavani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bhadreshpsavani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhadreshpsavani/subscriptions", "type": "User", "url": "https://api.github.com/users/bhadreshpsavani" }
Add web_split dataset for Paraphase and Rephrase benchmark
https://api.github.com/repos/huggingface/datasets/issues/2648/events
null
https://api.github.com/repos/huggingface/datasets/issues/2648/labels{/name}
2021-07-14T14:24:36Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/26653468?v=4", "events_url": "https://api.github.com/users/bhadreshpsavani/events{/privacy}", "followers_url": "https://api.github.com/users/bhadreshpsavani/followers", "following_url": "https://api.github.com/users/bhadreshpsavani/following{/other_user}", "gists_url": "https://api.github.com/users/bhadreshpsavani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bhadreshpsavani", "id": 26653468, "login": "bhadreshpsavani", "node_id": "MDQ6VXNlcjI2NjUzNDY4", "organizations_url": "https://api.github.com/users/bhadreshpsavani/orgs", "received_events_url": "https://api.github.com/users/bhadreshpsavani/received_events", "repos_url": "https://api.github.com/users/bhadreshpsavani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bhadreshpsavani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhadreshpsavani/subscriptions", "type": "User", "url": "https://api.github.com/users/bhadreshpsavani" }
null
944,484,522
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2648
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe: For getting simple sentences from complex sentence there are dataset and task like wiki_split that is available in hugging face datasets. This web_split is a very similar dataset. There some research paper which states that by combining these two datasets we if we train the model it will yield better results on both tests data. This dataset is made from web NLG data. All the dataset related details are provided in the below repository Github link: https://github.com/shashiongithub/Split-and-Rephrase
2021-07-14T14:26:12Z
https://github.com/huggingface/datasets/issues/2648
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2648/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2647/comments
https://api.github.com/repos/huggingface/datasets/issues/2647/timeline
2021-07-15T06:50:47Z
null
null
MDExOlB1bGxSZXF1ZXN0Njg5OTExMzky
closed
[]
false
2,647
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Fix anchor in README
https://api.github.com/repos/huggingface/datasets/issues/2647/events
{ "closed_at": "2021-07-21T15:36:49Z", "closed_issues": 29, "created_at": "2021-06-08T18:48:33Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-08-05T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/6", "id": 6836458, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/6/labels", "node_id": "MDk6TWlsZXN0b25lNjgzNjQ1OA==", "number": 6, "open_issues": 0, "state": "closed", "title": "1.10", "updated_at": "2021-07-21T15:36:49Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/6" }
https://api.github.com/repos/huggingface/datasets/issues/2647/labels{/name}
2021-07-14T13:22:44Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2647.diff", "html_url": "https://github.com/huggingface/datasets/pull/2647", "merged_at": "2021-07-15T06:50:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/2647.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2647" }
944,424,941
[]
https://api.github.com/repos/huggingface/datasets/issues/2647
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
I forgot to push this fix in #2611, so I'm sending it now.
2021-07-18T11:41:18Z
https://github.com/huggingface/datasets/pull/2647
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2647/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2646/comments
https://api.github.com/repos/huggingface/datasets/issues/2646/timeline
2022-08-04T08:28:24Z
null
completed
MDU6SXNzdWU5NDQzNzk5NTQ=
closed
[]
null
2,646
{ "avatar_url": "https://avatars.githubusercontent.com/u/66781249?v=4", "events_url": "https://api.github.com/users/vikrant7k/events{/privacy}", "followers_url": "https://api.github.com/users/vikrant7k/followers", "following_url": "https://api.github.com/users/vikrant7k/following{/other_user}", "gists_url": "https://api.github.com/users/vikrant7k/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/vikrant7k", "id": 66781249, "login": "vikrant7k", "node_id": "MDQ6VXNlcjY2NzgxMjQ5", "organizations_url": "https://api.github.com/users/vikrant7k/orgs", "received_events_url": "https://api.github.com/users/vikrant7k/received_events", "repos_url": "https://api.github.com/users/vikrant7k/repos", "site_admin": false, "starred_url": "https://api.github.com/users/vikrant7k/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vikrant7k/subscriptions", "type": "User", "url": "https://api.github.com/users/vikrant7k" }
downloading of yahoo_answers_topics dataset failed
https://api.github.com/repos/huggingface/datasets/issues/2646/events
null
https://api.github.com/repos/huggingface/datasets/issues/2646/labels{/name}
2021-07-14T12:31:05Z
null
false
null
null
944,379,954
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2646
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug I get an error datasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files when I try to download the yahoo_answers_topics dataset ## Steps to reproduce the bug self.dataset = load_dataset( 'yahoo_answers_topics', cache_dir=self.config['yahoo_cache_dir'], split='train[:90%]') # Sample code to reproduce the bug self.dataset = load_dataset( 'yahoo_answers_topics', cache_dir=self.config['yahoo_cache_dir'], split='train[:90%]') ## Expected results A clear and concise description of the expected results. ## Actual results Specify the actual results or traceback. datasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files
2022-08-04T08:28:24Z
https://github.com/huggingface/datasets/issues/2646
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2646/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2645/comments
https://api.github.com/repos/huggingface/datasets/issues/2645/timeline
2021-07-15T09:34:02Z
null
completed
MDU6SXNzdWU5NDQzNzQyODQ=
closed
[]
null
2,645
{ "avatar_url": "https://avatars.githubusercontent.com/u/40395156?v=4", "events_url": "https://api.github.com/users/fake-warrior8/events{/privacy}", "followers_url": "https://api.github.com/users/fake-warrior8/followers", "following_url": "https://api.github.com/users/fake-warrior8/following{/other_user}", "gists_url": "https://api.github.com/users/fake-warrior8/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/fake-warrior8", "id": 40395156, "login": "fake-warrior8", "node_id": "MDQ6VXNlcjQwMzk1MTU2", "organizations_url": "https://api.github.com/users/fake-warrior8/orgs", "received_events_url": "https://api.github.com/users/fake-warrior8/received_events", "repos_url": "https://api.github.com/users/fake-warrior8/repos", "site_admin": false, "starred_url": "https://api.github.com/users/fake-warrior8/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fake-warrior8/subscriptions", "type": "User", "url": "https://api.github.com/users/fake-warrior8" }
load_dataset processing failed with OS error after downloading a dataset
https://api.github.com/repos/huggingface/datasets/issues/2645/events
null
https://api.github.com/repos/huggingface/datasets/issues/2645/labels{/name}
2021-07-14T12:23:53Z
null
false
null
null
944,374,284
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2645
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug After downloading a dataset like opus100, there is a bug that OSError: Cannot find data file. Original error: dlopen: cannot load any more object with static TLS ## Steps to reproduce the bug ```python from datasets import load_dataset this_dataset = load_dataset('opus100', 'af-en') ``` ## Expected results there is no error when running load_dataset. ## Actual results Specify the actual results or traceback. Traceback (most recent call last): File "/home/anaconda3/lib/python3.6/site-packages/datasets/builder.py", line 652, in _download_and_prep self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/anaconda3/lib/python3.6/site-packages/datasets/builder.py", line 989, in _prepare_split example = self.info.features.encode_example(record) File "/home/anaconda3/lib/python3.6/site-packages/datasets/features.py", line 952, in encode_example example = cast_to_python_objects(example) File "/home/anaconda3/lib/python3.6/site-packages/datasets/features.py", line 219, in cast_to_python_ob return _cast_to_python_objects(obj)[0] File "/home/anaconda3/lib/python3.6/site-packages/datasets/features.py", line 165, in _cast_to_python_o import torch File "/home/anaconda3/lib/python3.6/site-packages/torch/__init__.py", line 188, in <module> _load_global_deps() File "/home/anaconda3/lib/python3.6/site-packages/torch/__init__.py", line 141, in _load_global_deps ctypes.CDLL(lib_path, mode=ctypes.RTLD_GLOBAL) File "/home/anaconda3/lib/python3.6/ctypes/__init__.py", line 348, in __init__ self._handle = _dlopen(self._name, mode) OSError: dlopen: cannot load any more object with static TLS During handling of the above exception, another exception occurred: Traceback (most recent call last): File "download_hub_opus100.py", line 9, in <module> this_dataset = load_dataset('opus100', language_pair) File "/home/anaconda3/lib/python3.6/site-packages/datasets/load.py", line 748, in load_dataset use_auth_token=use_auth_token, File "/home/anaconda3/lib/python3.6/site-packages/datasets/builder.py", line 575, in download_and_prepa dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "/home/anaconda3/lib/python3.6/site-packages/datasets/builder.py", line 658, in _download_and_prep + str(e) OSError: Cannot find data file. Original error: dlopen: cannot load any more object with static TLS ## Environment info - `datasets` version: 1.8.0 - Platform: Linux-3.13.0-32-generic-x86_64-with-debian-jessie-sid - Python version: 3.6.6 - PyArrow version: 3.0.0
2021-07-15T09:34:02Z
https://github.com/huggingface/datasets/issues/2645
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2645/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2644/comments
https://api.github.com/repos/huggingface/datasets/issues/2644/timeline
2021-07-26T14:55:15Z
null
completed
MDU6SXNzdWU5NDQyNTQ3NDg=
closed
[]
null
2,644
{ "avatar_url": "https://avatars.githubusercontent.com/u/1177582?v=4", "events_url": "https://api.github.com/users/pcuenca/events{/privacy}", "followers_url": "https://api.github.com/users/pcuenca/followers", "following_url": "https://api.github.com/users/pcuenca/following{/other_user}", "gists_url": "https://api.github.com/users/pcuenca/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/pcuenca", "id": 1177582, "login": "pcuenca", "node_id": "MDQ6VXNlcjExNzc1ODI=", "organizations_url": "https://api.github.com/users/pcuenca/orgs", "received_events_url": "https://api.github.com/users/pcuenca/received_events", "repos_url": "https://api.github.com/users/pcuenca/repos", "site_admin": false, "starred_url": "https://api.github.com/users/pcuenca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pcuenca/subscriptions", "type": "User", "url": "https://api.github.com/users/pcuenca" }
Batched `map` not allowed to return 0 items
https://api.github.com/repos/huggingface/datasets/issues/2644/events
null
https://api.github.com/repos/huggingface/datasets/issues/2644/labels{/name}
2021-07-14T09:58:19Z
null
false
null
null
944,254,748
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2644
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Describe the bug I'm trying to use `map` to filter a large dataset by selecting rows that match an expensive condition (files referenced by one of the columns need to exist in the filesystem, so we have to `stat` them). According to [the documentation](https://huggingface.co/docs/datasets/processing.html#augmenting-the-dataset), `a batch mapped function can take as input a batch of size N and return a batch of size M where M can be greater or less than N and can even be zero`. However, when the returned batch has a size of zero (neither item in the batch fulfilled the condition), we get an `index out of bounds` error. I think that `arrow_writer.py` is [trying to infer the returned types using the first element returned](https://github.com/huggingface/datasets/blob/master/src/datasets/arrow_writer.py#L100), but no elements were returned in this case. For this error to happen, I'm returning a dictionary that contains empty lists for the keys I want to keep, see below. If I return an empty dictionary instead (no keys), then a different error eventually occurs. ## Steps to reproduce the bug ```python def select_rows(examples): # `key` is a column name that exists in the original dataset # The following line simulates no matches found, so we return an empty batch result = {'key': []} return result filtered_dataset = dataset.map( select_rows, remove_columns = dataset.column_names, batched = True, num_proc = 1, desc = "Selecting rows with images that exist" ) ``` The code above immediately triggers the exception. If we use the following instead: ```python def select_rows(examples): # `key` is a column name that exists in the original dataset result = {'key': []} # or defaultdict or whatever # code to check for condition and append elements to result # some_items_found will be set to True if there were any matching elements in the batch return result if some_items_found else {} ``` Then it _seems_ to work, but it eventually fails with some sort of schema error. I believe it may happen when an empty batch is followed by a non-empty one, but haven't set up a test to verify it. In my opinion, returning a dictionary with empty lists and valid column names should be accepted as a valid result with zero items. ## Expected results The dataset would be filtered and only the matching fields would be returned. ## Actual results An exception is encountered, as described. Using a workaround makes it fail further along the line. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.9.1.dev0 - Platform: Linux-5.4.0-53-generic-x86_64-with-glibc2.17 - Python version: 3.8.10 - PyArrow version: 4.0.1
2021-07-26T14:55:15Z
https://github.com/huggingface/datasets/issues/2644
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2644/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2643/comments
https://api.github.com/repos/huggingface/datasets/issues/2643/timeline
null
null
null
MDU6SXNzdWU5NDQyMjAyNzM=
open
[]
null
2,643
{ "avatar_url": "https://avatars.githubusercontent.com/u/100702?v=4", "events_url": "https://api.github.com/users/jorgeecardona/events{/privacy}", "followers_url": "https://api.github.com/users/jorgeecardona/followers", "following_url": "https://api.github.com/users/jorgeecardona/following{/other_user}", "gists_url": "https://api.github.com/users/jorgeecardona/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jorgeecardona", "id": 100702, "login": "jorgeecardona", "node_id": "MDQ6VXNlcjEwMDcwMg==", "organizations_url": "https://api.github.com/users/jorgeecardona/orgs", "received_events_url": "https://api.github.com/users/jorgeecardona/received_events", "repos_url": "https://api.github.com/users/jorgeecardona/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jorgeecardona/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jorgeecardona/subscriptions", "type": "User", "url": "https://api.github.com/users/jorgeecardona" }
Enum used in map functions will raise a RecursionError with dill.
https://api.github.com/repos/huggingface/datasets/issues/2643/events
null
https://api.github.com/repos/huggingface/datasets/issues/2643/labels{/name}
2021-07-14T09:16:08Z
null
false
null
null
944,220,273
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2643
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Enums used in functions pass to `map` will fail at pickling with a maximum recursion exception as described here: https://github.com/uqfoundation/dill/issues/250#issuecomment-852566284 In my particular case, I use an enum to define an argument with fixed options using the `TraininigArguments` dataclass as base class and the `HfArgumentParser`. In the same file I use a `ds.map` that tries to pickle the content of the module including the definition of the enum that runs into the dill bug described above. ## Steps to reproduce the bug ```python from datasets import load_dataset from enum import Enum class A(Enum): a = 'a' def main(): a = A.a def f(x): return {} if a == a.a else x ds = load_dataset('cnn_dailymail', '3.0.0')['test'] ds = ds.map(f, num_proc=15) if __name__ == "__main__": main() ``` ## Expected results The known problem with dill could be prevented as explained in the link above (workaround.) Since `HFArgumentParser` nicely uses the enum class for choices it makes sense to also deal with this bug under the hood. ## Actual results ```python File "/home/xxxx/miniconda3/lib/python3.8/site-packages/dill/_dill.py", line 1373, in save_type pickler.save_reduce(_create_type, (type(obj), obj.__name__, File "/home/xxxx/miniconda3/lib/python3.8/pickle.py", line 690, in save_reduce save(args) File "/home/xxxx/miniconda3/lib/python3.8/pickle.py", line 558, in save f(self, obj) # Call unbound method with explicit self File "/home/xxxx/miniconda3/lib/python3.8/pickle.py", line 899, in save_tuple save(element) File "/home/xxxx/miniconda3/lib/python3.8/pickle.py", line 534, in save self.framer.commit_frame() File "/home/xxxx/miniconda3/lib/python3.8/pickle.py", line 220, in commit_frame if f.tell() >= self._FRAME_SIZE_TARGET or force: RecursionError: maximum recursion depth exceeded while calling a Python object ``` ## Environment info - `datasets` version: 1.8.0 - Platform: Linux-5.9.0-4-amd64-x86_64-with-glibc2.10 - Python version: 3.8.5 - PyArrow version: 3.0.0
2021-11-02T09:51:11Z
https://github.com/huggingface/datasets/issues/2643
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2643/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2642/comments
https://api.github.com/repos/huggingface/datasets/issues/2642/timeline
null
null
null
MDU6SXNzdWU5NDQxNzU2OTc=
open
[]
null
2,642
{ "avatar_url": "https://avatars.githubusercontent.com/u/31893406?v=4", "events_url": "https://api.github.com/users/cccntu/events{/privacy}", "followers_url": "https://api.github.com/users/cccntu/followers", "following_url": "https://api.github.com/users/cccntu/following{/other_user}", "gists_url": "https://api.github.com/users/cccntu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cccntu", "id": 31893406, "login": "cccntu", "node_id": "MDQ6VXNlcjMxODkzNDA2", "organizations_url": "https://api.github.com/users/cccntu/orgs", "received_events_url": "https://api.github.com/users/cccntu/received_events", "repos_url": "https://api.github.com/users/cccntu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cccntu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cccntu/subscriptions", "type": "User", "url": "https://api.github.com/users/cccntu" }
Support multi-worker with streaming dataset (IterableDataset).
https://api.github.com/repos/huggingface/datasets/issues/2642/events
null
https://api.github.com/repos/huggingface/datasets/issues/2642/labels{/name}
2021-07-14T08:22:58Z
null
false
null
null
944,175,697
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2642
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
**Is your feature request related to a problem? Please describe.** The current `.map` does not support multi-process, CPU can become bottleneck if the pre-processing is complex (e.g. t5 span masking). **Describe the solution you'd like** Ideally `.map` should support multi-worker like tfds, with `AUTOTUNE`. **Describe alternatives you've considered** A simpler solution is to shard the dataset and process it in parallel with pytorch dataloader. The shard does not need to be of equal size. * https://pytorch.org/docs/stable/data.html#torch.utils.data.IterableDataset **Additional context**
2021-07-15T09:37:34Z
https://github.com/huggingface/datasets/issues/2642
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2642/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2641/comments
https://api.github.com/repos/huggingface/datasets/issues/2641/timeline
2022-08-04T08:30:08Z
null
completed
MDU6SXNzdWU5NDM4MzgwODU=
closed
[]
null
2,641
{ "avatar_url": "https://avatars.githubusercontent.com/u/13956255?v=4", "events_url": "https://api.github.com/users/courtmckay/events{/privacy}", "followers_url": "https://api.github.com/users/courtmckay/followers", "following_url": "https://api.github.com/users/courtmckay/following{/other_user}", "gists_url": "https://api.github.com/users/courtmckay/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/courtmckay", "id": 13956255, "login": "courtmckay", "node_id": "MDQ6VXNlcjEzOTU2MjU1", "organizations_url": "https://api.github.com/users/courtmckay/orgs", "received_events_url": "https://api.github.com/users/courtmckay/received_events", "repos_url": "https://api.github.com/users/courtmckay/repos", "site_admin": false, "starred_url": "https://api.github.com/users/courtmckay/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/courtmckay/subscriptions", "type": "User", "url": "https://api.github.com/users/courtmckay" }
load_dataset("financial_phrasebank") NonMatchingChecksumError
https://api.github.com/repos/huggingface/datasets/issues/2641/events
null
https://api.github.com/repos/huggingface/datasets/issues/2641/labels{/name}
2021-07-13T21:21:49Z
null
false
null
null
943,838,085
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2641
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Attempting to download the financial_phrasebank dataset results in a NonMatchingChecksumError ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("financial_phrasebank", 'sentences_allagree') ``` ## Expected results I expect to see the financial_phrasebank dataset downloaded successfully ## Actual results NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://www.researchgate.net/profile/Pekka_Malo/publication/251231364_FinancialPhraseBank-v10/data/0c96051eee4fb1d56e000000/FinancialPhraseBank-v10.zip'] ## Environment info - `datasets` version: 1.9.0 - Platform: Linux-4.14.232-177.418.amzn2.x86_64-x86_64-with-debian-10.6 - Python version: 3.7.10 - PyArrow version: 4.0.1
2022-08-04T08:30:08Z
https://github.com/huggingface/datasets/issues/2641
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2641/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2640/comments
https://api.github.com/repos/huggingface/datasets/issues/2640/timeline
2021-07-15T06:06:12Z
null
null
MDExOlB1bGxSZXF1ZXN0Njg5MjAxMDkw
closed
[]
false
2,640
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix docstrings
https://api.github.com/repos/huggingface/datasets/issues/2640/events
{ "closed_at": "2021-07-21T15:36:49Z", "closed_issues": 29, "created_at": "2021-06-08T18:48:33Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-08-05T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/6", "id": 6836458, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/6/labels", "node_id": "MDk6TWlsZXN0b25lNjgzNjQ1OA==", "number": 6, "open_issues": 0, "state": "closed", "title": "1.10", "updated_at": "2021-07-21T15:36:49Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/6" }
https://api.github.com/repos/huggingface/datasets/issues/2640/labels{/name}
2021-07-13T16:09:14Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2640.diff", "html_url": "https://github.com/huggingface/datasets/pull/2640", "merged_at": "2021-07-15T06:06:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/2640.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2640" }
943,591,055
[]
https://api.github.com/repos/huggingface/datasets/issues/2640
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix rendering of some docstrings.
2021-07-15T06:51:01Z
https://github.com/huggingface/datasets/pull/2640
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2640/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2639/comments
https://api.github.com/repos/huggingface/datasets/issues/2639/timeline
2021-07-13T16:52:49Z
null
null
MDExOlB1bGxSZXF1ZXN0Njg5MTQ3NDE5
closed
[]
false
2,639
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Refactor patching to specific submodule
https://api.github.com/repos/huggingface/datasets/issues/2639/events
null
https://api.github.com/repos/huggingface/datasets/issues/2639/labels{/name}
2021-07-13T15:08:45Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2639.diff", "html_url": "https://github.com/huggingface/datasets/pull/2639", "merged_at": "2021-07-13T16:52:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/2639.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2639" }
943,527,463
[]
https://api.github.com/repos/huggingface/datasets/issues/2639
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Minor reorganization of the code, so that additional patching functions (not related to streaming) might be created. In relation with the initial approach followed in #2631.
2021-07-13T16:52:49Z
https://github.com/huggingface/datasets/pull/2639
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2639/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2638/comments
https://api.github.com/repos/huggingface/datasets/issues/2638/timeline
2021-07-16T15:59:31Z
null
null
MDExOlB1bGxSZXF1ZXN0Njg5MTA5NTg1
closed
[]
false
2,638
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Streaming for the Json loader
https://api.github.com/repos/huggingface/datasets/issues/2638/events
null
https://api.github.com/repos/huggingface/datasets/issues/2638/labels{/name}
2021-07-13T14:37:06Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2638.diff", "html_url": "https://github.com/huggingface/datasets/pull/2638", "merged_at": "2021-07-16T15:59:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/2638.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2638" }
943,484,913
[]
https://api.github.com/repos/huggingface/datasets/issues/2638
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
It was not using `open` in the builder. Therefore `pyarrow.json.read_json` was downloading the full file to start yielding rows. Moreover, it appeared that `pyarrow.json.read_json` was not really suited for streaming as it was downloading too much data and failing if `block_size` was not properly configured (related to #2573). So I switched to using `open` which is extended to support reading from remote file progressively, and I removed the pyarrow json reader which was not practical. Instead, I'm using the classical `json.loads` from the standard library.
2021-07-16T15:59:32Z
https://github.com/huggingface/datasets/pull/2638
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2638/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2636/comments
https://api.github.com/repos/huggingface/datasets/issues/2636/timeline
2021-07-13T14:37:23Z
null
null
MDExOlB1bGxSZXF1ZXN0Njg4NzEyMTY4
closed
[]
false
2,636
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Streaming for the Pandas loader
https://api.github.com/repos/huggingface/datasets/issues/2636/events
null
https://api.github.com/repos/huggingface/datasets/issues/2636/labels{/name}
2021-07-13T09:18:21Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2636.diff", "html_url": "https://github.com/huggingface/datasets/pull/2636", "merged_at": "2021-07-13T14:37:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/2636.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2636" }
943,044,514
[]
https://api.github.com/repos/huggingface/datasets/issues/2636
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
It was not using open in the builder. Therefore pd.read_pickle could fail when streaming from a private repo for example. Indeed, when streaming, open is extended to support reading from remote files and handles authentication to the HF Hub
2021-07-13T14:37:24Z
https://github.com/huggingface/datasets/pull/2636
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2636/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2635/comments
https://api.github.com/repos/huggingface/datasets/issues/2635/timeline
2021-07-13T15:19:37Z
null
null
MDExOlB1bGxSZXF1ZXN0Njg4Njk5OTM5
closed
[]
false
2,635
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Streaming for the CSV loader
https://api.github.com/repos/huggingface/datasets/issues/2635/events
null
https://api.github.com/repos/huggingface/datasets/issues/2635/labels{/name}
2021-07-13T09:08:58Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2635.diff", "html_url": "https://github.com/huggingface/datasets/pull/2635", "merged_at": "2021-07-13T15:19:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/2635.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2635" }
943,030,999
[]
https://api.github.com/repos/huggingface/datasets/issues/2635
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
It was not using `open` in the builder. Therefore `pd.read_csv` was downloading the full file to start yielding rows. Indeed, when streaming, `open` is extended to support reading from remote file progressively.
2021-07-13T15:19:38Z
https://github.com/huggingface/datasets/pull/2635
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2635/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2634/comments
https://api.github.com/repos/huggingface/datasets/issues/2634/timeline
2021-07-13T09:05:09Z
null
null
MDExOlB1bGxSZXF1ZXN0Njg4NDk2Mzc2
closed
[]
false
2,634
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Inject ASR template for lj_speech dataset
https://api.github.com/repos/huggingface/datasets/issues/2634/events
{ "closed_at": "2021-07-21T15:36:49Z", "closed_issues": 29, "created_at": "2021-06-08T18:48:33Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-08-05T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/6", "id": 6836458, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/6/labels", "node_id": "MDk6TWlsZXN0b25lNjgzNjQ1OA==", "number": 6, "open_issues": 0, "state": "closed", "title": "1.10", "updated_at": "2021-07-21T15:36:49Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/6" }
https://api.github.com/repos/huggingface/datasets/issues/2634/labels{/name}
2021-07-13T06:04:54Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2634.diff", "html_url": "https://github.com/huggingface/datasets/pull/2634", "merged_at": "2021-07-13T09:05:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/2634.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2634" }
942,805,621
[]
https://api.github.com/repos/huggingface/datasets/issues/2634
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Related to: #2565, #2633. cc: @lewtun
2021-07-13T09:05:09Z
https://github.com/huggingface/datasets/pull/2634
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2634/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2633/comments
https://api.github.com/repos/huggingface/datasets/issues/2633/timeline
2021-07-13T05:45:13Z
null
null
MDExOlB1bGxSZXF1ZXN0Njg4MTMwOTA5
closed
[]
false
2,633
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
Update ASR tags
https://api.github.com/repos/huggingface/datasets/issues/2633/events
{ "closed_at": "2021-07-21T15:36:49Z", "closed_issues": 29, "created_at": "2021-06-08T18:48:33Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-08-05T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/6", "id": 6836458, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/6/labels", "node_id": "MDk6TWlsZXN0b25lNjgzNjQ1OA==", "number": 6, "open_issues": 0, "state": "closed", "title": "1.10", "updated_at": "2021-07-21T15:36:49Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/6" }
https://api.github.com/repos/huggingface/datasets/issues/2633/labels{/name}
2021-07-12T19:58:31Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2633.diff", "html_url": "https://github.com/huggingface/datasets/pull/2633", "merged_at": "2021-07-13T05:45:13Z", "patch_url": "https://github.com/huggingface/datasets/pull/2633.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2633" }
942,396,414
[]
https://api.github.com/repos/huggingface/datasets/issues/2633
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This PR updates the ASR tags of the 5 datasets added in #2565 following the change of task categories in #2620
2021-07-13T05:45:26Z
https://github.com/huggingface/datasets/pull/2633
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2633/reactions" }
true