id
int64
1.14B
2.23B
labels_url
stringlengths
75
75
body
stringlengths
2
33.9k
βŒ€
updated_at
stringlengths
20
20
number
int64
3.76k
6.79k
milestone
dict
repository_url
stringclasses
1 value
draft
bool
2 classes
labels
listlengths
0
4
created_at
stringlengths
20
20
comments_url
stringlengths
70
70
assignee
dict
timeline_url
stringlengths
70
70
title
stringlengths
1
290
events_url
stringlengths
68
68
active_lock_reason
null
user
dict
assignees
listlengths
0
3
performed_via_github_app
null
state_reason
stringclasses
3 values
author_association
stringclasses
3 values
closed_at
stringlengths
20
20
βŒ€
pull_request
dict
node_id
stringlengths
18
19
comments
listlengths
0
30
reactions
dict
state
stringclasses
2 values
locked
bool
1 class
url
stringlengths
61
61
html_url
stringlengths
49
51
is_pull_request
bool
2 classes
1,400,530,412
https://api.github.com/repos/huggingface/datasets/issues/5088/labels{/name}
## Describe the bug I have a local file `*.json.gz` and it can be read by `pandas.read_json(lines=True)`, but cannot be read by `load_datasets("json")` (resulting in 0 lines) ## Steps to reproduce the bug ```python fpath = '/data/junwang/.cache/general/57b6f2314cbe0bc45dda5b78f0871df2/test.json.gz' ds_panda = DatasetDict( test=Dataset.from_pandas( pd.read_json(fpath, lines=True) ) ) ds_direct = load_dataset( 'json', data_files={ 'test': fpath }, features=Features( text_input=Value(dtype="string", id=None), text_output=Value(dtype="string", id=None) ) ) len(ds_panda['test']), len(ds_direct['test']) ``` ## Expected results Lines of `ds_panda['test']` and `ds_direct['test']` should match. ## Actual results ``` Using custom data configuration default-c0ef2598760968aa Downloading and preparing dataset json/default to /data/junwang/.cache/huggingface/datasets/json/default-c0ef2598760968aa/0.0.0/e6070c77f18f01a5ad4551a8b7edfba20b8438b7cad4d94e6ad9378022ce4aab... Dataset json downloaded and prepared to /data/junwang/.cache/huggingface/datasets/json/default-c0ef2598760968aa/0.0.0/e6070c77f18f01a5ad4551a8b7edfba20b8438b7cad4d94e6ad9378022ce4aab. Subsequent calls will reuse this data. (62087, 0) ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - Platform: Ubuntu 18.04.4 LTS - Python version: 3.8.13 - PyArrow version: 9.0.0
2022-10-07T14:43:16Z
5,088
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-10-07T02:16:58Z
https://api.github.com/repos/huggingface/datasets/issues/5088/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5088/timeline
load_datasets("json", ...) don't read local .json.gz properly
https://api.github.com/repos/huggingface/datasets/issues/5088/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/112650299?v=4", "events_url": "https://api.github.com/users/junwang-wish/events{/privacy}", "followers_url": "https://api.github.com/users/junwang-wish/followers", "following_url": "https://api.github.com/users/junwang-wish/following{/other_user}", "gists_url": "https://api.github.com/users/junwang-wish/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/junwang-wish", "id": 112650299, "login": "junwang-wish", "node_id": "U_kgDOBrboOw", "organizations_url": "https://api.github.com/users/junwang-wish/orgs", "received_events_url": "https://api.github.com/users/junwang-wish/received_events", "repos_url": "https://api.github.com/users/junwang-wish/repos", "site_admin": false, "starred_url": "https://api.github.com/users/junwang-wish/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/junwang-wish/subscriptions", "type": "User", "url": "https://api.github.com/users/junwang-wish" }
[]
null
null
NONE
null
null
I_kwDODunzps5TemXs
[ "Hi @junwang-wish, thanks for reporting.\r\n\r\nUnfortunately, I'm not able to reproduce the bug. Which version of `datasets` are you using? Does the problem persist if you update `datasets`?\r\n```shell\r\npip install -U datasets\r\n``` ", "Thanks @albertvillanova I updated `datasets` from `2.5.1` to `2.5.2` and...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5088/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5088
https://github.com/huggingface/datasets/issues/5088
false
1,400,487,967
https://api.github.com/repos/huggingface/datasets/issues/5087/labels{/name}
Fix #5085
2022-10-07T18:43:03Z
5,087
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-07T01:07:00Z
https://api.github.com/repos/huggingface/datasets/issues/5087/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5087/timeline
Fix filter with empty indices
https://api.github.com/repos/huggingface/datasets/issues/5087/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/23029765?v=4", "events_url": "https://api.github.com/users/Mouhanedg56/events{/privacy}", "followers_url": "https://api.github.com/users/Mouhanedg56/followers", "following_url": "https://api.github.com/users/Mouhanedg56/following{/other_user}", "gists_url": "https://api.github.com/users/Mouhanedg56/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Mouhanedg56", "id": 23029765, "login": "Mouhanedg56", "node_id": "MDQ6VXNlcjIzMDI5NzY1", "organizations_url": "https://api.github.com/users/Mouhanedg56/orgs", "received_events_url": "https://api.github.com/users/Mouhanedg56/received_events", "repos_url": "https://api.github.com/users/Mouhanedg56/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Mouhanedg56/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Mouhanedg56/subscriptions", "type": "User", "url": "https://api.github.com/users/Mouhanedg56" }
[]
null
null
CONTRIBUTOR
2022-10-07T18:40:26Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5087.diff", "html_url": "https://github.com/huggingface/datasets/pull/5087", "merged_at": "2022-10-07T18:40:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/5087.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5087" }
PR_kwDODunzps5AW-N9
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5087/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5087
https://github.com/huggingface/datasets/pull/5087
true
1,400,216,975
https://api.github.com/repos/huggingface/datasets/issues/5086/labels{/name}
## Describe the bug I was following chap 5 from huggingface course: https://huggingface.co/course/chapter5/6?fw=tf However, I'm not able to download the datasets, with a 404 erros <img width="1160" alt="iShot2022-10-06_15 54 50" src="https://user-images.githubusercontent.com/54015474/194406327-ae62c2f3-1da5-4686-8631-13d879a0edee.png"> ## Steps to reproduce the bug ```python from huggingface_hub import hf_hub_url data_files = hf_hub_url( repo_id="lewtun/github-issues", filename="datasets-issues-with-hf-doc-builder.jsonl", repo_type="dataset", ) from datasets import load_dataset issues_dataset = load_dataset("json", data_files=data_files, split="train") issues_dataset ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.5.2 - Platform: macOS-10.16-x86_64-i386-64bit - Python version: 3.9.12 - PyArrow version: 9.0.0 - Pandas version: 1.4.4
2022-10-07T15:12:01Z
5,086
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-10-06T19:48:58Z
https://api.github.com/repos/huggingface/datasets/issues/5086/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5086/timeline
HTTPError: 404 Client Error: Not Found for url
https://api.github.com/repos/huggingface/datasets/issues/5086/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/54015474?v=4", "events_url": "https://api.github.com/users/km5ar/events{/privacy}", "followers_url": "https://api.github.com/users/km5ar/followers", "following_url": "https://api.github.com/users/km5ar/following{/other_user}", "gists_url": "https://api.github.com/users/km5ar/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/km5ar", "id": 54015474, "login": "km5ar", "node_id": "MDQ6VXNlcjU0MDE1NDc0", "organizations_url": "https://api.github.com/users/km5ar/orgs", "received_events_url": "https://api.github.com/users/km5ar/received_events", "repos_url": "https://api.github.com/users/km5ar/repos", "site_admin": false, "starred_url": "https://api.github.com/users/km5ar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/km5ar/subscriptions", "type": "User", "url": "https://api.github.com/users/km5ar" }
[]
null
completed
NONE
2022-10-07T15:12:01Z
null
I_kwDODunzps5TdZ2P
[ "FYI @lewtun ", "Hi @km5ar, thanks for reporting.\r\n\r\nThis should be fixed in the notebook:\r\n- the filename `datasets-issues-with-hf-doc-builder.jsonl` no longer exists on the repo; instead, current filename is `datasets-issues-with-comments.jsonl`\r\n- see: https://huggingface.co/datasets/lewtun/github-issu...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5086/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5086
https://github.com/huggingface/datasets/issues/5086
false
1,400,113,569
https://api.github.com/repos/huggingface/datasets/issues/5085/labels{/name}
## Describe the bug When filtering a dataset twice, where the first result is an empty dataset, the second dataset seems corrupted. ## Steps to reproduce the bug ```python datasets = load_dataset("glue", "sst2") dataset_split = datasets['validation'] ds_filter_1 = dataset_split.filter(lambda x: False) # Some filtering condition that leads to an empty dataset assert ds_filter_1.num_rows == 0 sentences = ds_filter_1['sentence'] assert len(sentences) == 0 ds_filter_2 = ds_filter_1.filter(lambda x: False) # Some other filtering condition assert ds_filter_2.num_rows == 0 assert 'sentence' in ds_filter_2.column_names sentences = ds_filter_2['sentence'] ``` ## Expected results The last line should be returning an empty list, same as 4 lines above. ## Actual results The last line currently raises `IndexError: index out of bounds`. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.5.2 - Platform: macOS-11.6.6-x86_64-i386-64bit - Python version: 3.9.11 - PyArrow version: 7.0.0 - Pandas version: 1.4.1
2022-10-07T19:06:02Z
5,085
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "DF8D62", "default": false, "descrip...
2022-10-06T18:18:49Z
https://api.github.com/repos/huggingface/datasets/issues/5085/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/23029765?v=4", "events_url": "https://api.github.com/users/Mouhanedg56/events{/privacy}", "followers_url": "https://api.github.com/users/Mouhanedg56/followers", "following_url": "https://api.github.com/users/Mouhanedg56/following{/other_user}", "gists_url": "https://api.github.com/users/Mouhanedg56/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Mouhanedg56", "id": 23029765, "login": "Mouhanedg56", "node_id": "MDQ6VXNlcjIzMDI5NzY1", "organizations_url": "https://api.github.com/users/Mouhanedg56/orgs", "received_events_url": "https://api.github.com/users/Mouhanedg56/received_events", "repos_url": "https://api.github.com/users/Mouhanedg56/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Mouhanedg56/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Mouhanedg56/subscriptions", "type": "User", "url": "https://api.github.com/users/Mouhanedg56" }
https://api.github.com/repos/huggingface/datasets/issues/5085/timeline
Filtering on an empty dataset returns a corrupted dataset.
https://api.github.com/repos/huggingface/datasets/issues/5085/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36087158?v=4", "events_url": "https://api.github.com/users/gabegma/events{/privacy}", "followers_url": "https://api.github.com/users/gabegma/followers", "following_url": "https://api.github.com/users/gabegma/following{/other_user}", "gists_url": "https://api.github.com/users/gabegma/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/gabegma", "id": 36087158, "login": "gabegma", "node_id": "MDQ6VXNlcjM2MDg3MTU4", "organizations_url": "https://api.github.com/users/gabegma/orgs", "received_events_url": "https://api.github.com/users/gabegma/received_events", "repos_url": "https://api.github.com/users/gabegma/repos", "site_admin": false, "starred_url": "https://api.github.com/users/gabegma/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gabegma/subscriptions", "type": "User", "url": "https://api.github.com/users/gabegma" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/23029765?v=4", "events_url": "https://api.github.com/users/Mouhanedg56/events{/privacy}", "followers_url": "https://api.github.com/users/Mouhanedg56/followers", "following_url": "https://api.github.com/users/Mouhanedg56/following{/other_user}"...
null
completed
NONE
2022-10-07T18:40:26Z
null
I_kwDODunzps5TdAmh
[ "~~It seems like #5043 fix (merged recently) is the root cause of such behaviour. When we empty indices mapping (because the dataset length equals to zero), we can no longer get column item like: `ds_filter_2['sentence']` which uses\r\n`ds_filter_1._indices.column(0)`~~\r\n\r\n**UPDATE:**\r\nEmpty datasets are retu...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 3, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/5085/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5085
https://github.com/huggingface/datasets/issues/5085
false
1,400,016,229
https://api.github.com/repos/huggingface/datasets/issues/5084/labels{/name}
This code now returns a numpy array: ```python from datasets import load_dataset ds = load_dataset("imagenet-1k", split="train", streaming=True).with_format("np") print(next(iter(ds))["image"]) ``` It also works with "arrow", "pandas", "torch", "tf" and "jax" ### Implementation details: I'm using the existing code to format an Arrow Table to the right output format for simplicity. Therefore it's probbaly not the most optimized approach. For example to output PyTorch tensors it does this for every example: python data -> arrow table -> numpy extracted data -> pytorch formatted data ### Releasing this feature Even though I consider this as a bug/inconsistency, this change is a breaking change. And I'm sure some users were relying on the torch iterable dataset to return PIL Image and used data collators to convert to pytorch. So I guess this is `datasets` 3.0 ? ### TODO - [x] merge https://github.com/huggingface/datasets/pull/5072 - [ ] docs - [ ] tests Close https://github.com/huggingface/datasets/issues/5083
2023-09-24T10:06:51Z
5,084
null
https://api.github.com/repos/huggingface/datasets
true
[]
2022-10-06T16:53:38Z
https://api.github.com/repos/huggingface/datasets/issues/5084/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5084/timeline
IterableDataset formatting in numpy/torch/tf/jax
https://api.github.com/repos/huggingface/datasets/issues/5084/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-12-20T17:19:52Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5084.diff", "html_url": "https://github.com/huggingface/datasets/pull/5084", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5084.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5084" }
PR_kwDODunzps5AVXwm
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5084). All of your documentation changes will be reflected on that endpoint.", "Actually I'm not happy with this implementation. It always require the iterable dataset to have definite `features`, which removes a lot of flexibi...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5084/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5084
https://github.com/huggingface/datasets/pull/5084
true
1,399,842,514
https://api.github.com/repos/huggingface/datasets/issues/5083/labels{/name}
Right now `IterableDataset` doesn't do any formatting. In particular this code should return a numpy array: ```python from datasets import load_dataset ds = load_dataset("imagenet-1k", split="train", streaming=True).with_format("np") print(next(iter(ds))["image"]) ``` Right now it returns a PIL.Image. Setting `streaming=False` does return a numpy array after #5072
2023-10-09T12:42:15Z
5,083
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "fef2c0", "default": fals...
2022-10-06T15:14:58Z
https://api.github.com/repos/huggingface/datasets/issues/5083/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
https://api.github.com/repos/huggingface/datasets/issues/5083/timeline
Support numpy/torch/tf/jax formatting for IterableDataset
https://api.github.com/repos/huggingface/datasets/issues/5083/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
completed
MEMBER
2023-10-09T12:42:15Z
null
I_kwDODunzps5Tb-bS
[ "hii @lhoestq, can you assign this issue to me? Though i am new to open source still I would love to put my best foot forward. I can see there isn't anyone right now assigned to this issue.", "Hi @zutarich ! This issue was fixed by #5852 - sorry I forgot to close it\r\n\r\nFeel free to look for other issues and p...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5083/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5083
https://github.com/huggingface/datasets/issues/5083
false
1,399,379,777
https://api.github.com/repos/huggingface/datasets/issues/5082/labels{/name}
Fixing #514 . Hello @mariosasko πŸ‘‹, I have implemented what you have recommanded to fix the keep in memory problem for shuffle on the issue #514 .
2022-10-07T14:35:34Z
5,082
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-06T11:10:46Z
https://api.github.com/repos/huggingface/datasets/issues/5082/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5082/timeline
adding keep in memory
https://api.github.com/repos/huggingface/datasets/issues/5082/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/66799406?v=4", "events_url": "https://api.github.com/users/Mustapha-AJEGHRIR/events{/privacy}", "followers_url": "https://api.github.com/users/Mustapha-AJEGHRIR/followers", "following_url": "https://api.github.com/users/Mustapha-AJEGHRIR/following{/other_user}", "gists_url": "https://api.github.com/users/Mustapha-AJEGHRIR/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Mustapha-AJEGHRIR", "id": 66799406, "login": "Mustapha-AJEGHRIR", "node_id": "MDQ6VXNlcjY2Nzk5NDA2", "organizations_url": "https://api.github.com/users/Mustapha-AJEGHRIR/orgs", "received_events_url": "https://api.github.com/users/Mustapha-AJEGHRIR/received_events", "repos_url": "https://api.github.com/users/Mustapha-AJEGHRIR/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Mustapha-AJEGHRIR/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Mustapha-AJEGHRIR/subscriptions", "type": "User", "url": "https://api.github.com/users/Mustapha-AJEGHRIR" }
[]
null
null
CONTRIBUTOR
2022-10-07T14:32:54Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5082.diff", "html_url": "https://github.com/huggingface/datasets/pull/5082", "merged_at": "2022-10-07T14:32:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/5082.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5082" }
PR_kwDODunzps5ATJv-
[ "_The documentation is not available anymore as the PR was closed or merged._", "Hi @mariosasko , I have added a test for the `keep_in_memory` version. I have also removed the `Compatible with temp_seed` part in the scope of `dset_shuffled`, please verify if that makes sense." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5082/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5082
https://github.com/huggingface/datasets/pull/5082
true
1,399,340,050
https://api.github.com/repos/huggingface/datasets/issues/5081/labels{/name}
## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("sentence-transformers/parallel-sentences") ``` raises this: ``` /home/phmay/miniconda3/envs/paraphrase-mining/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py:697: FutureWarning: the 'mangle_dupe_cols' keyword is deprecated and will be removed in a future version. Please take steps to stop the use of 'mangle_dupe_cols' return pd.read_csv(xopen(filepath_or_buffer, "rb", use_auth_token=use_auth_token), **kwargs) /home/phmay/miniconda3/envs/paraphrase-mining/lib/python3.9/site-packages/datasets/download/streaming_download_manager.py:697: FutureWarning: the 'mangle_dupe_cols' keyword is deprecated and will be removed in a future version. Please take steps to stop the use of 'mangle_dupe_cols' return pd.read_csv(xopen(filepath_or_buffer, "rb", use_auth_token=use_auth_token), **kwargs) --------------------------------------------------------------------------- ValueError Traceback (most recent call last) Cell In [4], line 1 ----> 1 dataset = load_dataset("sentence-transformers/parallel-sentences", split="train") File ~/miniconda3/envs/paraphrase-mining/lib/python3.9/site-packages/datasets/load.py:1693, in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, **config_kwargs) 1690 try_from_hf_gcs = path not in _PACKAGED_DATASETS_MODULES 1692 # Download and prepare data -> 1693 builder_instance.download_and_prepare( 1694 download_config=download_config, 1695 download_mode=download_mode, 1696 ignore_verifications=ignore_verifications, 1697 try_from_hf_gcs=try_from_hf_gcs, 1698 use_auth_token=use_auth_token, 1699 ) 1701 # Build dataset for splits 1702 keep_in_memory = ( 1703 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size) 1704 ) File ~/miniconda3/envs/paraphrase-mining/lib/python3.9/site-packages/datasets/builder.py:807, in DatasetBuilder.download_and_prepare(self, output_dir, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, file_format, max_shard_size, storage_options, **download_and_prepare_kwargs) 801 if not downloaded_from_gcs: 802 prepare_split_kwargs = { 803 "file_format": file_format, 804 "max_shard_size": max_shard_size, 805 **download_and_prepare_kwargs, 806 } --> 807 self._download_and_prepare( 808 dl_manager=dl_manager, 809 verify_infos=verify_infos, 810 **prepare_split_kwargs, 811 **download_and_prepare_kwargs, 812 ) 813 # Sync info 814 self.info.dataset_size = sum(split.num_bytes for split in self.info.splits.values()) File ~/miniconda3/envs/paraphrase-mining/lib/python3.9/site-packages/datasets/builder.py:898, in DatasetBuilder._download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 894 split_dict.add(split_generator.split_info) 896 try: 897 # Prepare split will record examples associated to the split --> 898 self._prepare_split(split_generator, **prepare_split_kwargs) 899 except OSError as e: 900 raise OSError( 901 "Cannot find data file. " 902 + (self.manual_download_instructions or "") 903 + "\nOriginal error:\n" 904 + str(e) 905 ) from None File ~/miniconda3/envs/paraphrase-mining/lib/python3.9/site-packages/datasets/builder.py:1513, in ArrowBasedBuilder._prepare_split(self, split_generator, file_format, max_shard_size) 1506 shard_id += 1 1507 writer = writer_class( 1508 features=writer._features, 1509 path=fpath.replace("SSSSS", f"{shard_id:05d}"), 1510 storage_options=self._fs.storage_options, 1511 embed_local_files=embed_local_files, 1512 ) -> 1513 writer.write_table(table) 1514 finally: 1515 num_shards = shard_id + 1 File ~/miniconda3/envs/paraphrase-mining/lib/python3.9/site-packages/datasets/arrow_writer.py:540, in ArrowWriter.write_table(self, pa_table, writer_batch_size) 538 if self.pa_writer is None: 539 self._build_writer(inferred_schema=pa_table.schema) --> 540 pa_table = table_cast(pa_table, self._schema) 541 if self.embed_local_files: 542 pa_table = embed_table_storage(pa_table) File ~/miniconda3/envs/paraphrase-mining/lib/python3.9/site-packages/datasets/table.py:2044, in table_cast(table, schema) 2032 """Improved version of pa.Table.cast. 2033 2034 It supports casting to feature types stored in the schema metadata. (...) 2041 table (:obj:`pyarrow.Table`): the casted table 2042 """ 2043 if table.schema != schema: -> 2044 return cast_table_to_schema(table, schema) 2045 elif table.schema.metadata != schema.metadata: 2046 return table.replace_schema_metadata(schema.metadata) File ~/miniconda3/envs/paraphrase-mining/lib/python3.9/site-packages/datasets/table.py:2005, in cast_table_to_schema(table, schema) 2003 features = Features.from_arrow_schema(schema) 2004 if sorted(table.column_names) != sorted(features): -> 2005 raise ValueError(f"Couldn't cast\n{table.schema}\nto\n{features}\nbecause column names don't match") 2006 arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] 2007 return pa.Table.from_arrays(arrays, schema=schema) ValueError: Couldn't cast Action taken on Parliament's resolutions: see Minutes: string NΓ‘slednΓ½ postup na zΓ‘kladΔ› usnesenΓ­ Parlamentu: viz zΓ‘pis: string -- schema metadata -- pandas: '{"index_columns": [{"kind": "range", "name": null, "start": 0, "' + 742 to {'Membership of Parliament: see Minutes': Value(dtype='string', id=None), 'Π‘ΡŠΡΡ‚Π°Π² Π½Π° ΠŸΠ°Ρ€Π»Π°ΠΌΠ΅Π½Ρ‚Π°: Π²ΠΆ. ΠΏΡ€ΠΎΡ‚ΠΎΠΊΠΎΠ»ΠΈ': Value(dtype='string', id=None)} because column names don't match ``` ## Expected results no error ## Actual results error ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - Platform: Linux - Python version: Python 3.9.13 - PyArrow version: pyarrow 9.0.0 - transformers 4.22.2 - datasets 2.5.2
2022-10-11T10:00:48Z
5,081
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-10-06T10:47:51Z
https://api.github.com/repos/huggingface/datasets/issues/5081/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5081/timeline
Bug loading `sentence-transformers/parallel-sentences`
https://api.github.com/repos/huggingface/datasets/issues/5081/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/229382?v=4", "events_url": "https://api.github.com/users/PhilipMay/events{/privacy}", "followers_url": "https://api.github.com/users/PhilipMay/followers", "following_url": "https://api.github.com/users/PhilipMay/following{/other_user}", "gists_url": "https://api.github.com/users/PhilipMay/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/PhilipMay", "id": 229382, "login": "PhilipMay", "node_id": "MDQ6VXNlcjIyOTM4Mg==", "organizations_url": "https://api.github.com/users/PhilipMay/orgs", "received_events_url": "https://api.github.com/users/PhilipMay/received_events", "repos_url": "https://api.github.com/users/PhilipMay/repos", "site_admin": false, "starred_url": "https://api.github.com/users/PhilipMay/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/PhilipMay/subscriptions", "type": "User", "url": "https://api.github.com/users/PhilipMay" }
[]
null
null
CONTRIBUTOR
null
null
I_kwDODunzps5TaDwS
[ "tagging @nreimers ", "The dataset is sadly not really compatible to be loaded with `load_dataset`. So far it is better to git clone it and to use the files directly.\r\n\r\nA data loading script would be needed to be added to this dataset. But this was too much overhead / not really intuitive how to create it.",...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5081/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5081
https://github.com/huggingface/datasets/issues/5081
false
1,398,849,565
https://api.github.com/repos/huggingface/datasets/issues/5080/labels{/name}
## Is your feature request related to a problem? As previously discussed in our meeting with @Wauplin and agreed on our last datasets team sync meeting, I'm investigating how `datasets` can use `hfh` for caching. ## Describe the solution you'd like Due to the peculiarities of the `datasets` cache, I would propose adopting `hfh` caching system in stages. First, we could easily start using `hfh` caching for: - dataset Python scripts - dataset READMEs - dataset infos JSON files (now deprecated) Second, we could also use `hfh` caching for data files downloaded from the Hub. Further investigation is needed for: - files downloaded from non-Hub hosts - extracted files from downloaded archive/compressed files - generated Arrow files ## Additional context Docs about the `hfh` caching system: - [Manage huggingface_hub cache-system](https://huggingface.co/docs/huggingface_hub/main/en/how-to-cache) - [Cache-system reference](https://huggingface.co/docs/huggingface_hub/main/en/package_reference/cache) The `transformers` library has already adopted `hfh` for caching. See: - huggingface/transformers#18438 - huggingface/transformers#18857 - huggingface/transformers#18966
2022-10-06T14:26:05Z
5,080
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-10-06T05:51:58Z
https://api.github.com/repos/huggingface/datasets/issues/5080/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5080/timeline
Use hfh for caching
https://api.github.com/repos/huggingface/datasets/issues/5080/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
null
MEMBER
null
null
I_kwDODunzps5TYMAd
[ "There is some discussion in https://github.com/huggingface/huggingface_hub/pull/1088 if it can help :)" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5080/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5080
https://github.com/huggingface/datasets/issues/5080
false
1,398,609,305
https://api.github.com/repos/huggingface/datasets/issues/5079/labels{/name}
Closes #5074 Replaces `AssertionError` in the following files with more descriptive exceptions: - `src/datasets/arrow_reader.py` - `src/datasets/builder.py` - `src/datasets/utils/version.py` The issue listed more files that needed to be fixed, but the rest of them were contained in the top-level `datasets` directory, which was removed when #4974 was merged
2022-10-07T14:35:43Z
5,079
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-06T01:39:35Z
https://api.github.com/repos/huggingface/datasets/issues/5079/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5079/timeline
refactor: replace AssertionError with more meaningful exceptions (#5074)
https://api.github.com/repos/huggingface/datasets/issues/5079/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/20004072?v=4", "events_url": "https://api.github.com/users/galbwe/events{/privacy}", "followers_url": "https://api.github.com/users/galbwe/followers", "following_url": "https://api.github.com/users/galbwe/following{/other_user}", "gists_url": "https://api.github.com/users/galbwe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/galbwe", "id": 20004072, "login": "galbwe", "node_id": "MDQ6VXNlcjIwMDA0MDcy", "organizations_url": "https://api.github.com/users/galbwe/orgs", "received_events_url": "https://api.github.com/users/galbwe/received_events", "repos_url": "https://api.github.com/users/galbwe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/galbwe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/galbwe/subscriptions", "type": "User", "url": "https://api.github.com/users/galbwe" }
[]
null
null
CONTRIBUTOR
2022-10-07T14:33:10Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5079.diff", "html_url": "https://github.com/huggingface/datasets/pull/5079", "merged_at": "2022-10-07T14:33:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/5079.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5079" }
PR_kwDODunzps5AQemi
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5079/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5079
https://github.com/huggingface/datasets/pull/5079
true
1,398,335,148
https://api.github.com/repos/huggingface/datasets/issues/5078/labels{/name}
Fixes header level so `Dataset features` is the doc title instead of `The Audio type`: ![Screen Shot 2022-10-05 at 1 22 02 PM](https://user-images.githubusercontent.com/59462357/194155840-eeb5d62f-f4eb-411e-b281-8494c5fffdce.png)
2022-10-06T08:12:23Z
5,078
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-05T20:22:44Z
https://api.github.com/repos/huggingface/datasets/issues/5078/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5078/timeline
Fix header level in Audio docs
https://api.github.com/repos/huggingface/datasets/issues/5078/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
null
null
MEMBER
2022-10-06T08:09:41Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5078.diff", "html_url": "https://github.com/huggingface/datasets/pull/5078", "merged_at": "2022-10-06T08:09:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/5078.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5078" }
PR_kwDODunzps5APjkH
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5078/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5078
https://github.com/huggingface/datasets/pull/5078
true
1,398,080,859
https://api.github.com/repos/huggingface/datasets/issues/5077/labels{/name}
Fix passed `download_config` in `HubDatasetModuleFactoryWithoutScript`.
2022-10-06T05:31:22Z
5,077
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-05T16:42:36Z
https://api.github.com/repos/huggingface/datasets/issues/5077/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5077/timeline
Fix passed download_config in HubDatasetModuleFactoryWithoutScript
https://api.github.com/repos/huggingface/datasets/issues/5077/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-10-06T05:29:06Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5077.diff", "html_url": "https://github.com/huggingface/datasets/pull/5077", "merged_at": "2022-10-06T05:29:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/5077.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5077" }
PR_kwDODunzps5AOs9L
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5077/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5077
https://github.com/huggingface/datasets/pull/5077
true
1,397,918,092
https://api.github.com/repos/huggingface/datasets/issues/5076/labels{/name}
Status: Ready for review Description of Changes: Fixes #5075 Changes proposed in this pull request: - Throw EnvironmentError instead of OSError in `push_to_hub` when the Hub token is not present.
2022-10-07T14:35:57Z
5,076
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-05T14:46:29Z
https://api.github.com/repos/huggingface/datasets/issues/5076/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5076/timeline
fix: update exception throw from OSError to EnvironmentError in `push…
https://api.github.com/repos/huggingface/datasets/issues/5076/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/29496999?v=4", "events_url": "https://api.github.com/users/rahulXs/events{/privacy}", "followers_url": "https://api.github.com/users/rahulXs/followers", "following_url": "https://api.github.com/users/rahulXs/following{/other_user}", "gists_url": "https://api.github.com/users/rahulXs/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/rahulXs", "id": 29496999, "login": "rahulXs", "node_id": "MDQ6VXNlcjI5NDk2OTk5", "organizations_url": "https://api.github.com/users/rahulXs/orgs", "received_events_url": "https://api.github.com/users/rahulXs/received_events", "repos_url": "https://api.github.com/users/rahulXs/repos", "site_admin": false, "starred_url": "https://api.github.com/users/rahulXs/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rahulXs/subscriptions", "type": "User", "url": "https://api.github.com/users/rahulXs" }
[]
null
null
CONTRIBUTOR
2022-10-07T14:33:27Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5076.diff", "html_url": "https://github.com/huggingface/datasets/pull/5076", "merged_at": "2022-10-07T14:33:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/5076.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5076" }
PR_kwDODunzps5AOJp7
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5076/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5076
https://github.com/huggingface/datasets/pull/5076
true
1,397,865,501
https://api.github.com/repos/huggingface/datasets/issues/5075/labels{/name}
Throw EnvironmentError instead of OSError ([link](https://github.com/huggingface/datasets/blob/6ad430ba0cdeeb601170f732d4bd977f5c04594d/src/datasets/arrow_dataset.py#L4306) to the line) in `push_to_hub` when the Hub token is not present.
2022-10-07T14:33:28Z
5,075
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "7057ff", "default": true, "description": "Good for newcomers", "id": 1935892877, "name": "good first issue", "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue" }, { "color": "DF8D62", "defa...
2022-10-05T14:14:18Z
https://api.github.com/repos/huggingface/datasets/issues/5075/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5075/timeline
Throw EnvironmentError when token is not present
https://api.github.com/repos/huggingface/datasets/issues/5075/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
completed
CONTRIBUTOR
2022-10-07T14:33:28Z
null
I_kwDODunzps5TUbwd
[ "@mariosasko I've raised a PR #5076 against this issue. Please help to review. Thanks." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5075/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5075
https://github.com/huggingface/datasets/issues/5075
false
1,397,850,352
https://api.github.com/repos/huggingface/datasets/issues/5074/labels{/name}
Replace the AssertionErrors with more meaningful errors such as ValueError, TypeError, etc. The files with AssertionErrors that need to be replaced: ``` src/datasets/arrow_reader.py src/datasets/builder.py src/datasets/utils/version.py ```
2022-10-07T14:33:11Z
5,074
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "7057ff", "default": true, "description": "Good for newcomers", "id": 1935892877, "name": "good first issue", "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue" }, { "color": "DF8D62", "defa...
2022-10-05T14:03:55Z
https://api.github.com/repos/huggingface/datasets/issues/5074/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/20004072?v=4", "events_url": "https://api.github.com/users/galbwe/events{/privacy}", "followers_url": "https://api.github.com/users/galbwe/followers", "following_url": "https://api.github.com/users/galbwe/following{/other_user}", "gists_url": "https://api.github.com/users/galbwe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/galbwe", "id": 20004072, "login": "galbwe", "node_id": "MDQ6VXNlcjIwMDA0MDcy", "organizations_url": "https://api.github.com/users/galbwe/orgs", "received_events_url": "https://api.github.com/users/galbwe/received_events", "repos_url": "https://api.github.com/users/galbwe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/galbwe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/galbwe/subscriptions", "type": "User", "url": "https://api.github.com/users/galbwe" }
https://api.github.com/repos/huggingface/datasets/issues/5074/timeline
Replace AssertionErrors with more meaningful errors
https://api.github.com/repos/huggingface/datasets/issues/5074/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/20004072?v=4", "events_url": "https://api.github.com/users/galbwe/events{/privacy}", "followers_url": "https://api.github.com/users/galbwe/followers", "following_url": "https://api.github.com/users/galbwe/following{/other_user}", "gists_ur...
null
completed
CONTRIBUTOR
2022-10-07T14:33:11Z
null
I_kwDODunzps5TUYDw
[ "Hi, can I pick up this issue?", "#self-assign", "Looks like the top-level `datasource` directory was removed when https://github.com/huggingface/datasets/pull/4974 was merged, so there are 3 source files to fix." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5074/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5074
https://github.com/huggingface/datasets/issues/5074
false
1,397,832,183
https://api.github.com/repos/huggingface/datasets/issues/5073/labels{/name}
Hello! @mariosasko This pull request relates to issue #5050 and intends to add the format to datasets loaded from disk. All I did was add a set_format in the Dataset.load_from_disk, as DatasetDict.load_from_disk relies on the first. I don't know if I should add a test and where, so let me know if I should and I can work on that as well!
2022-10-11T16:55:07Z
5,073
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-05T13:51:47Z
https://api.github.com/repos/huggingface/datasets/issues/5073/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5073/timeline
Restore saved format state in `load_from_disk`
https://api.github.com/repos/huggingface/datasets/issues/5073/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/74454835?v=4", "events_url": "https://api.github.com/users/asofiaoliveira/events{/privacy}", "followers_url": "https://api.github.com/users/asofiaoliveira/followers", "following_url": "https://api.github.com/users/asofiaoliveira/following{/other_user}", "gists_url": "https://api.github.com/users/asofiaoliveira/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/asofiaoliveira", "id": 74454835, "login": "asofiaoliveira", "node_id": "MDQ6VXNlcjc0NDU0ODM1", "organizations_url": "https://api.github.com/users/asofiaoliveira/orgs", "received_events_url": "https://api.github.com/users/asofiaoliveira/received_events", "repos_url": "https://api.github.com/users/asofiaoliveira/repos", "site_admin": false, "starred_url": "https://api.github.com/users/asofiaoliveira/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/asofiaoliveira/subscriptions", "type": "User", "url": "https://api.github.com/users/asofiaoliveira" }
[]
null
null
CONTRIBUTOR
2022-10-11T16:49:23Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5073.diff", "html_url": "https://github.com/huggingface/datasets/pull/5073", "merged_at": "2022-10-11T16:49:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/5073.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5073" }
PR_kwDODunzps5AN3Gn
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5073/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5073
https://github.com/huggingface/datasets/pull/5073
true
1,397,765,531
https://api.github.com/repos/huggingface/datasets/issues/5072/labels{/name}
Added support for image and audio formatting for numpy, torch, tf and jax. For images, the dtype used is the one of the image (the one returned by PIL.Image), e.g. uint8 I also added support for string, binary and None types. In particular for torch and jax, strings are kept unchanged (previously it was returning an error because you can't create a tensor of strings)
2022-10-10T13:24:10Z
5,072
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-05T13:07:03Z
https://api.github.com/repos/huggingface/datasets/issues/5072/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5072/timeline
Image & Audio formatting for numpy/torch/tf/jax
https://api.github.com/repos/huggingface/datasets/issues/5072/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-10-10T13:21:32Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5072.diff", "html_url": "https://github.com/huggingface/datasets/pull/5072", "merged_at": "2022-10-10T13:21:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/5072.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5072" }
PR_kwDODunzps5ANoo5
[ "_The documentation is not available anymore as the PR was closed or merged._", "I just added a consolidation step so that numpy arrays or tensors of images are stacked together if the shapes match, instead of having lists of tensors\r\n\r\nFeel free to review @mariosasko :)", "I added a few lines in the docs a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5072/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5072
https://github.com/huggingface/datasets/pull/5072
true
1,397,301,270
https://api.github.com/repos/huggingface/datasets/issues/5071/labels{/name}
This PR supports defining a default config name, even if no predefined allowed config names are set. Fix #5070. CC: @stas00
2022-10-06T14:43:12Z
5,071
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-05T06:28:39Z
https://api.github.com/repos/huggingface/datasets/issues/5071/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5071/timeline
Support DEFAULT_CONFIG_NAME when no BUILDER_CONFIGS
https://api.github.com/repos/huggingface/datasets/issues/5071/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-10-06T14:40:26Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5071.diff", "html_url": "https://github.com/huggingface/datasets/pull/5071", "merged_at": "2022-10-06T14:40:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/5071.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5071" }
PR_kwDODunzps5AMG3g
[ "_The documentation is not available anymore as the PR was closed or merged._", "Super, thanks a lot for adding this support, Albert!" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5071/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5071
https://github.com/huggingface/datasets/pull/5071
true
1,396,765,647
https://api.github.com/repos/huggingface/datasets/issues/5070/labels{/name}
**Is your feature request related to a problem? Please describe.** As discussed with @stas00, we could support defining a default config name, even if no predefined allowed config names are set. That is, support `DEFAULT_CONFIG_NAME`, even when `BUILDER_CONFIGS` is not defined. **Additional context** In order to support creating configs on the fly **by name** (not using kwargs), the list of allowed builder configs `BUILDER_CONFIGS` must not be set. However, if so, then `DEFAULT_CONFIG_NAME` is not supported.
2022-10-06T14:40:26Z
5,070
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-10-04T19:49:35Z
https://api.github.com/repos/huggingface/datasets/issues/5070/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5070/timeline
Support default config name when no builder configs
https://api.github.com/repos/huggingface/datasets/issues/5070/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
MEMBER
2022-10-06T14:40:26Z
null
I_kwDODunzps5TQPPP
[ "Thank you for creating this feature request, Albert.\r\n\r\nFor context this is the datatest where Albert has been helping me to switch to on-the-fly split config https://huggingface.co/datasets/HuggingFaceM4/cm4-synthetic-testing\r\n\r\nand the attempt to switch on-the-fly splits was here: https://huggingface.co/...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5070/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5070
https://github.com/huggingface/datasets/issues/5070
false
1,396,361,768
https://api.github.com/repos/huggingface/datasets/issues/5067/labels{/name}
This PR updates the `CONTRIBUTING.md` guide, once the all dataset scripts have been removed from the GitHub repo and transferred to the HF Hub: - #4974 See diff here: https://github.com/huggingface/datasets/commit/e3291ecff9e54f09fcee3f313f051a03fdc3d94b Additionally, this PR fixes the line separator that by some previous mistake was CRLF instead of LF.
2022-10-06T06:14:43Z
5,067
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-04T14:16:05Z
https://api.github.com/repos/huggingface/datasets/issues/5067/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5067/timeline
Fix CONTRIBUTING once dataset scripts transferred to Hub
https://api.github.com/repos/huggingface/datasets/issues/5067/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-10-06T06:12:12Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5067.diff", "html_url": "https://github.com/huggingface/datasets/pull/5067", "merged_at": "2022-10-06T06:12:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/5067.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5067" }
PR_kwDODunzps5AI86d
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5067/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5067
https://github.com/huggingface/datasets/pull/5067
true
1,396,086,745
https://api.github.com/repos/huggingface/datasets/issues/5066/labels{/name}
This PR implements support for streaming out-of-the-box dataset scripts containing `gzip.open`. This has been a recurring issue. See, e.g.: - #5060 - #3191
2022-10-06T15:13:51Z
5,066
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-04T11:20:05Z
https://api.github.com/repos/huggingface/datasets/issues/5066/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5066/timeline
Support streaming gzip.open
https://api.github.com/repos/huggingface/datasets/issues/5066/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-10-06T15:11:29Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5066.diff", "html_url": "https://github.com/huggingface/datasets/pull/5066", "merged_at": "2022-10-06T15:11:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/5066.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5066" }
PR_kwDODunzps5AIDWj
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5066/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5066
https://github.com/huggingface/datasets/pull/5066
true
1,396,003,362
https://api.github.com/repos/huggingface/datasets/issues/5065/labels{/name}
Added a CI job for python 3.10 Some dependencies don't work on 3.10 like apache beam, so I remove them from the extras in this case. I also removed some s3 fixtures that we don't use anymore (and that don't work on 3.10 anyway)
2022-11-29T15:28:05Z
5,065
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-04T10:13:51Z
https://api.github.com/repos/huggingface/datasets/issues/5065/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5065/timeline
Ci py3.10
https://api.github.com/repos/huggingface/datasets/issues/5065/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-11-29T15:25:26Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5065.diff", "html_url": "https://github.com/huggingface/datasets/pull/5065", "merged_at": "2022-11-29T15:25:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/5065.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5065" }
PR_kwDODunzps5AHxlQ
[ "_The documentation is not available anymore as the PR was closed or merged._", "Does it sound good to you @albertvillanova ?" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5065/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5065
https://github.com/huggingface/datasets/pull/5065
true
1,395,978,143
https://api.github.com/repos/huggingface/datasets/issues/5064/labels{/name}
This PR aligns the signature of `create_repo`/`delete_repo` with the current one in hfh, by removing deprecated `name` and `organization`, and using `repo_id` instead. Related to: - #5063 CC: @lhoestq
2022-10-07T17:02:11Z
5,064
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-04T09:54:53Z
https://api.github.com/repos/huggingface/datasets/issues/5064/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5064/timeline
Align signature of create/delete_repo with latest hfh
https://api.github.com/repos/huggingface/datasets/issues/5064/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-10-07T16:59:30Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5064.diff", "html_url": "https://github.com/huggingface/datasets/pull/5064", "merged_at": "2022-10-07T16:59:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/5064.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5064" }
PR_kwDODunzps5AHsP0
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5064/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5064
https://github.com/huggingface/datasets/pull/5064
true
1,395,895,463
https://api.github.com/repos/huggingface/datasets/issues/5063/labels{/name}
This PR aligns the signature of `list_repo_files` with the current one in `hfh`, by renaming deprecated `token` to `use_auth_token`. This is already the case for `dataset_info`. CC: @lhoestq
2022-10-07T16:42:57Z
5,063
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-04T08:51:46Z
https://api.github.com/repos/huggingface/datasets/issues/5063/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5063/timeline
Align signature of list_repo_files with latest hfh
https://api.github.com/repos/huggingface/datasets/issues/5063/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-10-07T16:40:16Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5063.diff", "html_url": "https://github.com/huggingface/datasets/pull/5063", "merged_at": "2022-10-07T16:40:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/5063.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5063" }
PR_kwDODunzps5AHasG
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5063/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5063
https://github.com/huggingface/datasets/pull/5063
true
1,395,739,417
https://api.github.com/repos/huggingface/datasets/issues/5062/labels{/name}
In our CI, we get warnings from `hfh` about using deprecated `token`: https://github.com/huggingface/datasets/actions/runs/3174626525/jobs/5171672431 ``` tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_dict_to_hub_private tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_dict_to_hub tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_dict_to_hub_multiple_files tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_dict_to_hub_multiple_files_with_max_shard_size tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_dict_to_hub_overwrite_files C:\hostedtoolcache\windows\Python\3.7.9\x64\lib\site-packages\huggingface_hub\utils\_deprecation.py:97: FutureWarning: Deprecated argument(s) used in 'dataset_info': token. Will not be supported from version '0.12'. warnings.warn(message, FutureWarning) ``` This PR fixes the tests in `TestPushToHub` so that we fix these warnings. Continuation of: - #5031 CC: @lhoestq
2022-10-04T08:58:15Z
5,062
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-04T06:36:54Z
https://api.github.com/repos/huggingface/datasets/issues/5062/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5062/timeline
Fix CI hfh token warning
https://api.github.com/repos/huggingface/datasets/issues/5062/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-10-04T08:42:31Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5062.diff", "html_url": "https://github.com/huggingface/datasets/pull/5062", "merged_at": "2022-10-04T08:42:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/5062.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5062" }
PR_kwDODunzps5AG6SA
[ "_The documentation is not available anymore as the PR was closed or merged._", "good catch !" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5062/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5062
https://github.com/huggingface/datasets/pull/5062
true
1,395,476,770
https://api.github.com/repos/huggingface/datasets/issues/5061/labels{/name}
## Describe the bug When I `map` with multiple processes, this error occurs. The `.name` of the `logger` that fails to pickle in the final line is `datasets.fingerprint`. ``` File "~/project/dataset.py", line 204, in <dictcomp> split: dataset.map( File ".../site-packages/datasets/arrow_dataset.py", line 2489, in map transformed_shards[index] = async_result.get() File ".../site-packages/multiprocess/pool.py", line 771, in get raise self._value File ".../site-packages/multiprocess/pool.py", line 537, in _handle_tasks put(task) File ".../site-packages/multiprocess/connection.py", line 214, in send self._send_bytes(_ForkingPickler.dumps(obj)) File ".../site-packages/multiprocess/reduction.py", line 54, in dumps cls(buf, protocol, *args, **kwds).dump(obj) File ".../site-packages/dill/_dill.py", line 620, in dump StockPickler.dump(self, obj) File ".../pickle.py", line 487, in dump self.save(obj) File ".../pickle.py", line 560, in save f(self, obj) # Call unbound method with explicit self File ".../pickle.py", line 902, in save_tuple save(element) File ".../pickle.py", line 560, in save f(self, obj) # Call unbound method with explicit self File ".../site-packages/dill/_dill.py", line 1963, in save_function _save_with_postproc(pickler, (_create_function, ( File ".../site-packages/dill/_dill.py", line 1140, in _save_with_postproc pickler.save_reduce(*reduction, obj=obj) File ".../pickle.py", line 717, in save_reduce save(state) File ".../pickle.py", line 560, in save f(self, obj) # Call unbound method with explicit self File ".../pickle.py", line 887, in save_tuple save(element) File ".../pickle.py", line 560, in save f(self, obj) # Call unbound method with explicit self File ".../site-packages/dill/_dill.py", line 1251, in save_module_dict StockPickler.save_dict(pickler, obj) File ".../pickle.py", line 972, in save_dict self._batch_setitems(obj.items()) File ".../pickle.py", line 998, in _batch_setitems save(v) File ".../pickle.py", line 560, in save f(self, obj) # Call unbound method with explicit self File ".../site-packages/dill/_dill.py", line 1963, in save_function _save_with_postproc(pickler, (_create_function, ( File ".../site-packages/dill/_dill.py", line 1140, in _save_with_postproc pickler.save_reduce(*reduction, obj=obj) File ".../pickle.py", line 717, in save_reduce save(state) File ".../pickle.py", line 560, in save f(self, obj) # Call unbound method with explicit self File ".../pickle.py", line 887, in save_tuple save(element) File ".../pickle.py", line 560, in save f(self, obj) # Call unbound method with explicit self File ".../site-packages/dill/_dill.py", line 1251, in save_module_dict StockPickler.save_dict(pickler, obj) File ".../pickle.py", line 972, in save_dict self._batch_setitems(obj.items()) File ".../pickle.py", line 998, in _batch_setitems save(v) File ".../pickle.py", line 560, in save f(self, obj) # Call unbound method with explicit self File ".../site-packages/dill/_dill.py", line 1963, in save_function _save_with_postproc(pickler, (_create_function, ( File ".../site-packages/dill/_dill.py", line 1154, in _save_with_postproc pickler._batch_setitems(iter(source.items())) File ".../pickle.py", line 998, in _batch_setitems save(v) File ".../pickle.py", line 578, in save rv = reduce(self.proto) File ".../logging/__init__.py", line 1774, in __reduce__ raise pickle.PicklingError('logger cannot be pickled') _pickle.PicklingError: logger cannot be pickled ``` ## Steps to reproduce the bug Sorry I failed to have a minimal reproducible example, but the offending line on my end is ```python dataset.map( lambda examples: self.tokenize(examples), # this doesn't matter, lambda e: [1] * len(...) also breaks. In fact I'm pretty sure it breaks before executing this lambda batched=True, num_proc=4, ) ``` This does work when `num_proc=1`, so it's likely a multiprocessing thing. ## Expected results `map` succeeds ## Actual results The error trace above. ## Environment info - `datasets` version: 1.16.1 and 2.5.1 both failed - Platform: Ubuntu 20.04.4 LTS - Python version: 3.10.4 - PyArrow version: 9.0.0
2023-07-21T14:43:35Z
5,061
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-10-03T23:51:38Z
https://api.github.com/repos/huggingface/datasets/issues/5061/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5061/timeline
`_pickle.PicklingError: logger cannot be pickled` in multiprocessing `map`
https://api.github.com/repos/huggingface/datasets/issues/5061/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/11954789?v=4", "events_url": "https://api.github.com/users/ZhaofengWu/events{/privacy}", "followers_url": "https://api.github.com/users/ZhaofengWu/followers", "following_url": "https://api.github.com/users/ZhaofengWu/following{/other_user}", "gists_url": "https://api.github.com/users/ZhaofengWu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ZhaofengWu", "id": 11954789, "login": "ZhaofengWu", "node_id": "MDQ6VXNlcjExOTU0Nzg5", "organizations_url": "https://api.github.com/users/ZhaofengWu/orgs", "received_events_url": "https://api.github.com/users/ZhaofengWu/received_events", "repos_url": "https://api.github.com/users/ZhaofengWu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ZhaofengWu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZhaofengWu/subscriptions", "type": "User", "url": "https://api.github.com/users/ZhaofengWu" }
[]
null
completed
NONE
2023-07-21T14:43:34Z
null
I_kwDODunzps5TLUki
[ "This is maybe related to python 3.10, do you think you could try on 3.8 ?\r\n\r\nIn the meantime we'll keep improving the support for 3.10. Let me add a dedicated CI", "I did some binary search and seems like the root cause is either `multiprocess` or `dill`. python 3.10 is fine. Specifically:\r\n- `multiprocess...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5061/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5061
https://github.com/huggingface/datasets/issues/5061
false
1,395,382,940
https://api.github.com/repos/huggingface/datasets/issues/5060/labels{/name}
## Describe the bug I have uploaded a [dataset](https://huggingface.co/datasets/zpn/pubchem_selfies) and followed the instructions from the [dataset_loader](https://huggingface.co/docs/datasets/dataset_script#download-data-files-and-organize-splits) tutorial. In that tutorial, it says ``` If the data files live in the same folder or repository of the dataset script, you can just pass the relative paths to the files instead of URLs. ``` Accordingly, I put the [relative path](https://huggingface.co/datasets/zpn/pubchem_selfies/blob/main/pubchem_selfies.py#L76) to the data to be used. I was able to test the dataset and generate the metadata locally with `datasets-cli test path/to/<your-dataset-loading-script> --save_infos --all_configs` However, if I try to load the data using `load_dataset`, I get the following error ``` with gzip.open(filepath, mode="rt") as f: File "/usr/local/Cellar/python@3.9/3.9.7_1/Frameworks/Python.framework/Versions/3.9/lib/python3.9/gzip.py", line 58, in open binary_file = GzipFile(filename, gz_mode, compresslevel) File "/usr/local/Cellar/python@3.9/3.9.7_1/Frameworks/Python.framework/Versions/3.9/lib/python3.9/gzip.py", line 173, in __init__ fileobj = self.myfileobj = builtins.open(filename, mode or 'rb') FileNotFoundError: [Errno 2] No such file or directory: 'https://huggingface.co/datasets/zpn/pubchem_selfies/resolve/main/data/Compound_021000001_021500000/Compound_021000001_021500000_SELFIES.jsonl.gz' ``` ## Steps to reproduce the bug ```python >>> from datasets import load_dataset >>> dataset = load_dataset("zpn/pubchem_selfies", streaming=True) >>> t = dataset["train"] >>> for item in t: ...... print(item) ...... break Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/Users/zachnussbaum/env/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 723, in __iter__ for key, example in self._iter(): File "/Users/zachnussbaum/env/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 713, in _iter yield from ex_iterable File "/Users/zachnussbaum/env/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 113, in __iter__ yield from self.generate_examples_fn(**self.kwargs) File "/Users/zachnussbaum/.cache/huggingface/modules/datasets_modules/datasets/zpn--pubchem_selfies/d2571f35996765aea70fd3f3f8e3882d59c401fb738615c79282e2eb1d9f7a25/pubchem_selfies.py", line 475, in _generate_examples with gzip.open(filepath, mode="rt") as f: File "/usr/local/Cellar/python@3.9/3.9.7_1/Frameworks/Python.framework/Versions/3.9/lib/python3.9/gzip.py", line 58, in open binary_file = GzipFile(filename, gz_mode, compresslevel) File "/usr/local/Cellar/python@3.9/3.9.7_1/Frameworks/Python.framework/Versions/3.9/lib/python3.9/gzip.py", line 173, in __init__ fileobj = self.myfileobj = builtins.open(filename, mode or 'rb') FileNotFoundError: [Errno 2] No such file or directory: 'https://huggingface.co/datasets/zpn/pubchem_selfies/resolve/main/data/Compound_021000001_021500000/Compound_021000001_021500000_SELFIES.jsonl.gz' ```` ``` ## Expected results A clear and concise description of the expected results. ## Actual results Specify the actual results or traceback. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.5.1 - Platform: macOS-12.5.1-x86_64-i386-64bit - Python version: 3.9.7 - PyArrow version: 9.0.0 - Pandas version: 1.5.0
2022-10-06T14:29:18Z
5,060
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-10-03T21:55:16Z
https://api.github.com/repos/huggingface/datasets/issues/5060/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5060/timeline
Unable to Use Custom Dataset Locally
https://api.github.com/repos/huggingface/datasets/issues/5060/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/33707069?v=4", "events_url": "https://api.github.com/users/zanussbaum/events{/privacy}", "followers_url": "https://api.github.com/users/zanussbaum/followers", "following_url": "https://api.github.com/users/zanussbaum/following{/other_user}", "gists_url": "https://api.github.com/users/zanussbaum/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zanussbaum", "id": 33707069, "login": "zanussbaum", "node_id": "MDQ6VXNlcjMzNzA3MDY5", "organizations_url": "https://api.github.com/users/zanussbaum/orgs", "received_events_url": "https://api.github.com/users/zanussbaum/received_events", "repos_url": "https://api.github.com/users/zanussbaum/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zanussbaum/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zanussbaum/subscriptions", "type": "User", "url": "https://api.github.com/users/zanussbaum" }
[]
null
completed
CONTRIBUTOR
2022-10-06T14:29:17Z
null
I_kwDODunzps5TK9qc
[ "Hi ! I opened a PR in your repo to fix this :)\r\nhttps://huggingface.co/datasets/zpn/pubchem_selfies/discussions/7\r\n\r\nbasically you need to use `open` for streaming to work properly", "Thank you so much for this! Naive question, is this a feature of `open` or have you all overloaded it to be able to read fr...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5060/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5060
https://github.com/huggingface/datasets/issues/5060
false
1,395,050,876
https://api.github.com/repos/huggingface/datasets/issues/5059/labels{/name}
Fixes a small typo :)
2022-10-03T17:34:40Z
5,059
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-03T17:05:25Z
https://api.github.com/repos/huggingface/datasets/issues/5059/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5059/timeline
Fix typo
https://api.github.com/repos/huggingface/datasets/issues/5059/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
null
null
MEMBER
2022-10-03T17:32:27Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5059.diff", "html_url": "https://github.com/huggingface/datasets/pull/5059", "merged_at": "2022-10-03T17:32:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/5059.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5059" }
PR_kwDODunzps5AEoX7
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5059/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5059
https://github.com/huggingface/datasets/pull/5059
true
1,394,962,424
https://api.github.com/repos/huggingface/datasets/issues/5058/labels{/name}
To make CI more robust, we could mark as xfail when the Hub raises a 502 error (besides 500 error): - FAILED tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_to_hub_skip_identical_files - https://github.com/huggingface/datasets/actions/runs/3174626525/jobs/5171672431 ``` > raise HTTPError(http_error_msg, response=self) E requests.exceptions.HTTPError: 502 Server Error: Bad Gateway for url: https://hub-ci.huggingface.co/datasets/__DUMMY_TRANSFORMERS_USER__/test-16648055339047.git/info/lfs/objects/batch ``` - FAILED tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_dict_to_hub_overwrite_files - https://github.com/huggingface/datasets/actions/runs/3145587033/jobs/5113074889 ``` > raise HTTPError(http_error_msg, response=self) E requests.exceptions.HTTPError: 502 Server Error: Bad Gateway for url: https://hub-ci.huggingface.co/datasets/__DUMMY_TRANSFORMERS_USER__/test-16643866807322.git/info/lfs/objects/verify ``` Currently, we mark as xfail when 500 error: - #4845
2022-10-04T10:03:23Z
5,058
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-03T15:53:55Z
https://api.github.com/repos/huggingface/datasets/issues/5058/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5058/timeline
Mark CI tests as xfail when 502 error
https://api.github.com/repos/huggingface/datasets/issues/5058/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-10-04T10:01:23Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5058.diff", "html_url": "https://github.com/huggingface/datasets/pull/5058", "merged_at": "2022-10-04T10:01:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/5058.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5058" }
PR_kwDODunzps5AEVWn
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5058/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5058
https://github.com/huggingface/datasets/pull/5058
true
1,394,827,216
https://api.github.com/repos/huggingface/datasets/issues/5057/labels{/name}
Add the `converters` param to `CsvBuilder`, to help in situations like [this one](https://discuss.huggingface.co/t/typeerror-in-load-dataset-related-to-a-sequence-of-strings/23545).
2022-10-04T11:19:28Z
5,057
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-03T14:23:21Z
https://api.github.com/repos/huggingface/datasets/issues/5057/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5057/timeline
Support `converters` in `CsvBuilder`
https://api.github.com/repos/huggingface/datasets/issues/5057/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-10-04T11:17:32Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5057.diff", "html_url": "https://github.com/huggingface/datasets/pull/5057", "merged_at": "2022-10-04T11:17:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/5057.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5057" }
PR_kwDODunzps5AD4c6
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5057/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5057
https://github.com/huggingface/datasets/pull/5057
true
1,394,713,173
https://api.github.com/repos/huggingface/datasets/issues/5056/labels{/name}
This PR fixes the broken URL's in GEM. cc. @lhoestq, @albertvillanova
2022-10-04T13:49:00Z
5,056
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-03T13:13:22Z
https://api.github.com/repos/huggingface/datasets/issues/5056/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5056/timeline
Fix broken URL's (GEM)
https://api.github.com/repos/huggingface/datasets/issues/5056/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6687858?v=4", "events_url": "https://api.github.com/users/manandey/events{/privacy}", "followers_url": "https://api.github.com/users/manandey/followers", "following_url": "https://api.github.com/users/manandey/following{/other_user}", "gists_url": "https://api.github.com/users/manandey/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/manandey", "id": 6687858, "login": "manandey", "node_id": "MDQ6VXNlcjY2ODc4NTg=", "organizations_url": "https://api.github.com/users/manandey/orgs", "received_events_url": "https://api.github.com/users/manandey/received_events", "repos_url": "https://api.github.com/users/manandey/repos", "site_admin": false, "starred_url": "https://api.github.com/users/manandey/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/manandey/subscriptions", "type": "User", "url": "https://api.github.com/users/manandey" }
[]
null
null
CONTRIBUTOR
2022-10-04T13:48:59Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5056.diff", "html_url": "https://github.com/huggingface/datasets/pull/5056", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5056.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5056" }
PR_kwDODunzps5ADfxN
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5056). All of your documentation changes will be reflected on that endpoint.", "Thanks, @manandey. We have removed all dataset scripts from this repo. Subsequent PRs should be opened directly on the Hugging Face Hub." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5056/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5056
https://github.com/huggingface/datasets/pull/5056
true
1,394,503,844
https://api.github.com/repos/huggingface/datasets/issues/5055/labels{/name}
While working on https://github.com/huggingface/datasets/pull/5018 I noticed a small bug introduced in #4926 regarding backward compatibility for dataset_infos.json Indeed, when a dataset repo had both dataset_infos.json and README.md, the JSON file was ignored. This is unexpected: in practice it should be ignored only if the README.md has a dataset_info field, which has precedence over the data in the JSON file.
2022-10-03T13:43:55Z
5,055
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-10-03T10:30:14Z
https://api.github.com/repos/huggingface/datasets/issues/5055/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5055/timeline
Fix backward compatibility for dataset_infos.json
https://api.github.com/repos/huggingface/datasets/issues/5055/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-10-03T13:41:32Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5055.diff", "html_url": "https://github.com/huggingface/datasets/pull/5055", "merged_at": "2022-10-03T13:41:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/5055.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5055" }
PR_kwDODunzps5ACyVU
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5055/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5055
https://github.com/huggingface/datasets/pull/5055
true
1,394,152,728
https://api.github.com/repos/huggingface/datasets/issues/5054/labels{/name}
This PR fixes the license/citation information of squadshifts dataset card, once the dataset owners have responded to our request for information: - https://github.com/modestyachts/squadshifts-website/issues/1 Additionally, we have updated the mention in their website to our `datasets` library (they were referring old name `nlp`): - https://github.com/modestyachts/squadshifts-website/pull/2#event-7500953009
2022-10-03T09:26:49Z
5,054
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-10-03T05:19:13Z
https://api.github.com/repos/huggingface/datasets/issues/5054/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5054/timeline
Fix license/citation information of squadshifts dataset card
https://api.github.com/repos/huggingface/datasets/issues/5054/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-10-03T09:24:30Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5054.diff", "html_url": "https://github.com/huggingface/datasets/pull/5054", "merged_at": "2022-10-03T09:24:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/5054.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5054" }
PR_kwDODunzps5ABnd3
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5054/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5054
https://github.com/huggingface/datasets/pull/5054
true
1,393,739,882
https://api.github.com/repos/huggingface/datasets/issues/5053/labels{/name}
## Describe the bug I have an intermittent error when streaming the Pile, where I get a JSON parse error which causes my program to crash. This is intermittent - when I rerun the program with the same random seed it does not crash in the same way. The exact point this happens also varied - it happened to me 11B tokens and 4 days into a training run, and now just happened 2 minutes into one, but I can't reliably reproduce it. I'm using a remote machine with 8 A6000 GPUs via runpod.io ## Expected results I have a DataLoader which can iterate through the whole Pile ## Actual results Stack trace: ``` FailedΒ toΒ readΒ fileΒ 'zstd://12.jsonl::https://the-eye.eu/public/AI/pile/train/12.jsonl.zst'Β withΒ errorΒ <classΒ 'pyarrow.lib.ArrowInvalid'>:Β JSONΒ parseΒ error:Β InvalidΒ value.Β inΒ rowΒ 0 ``` I'm currently using HuggingFace accelerate, which also gave me the following stack trace, but I've also experienced this problem intermittently when using DataParallel, so I don't think it's to do with parallelisation ``` Traceback (most recent call last): File "ddp_script.py", line 1258, in <module> main() File "ddp_script.py", line 1143, in main for c, batch in tqdm.tqdm(enumerate(data_iter)): File "/opt/conda/lib/python3.7/site-packages/tqdm/std.py", line 1195, in __iter__ for obj in iterable: File "/opt/conda/lib/python3.7/site-packages/accelerate/data_loader.py", line 503, in __iter__ next_batch, next_batch_info, next_skip = self._fetch_batches(main_iterator) File "/opt/conda/lib/python3.7/site-packages/accelerate/data_loader.py", line 454, in _fetch_batches broadcast_object_list(batch_info) File "/opt/conda/lib/python3.7/site-packages/accelerate/utils/operations.py", line 333, in broadcast_object_list torch.distributed.broadcast_object_list(object_list, src=from_process) File "/opt/conda/lib/python3.7/site-packages/torch/distributed/distributed_c10d.py", line 1900, in broadcast_object_list object_list[i] = _tensor_to_object(obj_view, obj_size) File "/opt/conda/lib/python3.7/site-packages/torch/distributed/distributed_c10d.py", line 1571, in _tensor_to_object return _unpickler(io.BytesIO(buf)).load() _pickle.UnpicklingError: invalid load key, '@'. ``` ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset( cfg["dataset_name"], streaming=True, split="train") dataset = dataset.remove_columns("meta") dataset = dataset.map(tokenize_and_concatenate, batched=True) dataset = dataset.with_format(type="torch") train_data_loader = DataLoader( dataset, batch_size=cfg["batch_size"], num_workers=3) for batch in train_data_loader: continue ``` `tokenize_and_concatenate` is a custom tokenization function I defined on the GPT-NeoX tokenizer to tokenize the text, separated by endoftext tokens, and reshape to have length batch_size, I don't think this is related to tokenization: ``` import numpy as np import einops import torch def tokenize_and_concatenate(examples): texts = examples["text"] full_text = tokenizer.eos_token.join(texts) div = 20 length = len(full_text) // div text_list = [full_text[i * length: (i + 1) * length] for i in range(div)] tokens = tokenizer(text_list, return_tensors="np", padding=True)[ "input_ids" ].flatten() tokens = tokens[tokens != tokenizer.pad_token_id] n = len(tokens) curr_batch_size = n // (seq_len - 1) tokens = tokens[: (seq_len - 1) * curr_batch_size] tokens = einops.rearrange( tokens, "(batch_size seq) -> batch_size seq", batch_size=curr_batch_size, seq=seq_len - 1, ) prefix = np.ones((curr_batch_size, 1), dtype=np.int64) * \ tokenizer.bos_token_id return { "text": np.concatenate([prefix, tokens], axis=1) } ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.4.0 - Platform: Linux-5.4.0-105-generic-x86_64-with-debian-buster-sid - Python version: 3.7.13 - PyArrow version: 9.0.0 - Pandas version: 1.3.5 ZStandard data: Version: 0.18.0 Summary: Zstandard bindings for Python Home-page: https://github.com/indygreg/python-zstandard Author: Gregory Szorc Author-email: gregory.szorc@gmail.com License: BSD Location: /opt/conda/lib/python3.7/site-packages Requires: Required-by:
2022-10-04T17:59:03Z
5,053
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-10-02T11:56:46Z
https://api.github.com/repos/huggingface/datasets/issues/5053/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5053/timeline
Intermittent JSON parse error when streaming the Pile
https://api.github.com/repos/huggingface/datasets/issues/5053/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/77788841?v=4", "events_url": "https://api.github.com/users/neelnanda-io/events{/privacy}", "followers_url": "https://api.github.com/users/neelnanda-io/followers", "following_url": "https://api.github.com/users/neelnanda-io/following{/other_user}", "gists_url": "https://api.github.com/users/neelnanda-io/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/neelnanda-io", "id": 77788841, "login": "neelnanda-io", "node_id": "MDQ6VXNlcjc3Nzg4ODQx", "organizations_url": "https://api.github.com/users/neelnanda-io/orgs", "received_events_url": "https://api.github.com/users/neelnanda-io/received_events", "repos_url": "https://api.github.com/users/neelnanda-io/repos", "site_admin": false, "starred_url": "https://api.github.com/users/neelnanda-io/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neelnanda-io/subscriptions", "type": "User", "url": "https://api.github.com/users/neelnanda-io" }
[]
null
null
NONE
null
null
I_kwDODunzps5TEshq
[ "Maybe #2838 can help. In this PR we allow to skip bad chunks of JSON data to not crash the training\r\n\r\nDid you have warning messages before the error ?\r\n\r\nsomething like this maybe ?\r\n```\r\n03/24/2022 02:19:46 - WARNING - datasets.utils.streaming_download_manager - Got disconnected from remote data host...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5053/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5053
https://github.com/huggingface/datasets/issues/5053
false
1,393,076,765
https://api.github.com/repos/huggingface/datasets/issues/5052/labels{/name}
Hello, This resolves issues #4988. I added a method `from_generator` to class `IterableDataset`. I modified the `read` method of input stream generator to also return Iterable_dataset.
2022-10-05T12:51:48Z
5,052
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-30T22:14:05Z
https://api.github.com/repos/huggingface/datasets/issues/5052/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5052/timeline
added from_generator method to IterableDataset class.
https://api.github.com/repos/huggingface/datasets/issues/5052/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/56002455?v=4", "events_url": "https://api.github.com/users/hamid-vakilzadeh/events{/privacy}", "followers_url": "https://api.github.com/users/hamid-vakilzadeh/followers", "following_url": "https://api.github.com/users/hamid-vakilzadeh/following{/other_user}", "gists_url": "https://api.github.com/users/hamid-vakilzadeh/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hamid-vakilzadeh", "id": 56002455, "login": "hamid-vakilzadeh", "node_id": "MDQ6VXNlcjU2MDAyNDU1", "organizations_url": "https://api.github.com/users/hamid-vakilzadeh/orgs", "received_events_url": "https://api.github.com/users/hamid-vakilzadeh/received_events", "repos_url": "https://api.github.com/users/hamid-vakilzadeh/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hamid-vakilzadeh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hamid-vakilzadeh/subscriptions", "type": "User", "url": "https://api.github.com/users/hamid-vakilzadeh" }
[]
null
null
CONTRIBUTOR
2022-10-05T12:10:48Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5052.diff", "html_url": "https://github.com/huggingface/datasets/pull/5052", "merged_at": "2022-10-05T12:10:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/5052.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5052" }
PR_kwDODunzps4_-PZw
[ "_The documentation is not available anymore as the PR was closed or merged._", "I added a test and moved the `streaming` param from `read` to `__init_`. Then, I also decided to update the `read` method of the rest of the packaged modules to account for this param. \r\n\r\n@hamid-vakilzadeh Are you OK with these ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5052/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5052
https://github.com/huggingface/datasets/pull/5052
true
1,392,559,503
https://api.github.com/repos/huggingface/datasets/issues/5051/labels{/name}
Reverts the removal of `task_templates` in the folder-based builders. I also added the `AudioClassifaction` task for consistency. This is needed to fix https://github.com/huggingface/transformers/issues/19177. I think we should soon deprecate and remove the current task API (and investigate if it's possible to integrate the `train eval index` API), but we need to update the Transformers examples before that so we don't break them. cc @NielsRogge
2022-10-03T12:23:35Z
5,051
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-30T14:50:03Z
https://api.github.com/repos/huggingface/datasets/issues/5051/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5051/timeline
Revert task removal in folder-based builders
https://api.github.com/repos/huggingface/datasets/issues/5051/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-10-03T12:21:31Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5051.diff", "html_url": "https://github.com/huggingface/datasets/pull/5051", "merged_at": "2022-10-03T12:21:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/5051.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5051" }
PR_kwDODunzps4_8drw
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5051/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5051
https://github.com/huggingface/datasets/pull/5051
true
1,392,381,882
https://api.github.com/repos/huggingface/datasets/issues/5050/labels{/name}
Even though we save the `format` state in `save_to_disk`, we don't restore it in `load_from_disk`. We should fix that. Reported here: https://discuss.huggingface.co/t/save-to-disk-loses-formatting-information/23815
2022-10-11T16:49:24Z
5,050
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "7057ff", "default": true, "descript...
2022-09-30T12:40:07Z
https://api.github.com/repos/huggingface/datasets/issues/5050/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/74454835?v=4", "events_url": "https://api.github.com/users/asofiaoliveira/events{/privacy}", "followers_url": "https://api.github.com/users/asofiaoliveira/followers", "following_url": "https://api.github.com/users/asofiaoliveira/following{/other_user}", "gists_url": "https://api.github.com/users/asofiaoliveira/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/asofiaoliveira", "id": 74454835, "login": "asofiaoliveira", "node_id": "MDQ6VXNlcjc0NDU0ODM1", "organizations_url": "https://api.github.com/users/asofiaoliveira/orgs", "received_events_url": "https://api.github.com/users/asofiaoliveira/received_events", "repos_url": "https://api.github.com/users/asofiaoliveira/repos", "site_admin": false, "starred_url": "https://api.github.com/users/asofiaoliveira/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/asofiaoliveira/subscriptions", "type": "User", "url": "https://api.github.com/users/asofiaoliveira" }
https://api.github.com/repos/huggingface/datasets/issues/5050/timeline
Restore saved format state in `load_from_disk`
https://api.github.com/repos/huggingface/datasets/issues/5050/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/74454835?v=4", "events_url": "https://api.github.com/users/asofiaoliveira/events{/privacy}", "followers_url": "https://api.github.com/users/asofiaoliveira/followers", "following_url": "https://api.github.com/users/asofiaoliveira/following{/oth...
null
completed
CONTRIBUTOR
2022-10-11T16:49:24Z
null
I_kwDODunzps5S_g-6
[ "Hi, can I work on this?", "Hi, sure! Let us know if you need some pointers/help." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5050/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5050
https://github.com/huggingface/datasets/issues/5050
false
1,392,361,381
https://api.github.com/repos/huggingface/datasets/issues/5049/labels{/name}
Add the `kwargs` param to `from_generator` to align it with the rest of the `from_` methods (this param allows passing custom `writer_batch_size` for instance).
2022-10-03T11:00:11Z
5,049
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-30T12:24:27Z
https://api.github.com/repos/huggingface/datasets/issues/5049/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5049/timeline
Add `kwargs` to `Dataset.from_generator`
https://api.github.com/repos/huggingface/datasets/issues/5049/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-10-03T10:58:15Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5049.diff", "html_url": "https://github.com/huggingface/datasets/pull/5049", "merged_at": "2022-10-03T10:58:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/5049.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5049" }
PR_kwDODunzps4_7zOY
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5049/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5049
https://github.com/huggingface/datasets/pull/5049
true
1,392,170,680
https://api.github.com/repos/huggingface/datasets/issues/5048/labels{/name}
Fix for a critical bug in the EURLEX dataset label list to make LexGLUE EURLEX results replicable. In LexGLUE (Chalkidis et al., 2022), the following is mentioned w.r.t. EUR-LEX: _"It supports four different label granularities, comprising 21, 127, 567, 7390 EuroVoc concepts, respectively. We use the 100 most frequent concepts from level 2 [...]”._ The current label list has all 127 labels, which leads to different (lower) results, as communicated by users. Thanks!
2022-09-30T16:30:25Z
5,048
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-30T09:47:12Z
https://api.github.com/repos/huggingface/datasets/issues/5048/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5048/timeline
Fix bug with labels of eurlex config of lex_glue dataset
https://api.github.com/repos/huggingface/datasets/issues/5048/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1626984?v=4", "events_url": "https://api.github.com/users/iliaschalkidis/events{/privacy}", "followers_url": "https://api.github.com/users/iliaschalkidis/followers", "following_url": "https://api.github.com/users/iliaschalkidis/following{/other_user}", "gists_url": "https://api.github.com/users/iliaschalkidis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/iliaschalkidis", "id": 1626984, "login": "iliaschalkidis", "node_id": "MDQ6VXNlcjE2MjY5ODQ=", "organizations_url": "https://api.github.com/users/iliaschalkidis/orgs", "received_events_url": "https://api.github.com/users/iliaschalkidis/received_events", "repos_url": "https://api.github.com/users/iliaschalkidis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/iliaschalkidis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iliaschalkidis/subscriptions", "type": "User", "url": "https://api.github.com/users/iliaschalkidis" }
[]
null
null
CONTRIBUTOR
2022-09-30T16:21:41Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5048.diff", "html_url": "https://github.com/huggingface/datasets/pull/5048", "merged_at": "2022-09-30T16:21:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/5048.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5048" }
PR_kwDODunzps4_7KI2
[ "_The documentation is not available anymore as the PR was closed or merged._", "@JamesLYC88 here is the fix! Thanks again!", "Thanks, @albertvillanova. When do you expect that this change will take effect when someone downloads the dataset?", "The change is immediately available now, since this change we mad...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5048/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5048
https://github.com/huggingface/datasets/pull/5048
true
1,392,088,398
https://api.github.com/repos/huggingface/datasets/issues/5047/labels{/name}
Reported in https://github.com/huggingface/datasets/pull/3878 I updated the number of examples
2022-09-30T10:23:22Z
5,047
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-30T08:47:29Z
https://api.github.com/repos/huggingface/datasets/issues/5047/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5047/timeline
Fix cats_vs_dogs
https://api.github.com/repos/huggingface/datasets/issues/5047/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-30T09:34:28Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5047.diff", "html_url": "https://github.com/huggingface/datasets/pull/5047", "merged_at": "2022-09-30T09:34:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/5047.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5047" }
PR_kwDODunzps4_64bS
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5047/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5047
https://github.com/huggingface/datasets/pull/5047
true
1,391,372,519
https://api.github.com/repos/huggingface/datasets/issues/5046/labels{/name}
## Describe the bug When audio files are at the same level as the metadata (`metadata.csv` or `metadata.jsonl` ), the `load_dataset` returns a `DatasetDict` with no rows but the correct columns. https://github.com/huggingface/datasets/blob/1ea4d091b7a4b83a85b2eeb8df65115d39af3766/docs/source/audio_dataset.mdx?plain=1#L88 ## Steps to reproduce the bug `metadata.csv`: ```csv file_name,duration,transcription ./2063_fe9936e7-62b2-4e62-a276-acbd344480ce_1.wav,10.768,hello ``` ```python >>> audio_dataset = load_dataset("audiofolder", data_dir="/audio-data/") >>> audio_dataset DatasetDict({ train: Dataset({ features: ['audio', 'duration', 'transcription'], num_rows: 0 }) validation: Dataset({ features: ['audio', 'duration', 'transcription'], num_rows: 0 }) }) ``` I've tried, with no success,: - setting `split` to something else so I don't get a `DatasetDict`, - removing the `./`, - using `.jsonl`. ## Expected results ``` Dataset({ features: ['audio', 'duration', 'transcription'], num_rows: 1 }) ``` ## Actual results ``` DatasetDict({ train: Dataset({ features: ['audio', 'duration', 'transcription'], num_rows: 0 }) validation: Dataset({ features: ['audio', 'duration', 'transcription'], num_rows: 0 }) }) ``` ## Environment info - `datasets` version: 2.5.1 - Platform: Linux-5.13.0-1025-aws-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 9.0.0 - Pandas version: 1.5.0
2022-10-28T13:05:07Z
5,046
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "7057ff", "default": true, "descript...
2022-09-29T19:17:23Z
https://api.github.com/repos/huggingface/datasets/issues/5046/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/9295277?v=4", "events_url": "https://api.github.com/users/riccardobucco/events{/privacy}", "followers_url": "https://api.github.com/users/riccardobucco/followers", "following_url": "https://api.github.com/users/riccardobucco/following{/other_user}", "gists_url": "https://api.github.com/users/riccardobucco/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/riccardobucco", "id": 9295277, "login": "riccardobucco", "node_id": "MDQ6VXNlcjkyOTUyNzc=", "organizations_url": "https://api.github.com/users/riccardobucco/orgs", "received_events_url": "https://api.github.com/users/riccardobucco/received_events", "repos_url": "https://api.github.com/users/riccardobucco/repos", "site_admin": false, "starred_url": "https://api.github.com/users/riccardobucco/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/riccardobucco/subscriptions", "type": "User", "url": "https://api.github.com/users/riccardobucco" }
https://api.github.com/repos/huggingface/datasets/issues/5046/timeline
Audiofolder creates empty Dataset if files same level as metadata
https://api.github.com/repos/huggingface/datasets/issues/5046/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/577139?v=4", "events_url": "https://api.github.com/users/msis/events{/privacy}", "followers_url": "https://api.github.com/users/msis/followers", "following_url": "https://api.github.com/users/msis/following{/other_user}", "gists_url": "https://api.github.com/users/msis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/msis", "id": 577139, "login": "msis", "node_id": "MDQ6VXNlcjU3NzEzOQ==", "organizations_url": "https://api.github.com/users/msis/orgs", "received_events_url": "https://api.github.com/users/msis/received_events", "repos_url": "https://api.github.com/users/msis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/msis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/msis/subscriptions", "type": "User", "url": "https://api.github.com/users/msis" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/9295277?v=4", "events_url": "https://api.github.com/users/riccardobucco/events{/privacy}", "followers_url": "https://api.github.com/users/riccardobucco/followers", "following_url": "https://api.github.com/users/riccardobucco/following{/other_u...
null
completed
NONE
2022-10-28T13:05:07Z
null
I_kwDODunzps5S7qjn
[ "Hi! Unfortunately, I can't reproduce this behavior. Instead, I get `ValueError: audio at 2063_fe9936e7-62b2-4e62-a276-acbd344480ce_1.wav doesn't have metadata in /audio-data/metadata.csv`, which can be fixed by removing the `./` from the file name.\r\n\r\n(Link to a Colab that tries to reproduce this behavior: htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5046/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5046
https://github.com/huggingface/datasets/issues/5046
false
1,391,287,609
https://api.github.com/repos/huggingface/datasets/issues/5045/labels{/name}
**Is your feature request related to a problem? Please describe.** I pushed a modification of a large dataset (remove a column) to the hub. The push was interrupted after some files were committed to the repo. This left the dataset to raise an error on load_dataset() (ValueError couldn’t cast … because column names don’t match). Only by specifying the previous (complete) commit as revision=commit_hash in load_data(), I was able to repair this and after a successful, complete push, the dataset loads without error again. **Describe the solution you'd like** Would it make sense to detect an incomplete push_to_hub() and automatically revert to the previous commit/revision? **Describe alternatives you've considered** Leave everything as is, the revision parameter in load_dataset() allows to manually fix this problem. **Additional context** Provide useful defaults
2023-10-16T13:30:49Z
5,045
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-29T18:08:12Z
https://api.github.com/repos/huggingface/datasets/issues/5045/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5045/timeline
Automatically revert to last successful commit to hub when a push_to_hub is interrupted
https://api.github.com/repos/huggingface/datasets/issues/5045/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/13120204?v=4", "events_url": "https://api.github.com/users/jorahn/events{/privacy}", "followers_url": "https://api.github.com/users/jorahn/followers", "following_url": "https://api.github.com/users/jorahn/following{/other_user}", "gists_url": "https://api.github.com/users/jorahn/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jorahn", "id": 13120204, "login": "jorahn", "node_id": "MDQ6VXNlcjEzMTIwMjA0", "organizations_url": "https://api.github.com/users/jorahn/orgs", "received_events_url": "https://api.github.com/users/jorahn/received_events", "repos_url": "https://api.github.com/users/jorahn/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jorahn/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jorahn/subscriptions", "type": "User", "url": "https://api.github.com/users/jorahn" }
[]
null
completed
NONE
2023-10-16T13:30:49Z
null
I_kwDODunzps5S7V05
[ "Could you share the error you got please ? Maybe the full stack trace if you have it ?\r\n\r\nMaybe `push_to_hub` be implemented as a single commit @Wauplin ? This way if it fails, the repo is still at the previous (valid) state instead of ending-up in an invalid/incimplete state.", "> Maybe push_to_hub be imple...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5045/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5045
https://github.com/huggingface/datasets/issues/5045
false
1,391,242,908
https://api.github.com/repos/huggingface/datasets/issues/5044/labels{/name}
**Is your feature request related to a problem? Please describe.** Is it possible to make `load_dataset` more universal similar to `from_pretrained` in `transformers` so that it can handle the hub, and the local path datasets of all supported types? Currently one has to choose a different loader depending on how the dataset has been created. e.g. this won't work: ``` $ git clone https://huggingface.co/datasets/severo/test-parquet $ python -c 'from datasets import load_dataset; ds=load_dataset("test-parquet"); \ ds.save_to_disk("my_dataset"); load_dataset("my_dataset")' [...] Traceback (most recent call last): File "<string>", line 1, in <module> File "/home/stas/anaconda3/envs/py38-pt112/lib/python3.8/site-packages/datasets/load.py", line 1746, in load_dataset builder_instance.download_and_prepare( File "/home/stas/anaconda3/envs/py38-pt112/lib/python3.8/site-packages/datasets/builder.py", line 704, in download_and_prepare self._download_and_prepare( File "/home/stas/anaconda3/envs/py38-pt112/lib/python3.8/site-packages/datasets/builder.py", line 793, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/stas/anaconda3/envs/py38-pt112/lib/python3.8/site-packages/datasets/builder.py", line 1277, in _prepare_split writer.write_table(table) File "/home/stas/anaconda3/envs/py38-pt112/lib/python3.8/site-packages/datasets/arrow_writer.py", line 524, in write_table pa_table = table_cast(pa_table, self._schema) File "/home/stas/anaconda3/envs/py38-pt112/lib/python3.8/site-packages/datasets/table.py", line 2005, in table_cast return cast_table_to_schema(table, schema) File "/home/stas/anaconda3/envs/py38-pt112/lib/python3.8/site-packages/datasets/table.py", line 1968, in cast_table_to_schema raise ValueError(f"Couldn't cast\n{table.schema}\nto\n{features}\nbecause column names don't match") ValueError: Couldn't cast _data_files: list<item: struct<filename: string>> child 0, item: struct<filename: string> child 0, filename: string ``` both times the dataset is being loaded from disk. Why does it fail the second time? Why can't `save_to_disk` generate a dataset that can be immediately loaded by `load_dataset`? e.g. the simplest hack would be to have `save_to_disk` add some flag to the saved dataset, that tells `load_dataset` to internally call `load_from_disk`. like having `save_to_disk` create a `load_me_with_load_from_disk.txt` file ;) and `load_dataset` will support that feature from saved datasets from new `datasets` versions. The old ones will still need to use `load_from_disk` explicitly. Unless the flag is not needed and one can immediately tell by looking at the saved dataset that it was saved via `save_to_disk` and thus use `load_from_disk` internally. The use-case is defining a simple API where the user only ever needs to pass a `dataset_name_or_path` and it will always just work. Currently one needs to manually add additional switches telling the system whether to use one loading method or the other which works but it's not smooth. Thank you!
2024-02-12T15:03:27Z
5,044
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-29T17:37:12Z
https://api.github.com/repos/huggingface/datasets/issues/5044/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5044/timeline
integrate `load_from_disk` into `load_dataset`
https://api.github.com/repos/huggingface/datasets/issues/5044/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stas00", "id": 10676103, "login": "stas00", "node_id": "MDQ6VXNlcjEwNjc2MTAz", "organizations_url": "https://api.github.com/users/stas00/orgs", "received_events_url": "https://api.github.com/users/stas00/received_events", "repos_url": "https://api.github.com/users/stas00/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "type": "User", "url": "https://api.github.com/users/stas00" }
[]
null
null
CONTRIBUTOR
null
null
I_kwDODunzps5S7K6c
[ "I agree the situation is not ideal and it would be awesome to use `load_dataset` to reload a dataset saved locally !\r\n\r\nFor context:\r\n\r\n- `load_dataset` works in three steps: download the dataset, then prepare it as an arrow dataset, and finally return a memory mapped arrow dataset. In particular it create...
{ "+1": 4, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 4, "url": "https://api.github.com/repos/huggingface/datasets/issues/5044/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5044
https://github.com/huggingface/datasets/issues/5044
false
1,391,141,773
https://api.github.com/repos/huggingface/datasets/issues/5043/labels{/name}
Fix #5038
2022-09-30T15:46:39Z
5,043
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-29T16:17:28Z
https://api.github.com/repos/huggingface/datasets/issues/5043/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5043/timeline
Fix `flatten_indices` with empty indices mapping
https://api.github.com/repos/huggingface/datasets/issues/5043/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-09-30T15:44:25Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5043.diff", "html_url": "https://github.com/huggingface/datasets/pull/5043", "merged_at": "2022-09-30T15:44:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/5043.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5043" }
PR_kwDODunzps4_3uzy
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5043/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5043
https://github.com/huggingface/datasets/pull/5043
true
1,390,762,877
https://api.github.com/repos/huggingface/datasets/issues/5042/labels{/name}
I forgot to add the new citation.
2022-09-30T07:14:00Z
5,042
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-29T12:10:02Z
https://api.github.com/repos/huggingface/datasets/issues/5042/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5042/timeline
Update swiss judgment prediction
https://api.github.com/repos/huggingface/datasets/issues/5042/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3775944?v=4", "events_url": "https://api.github.com/users/JoelNiklaus/events{/privacy}", "followers_url": "https://api.github.com/users/JoelNiklaus/followers", "following_url": "https://api.github.com/users/JoelNiklaus/following{/other_user}", "gists_url": "https://api.github.com/users/JoelNiklaus/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JoelNiklaus", "id": 3775944, "login": "JoelNiklaus", "node_id": "MDQ6VXNlcjM3NzU5NDQ=", "organizations_url": "https://api.github.com/users/JoelNiklaus/orgs", "received_events_url": "https://api.github.com/users/JoelNiklaus/received_events", "repos_url": "https://api.github.com/users/JoelNiklaus/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JoelNiklaus/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JoelNiklaus/subscriptions", "type": "User", "url": "https://api.github.com/users/JoelNiklaus" }
[]
null
null
CONTRIBUTOR
2022-09-29T14:32:02Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5042.diff", "html_url": "https://github.com/huggingface/datasets/pull/5042", "merged_at": "2022-09-29T14:32:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/5042.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5042" }
PR_kwDODunzps4_2eqa
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5042/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5042
https://github.com/huggingface/datasets/pull/5042
true
1,390,722,230
https://api.github.com/repos/huggingface/datasets/issues/5041/labels{/name}
This PR: - supports streaming - fixes the description section of the dataset card
2022-09-30T07:13:38Z
5,041
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-29T11:37:58Z
https://api.github.com/repos/huggingface/datasets/issues/5041/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5041/timeline
Support streaming hendrycks_test dataset.
https://api.github.com/repos/huggingface/datasets/issues/5041/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-29T12:07:29Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5041.diff", "html_url": "https://github.com/huggingface/datasets/pull/5041", "merged_at": "2022-09-29T12:07:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/5041.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5041" }
PR_kwDODunzps4_2WES
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5041/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5041
https://github.com/huggingface/datasets/pull/5041
true
1,390,566,428
https://api.github.com/repos/huggingface/datasets/issues/5040/labels{/name}
Update metadata JSON. Fix #5039.
2022-09-29T10:06:22Z
5,040
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-29T09:37:43Z
https://api.github.com/repos/huggingface/datasets/issues/5040/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5040/timeline
Fix NonMatchingChecksumError in hendrycks_test dataset
https://api.github.com/repos/huggingface/datasets/issues/5040/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-29T10:04:19Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5040.diff", "html_url": "https://github.com/huggingface/datasets/pull/5040", "merged_at": "2022-09-29T10:04:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/5040.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5040" }
PR_kwDODunzps4_11O2
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5040/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5040
https://github.com/huggingface/datasets/pull/5040
true
1,390,353,315
https://api.github.com/repos/huggingface/datasets/issues/5039/labels{/name}
Hi, The checksum for [hendrycks_test](https://huggingface.co/datasets/hendrycks_test) does not compare correctly, I guess it has been updated on the remote. ``` datasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://people.eecs.berkeley.edu/~hendrycks/data.tar'] ```
2022-09-29T10:23:30Z
5,039
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
2022-09-29T06:56:20Z
https://api.github.com/repos/huggingface/datasets/issues/5039/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5039/timeline
Hendrycks Checksum
https://api.github.com/repos/huggingface/datasets/issues/5039/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/9974388?v=4", "events_url": "https://api.github.com/users/DanielHesslow/events{/privacy}", "followers_url": "https://api.github.com/users/DanielHesslow/followers", "following_url": "https://api.github.com/users/DanielHesslow/following{/other_user}", "gists_url": "https://api.github.com/users/DanielHesslow/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/DanielHesslow", "id": 9974388, "login": "DanielHesslow", "node_id": "MDQ6VXNlcjk5NzQzODg=", "organizations_url": "https://api.github.com/users/DanielHesslow/orgs", "received_events_url": "https://api.github.com/users/DanielHesslow/received_events", "repos_url": "https://api.github.com/users/DanielHesslow/repos", "site_admin": false, "starred_url": "https://api.github.com/users/DanielHesslow/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DanielHesslow/subscriptions", "type": "User", "url": "https://api.github.com/users/DanielHesslow" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
NONE
2022-09-29T10:04:20Z
null
I_kwDODunzps5S3xuj
[ "Thanks for reporting, @DanielHesslow. We are fixing it. ", "@albertvillanova thanks for taking care of this so quickly!", "The dataset metadata is fixed. You can download it normally." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5039/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5039
https://github.com/huggingface/datasets/issues/5039
false
1,389,631,122
https://api.github.com/repos/huggingface/datasets/issues/5038/labels{/name}
## Describe the bug After filtering a dataset, and if no samples remain, `Dataset.unique` will return the unique values of the unfiltered dataset. ## Steps to reproduce the bug ```python from datasets import Dataset dataset = Dataset.from_dict({'id': [0]}) dataset = dataset.filter(lambda _: False) print(dataset.unique('id')) ``` ## Expected results The above code should return an empty list since the dataset is empty. ## Actual results ```bash [0] ``` ## Environment info - `datasets` version: 2.5.1 - Platform: Linux-5.18.19-100.fc35.x86_64-x86_64-with-glibc2.34 - Python version: 3.9.14 - PyArrow version: 7.0.0 - Pandas version: 1.3.5
2022-09-30T15:44:25Z
5,038
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-28T16:20:35Z
https://api.github.com/repos/huggingface/datasets/issues/5038/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5038/timeline
`Dataset.unique` showing wrong output after filtering
https://api.github.com/repos/huggingface/datasets/issues/5038/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/4904985?v=4", "events_url": "https://api.github.com/users/mxschmdt/events{/privacy}", "followers_url": "https://api.github.com/users/mxschmdt/followers", "following_url": "https://api.github.com/users/mxschmdt/following{/other_user}", "gists_url": "https://api.github.com/users/mxschmdt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mxschmdt", "id": 4904985, "login": "mxschmdt", "node_id": "MDQ6VXNlcjQ5MDQ5ODU=", "organizations_url": "https://api.github.com/users/mxschmdt/orgs", "received_events_url": "https://api.github.com/users/mxschmdt/received_events", "repos_url": "https://api.github.com/users/mxschmdt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mxschmdt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxschmdt/subscriptions", "type": "User", "url": "https://api.github.com/users/mxschmdt" }
[]
null
completed
CONTRIBUTOR
2022-09-30T15:44:25Z
null
I_kwDODunzps5S1BaS
[ "Hi! It seems like `flatten_indices` (called in `unique`) doesn't know how to handle empty indices mappings. I'm working on the fix.", "Thanks, that was fast!" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5038/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5038
https://github.com/huggingface/datasets/issues/5038
false
1,389,244,722
https://api.github.com/repos/huggingface/datasets/issues/5037/labels{/name}
This PR improves PackagedDatasetTest CI performance speed. For Ubuntu (latest): - Duration (without parallelism) before: 334.78s (5.58m) - Duration (without parallelism) afterwards: 0.48s The approach is passing a dummy `data_files` argument to load the builder, so that it avoids the slow inferring of it over the entire root directory of the repo. ## Total duration of PackagedDatasetTest | | Before | Afterwards | Improvement |---|---:|---:|---:| | Linux | 334.78s | 0.48s | x700 | Windows | 513.02s | 1.09s | x500 ## Durations by each individual sub-test More accurate durations, running them on GitHub, for Linux (latest). Before this PR, the total test time (without parallelism) for `tests/test_dataset_common.py::PackagedDatasetTest` is 334.78s (5.58m) ``` 39.07s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_imagefolder 38.94s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_audiofolder 34.18s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_parquet 34.12s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_csv 34.00s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_pandas 34.00s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_text 33.86s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_json 10.39s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_class_audiofolder 6.50s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_configs_audiofolder 6.46s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_configs_imagefolder 6.40s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_class_imagefolder 5.77s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_class_csv 5.77s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_class_text 5.74s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_configs_parquet 5.69s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_class_json 5.68s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_configs_pandas 5.67s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_class_parquet 5.67s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_class_pandas 5.66s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_configs_json 5.66s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_configs_csv 5.55s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_configs_text (42 durations < 0.005s hidden.) ``` With this PR: 0.48s ``` 0.09s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_audiofolder 0.08s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_csv 0.08s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_imagefolder 0.06s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_json 0.05s call tests/test_dataset_common.py::PackagedDatasetTest::test_builder_class_audiofolder 0.05s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_parquet 0.04s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_pandas 0.03s call tests/test_dataset_common.py::PackagedDatasetTest::test_load_dataset_offline_text (55 durations < 0.005s hidden.) ```
2022-09-30T16:05:42Z
5,037
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-28T12:08:16Z
https://api.github.com/repos/huggingface/datasets/issues/5037/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5037/timeline
Improve CI performance speed of PackagedDatasetTest
https://api.github.com/repos/huggingface/datasets/issues/5037/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-30T16:03:24Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5037.diff", "html_url": "https://github.com/huggingface/datasets/pull/5037", "merged_at": "2022-09-30T16:03:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/5037.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5037" }
PR_kwDODunzps4_xcp0
[ "_The documentation is not available anymore as the PR was closed or merged._", "There was a CI error which seemed unrelated: https://github.com/huggingface/datasets/actions/runs/3143581330/jobs/5111807056\r\n```\r\nFAILED tests/test_load.py::test_load_dataset_private_zipped_images[True] - FileNotFoundError: http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 1, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/5037/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5037
https://github.com/huggingface/datasets/pull/5037
true
1,389,094,075
https://api.github.com/repos/huggingface/datasets/issues/5036/labels{/name}
Hello everyone, Following the issue #4893 and the PR #4831, I propose here an oversampling strategy for a `IterableDataset` list. The `all_exhausted` strategy stops building the new dataset as soon as all samples in each dataset have been added at least once. It follows roughly the same logic behind #4831, namely: - if ``probabilities`` is `None` and the strategy is `all_exhausted`, it simply performs a round robin interleaving that stops when the longest dataset is out of samples. Here the new dataset length will be $maxLengthDataset*nbDataset$. - if ``probabilities`` is not `None` and the strategy is `all_exhausted`, it keeps trace of the datasets which were out of samples but continues to add them to the new dataset, and stops as soons as every dataset runs out of samples at least once. In order to be consistent and also to align with the `Dataset` behavior, please note that the behavior of the default strategy (`first_exhausted`) has been changed. Namely, it really stops when a dataset is out of samples whereas it used to stop when receiving the `StopIteration` error. To give an example of the last note, with the following snippet: ``` >>> from tests.test_iterable_dataset import * >>> d1 = IterableDataset(ExamplesIterable((lambda: (yield from [(i, {"a": i}) for i in [0, 1, 2]])), {})) >>> d2 = IterableDataset(ExamplesIterable((lambda: (yield from [(i, {"a": i}) for i in [10, 11, 12, 13]])), {})) >>> d3 = IterableDataset(ExamplesIterable((lambda: (yield from [(i, {"a": i}) for i in [20, 21, 22, 23, 24]])), {})) >>> dataset = interleave_datasets([d1, d2, d3]) >>> [x["a"] for x in dataset] ``` The result here will then be `[10, 0, 11, 1, 2]` instead of `[10, 0, 11, 1, 2, 20, 12, 13]`. I modified the behavior because I found it to be consistent with the under/oversampling approach and because it unified the undersampling and oversampling code, but I stay open to any suggestions.
2022-09-30T12:30:48Z
5,036
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-28T10:10:23Z
https://api.github.com/repos/huggingface/datasets/issues/5036/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5036/timeline
Add oversampling strategy iterable datasets interleave
https://api.github.com/repos/huggingface/datasets/issues/5036/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/52246514?v=4", "events_url": "https://api.github.com/users/ylacombe/events{/privacy}", "followers_url": "https://api.github.com/users/ylacombe/followers", "following_url": "https://api.github.com/users/ylacombe/following{/other_user}", "gists_url": "https://api.github.com/users/ylacombe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ylacombe", "id": 52246514, "login": "ylacombe", "node_id": "MDQ6VXNlcjUyMjQ2NTE0", "organizations_url": "https://api.github.com/users/ylacombe/orgs", "received_events_url": "https://api.github.com/users/ylacombe/received_events", "repos_url": "https://api.github.com/users/ylacombe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ylacombe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ylacombe/subscriptions", "type": "User", "url": "https://api.github.com/users/ylacombe" }
[]
null
null
CONTRIBUTOR
2022-09-30T12:28:23Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5036.diff", "html_url": "https://github.com/huggingface/datasets/pull/5036", "merged_at": "2022-09-30T12:28:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/5036.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5036" }
PR_kwDODunzps4_w8Bs
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5036/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5036
https://github.com/huggingface/datasets/pull/5036
true
1,388,914,476
https://api.github.com/repos/huggingface/datasets/issues/5035/labels{/name}
Minor fix of typos in load docstrings and comments
2022-09-28T17:28:40Z
5,035
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-28T08:05:07Z
https://api.github.com/repos/huggingface/datasets/issues/5035/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5035/timeline
Fix typos in load docstrings and comments
https://api.github.com/repos/huggingface/datasets/issues/5035/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-28T17:26:15Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5035.diff", "html_url": "https://github.com/huggingface/datasets/pull/5035", "merged_at": "2022-09-28T17:26:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/5035.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5035" }
PR_kwDODunzps4_wVie
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5035/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5035
https://github.com/huggingface/datasets/pull/5035
true
1,388,855,136
https://api.github.com/repos/huggingface/datasets/issues/5034/labels{/name}
null
2022-10-06T15:56:05Z
5,034
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-28T07:17:33Z
https://api.github.com/repos/huggingface/datasets/issues/5034/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5034/timeline
Update README.md of yahoo_answers_topics dataset
https://api.github.com/repos/huggingface/datasets/issues/5034/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6416600?v=4", "events_url": "https://api.github.com/users/borgr/events{/privacy}", "followers_url": "https://api.github.com/users/borgr/followers", "following_url": "https://api.github.com/users/borgr/following{/other_user}", "gists_url": "https://api.github.com/users/borgr/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/borgr", "id": 6416600, "login": "borgr", "node_id": "MDQ6VXNlcjY0MTY2MDA=", "organizations_url": "https://api.github.com/users/borgr/orgs", "received_events_url": "https://api.github.com/users/borgr/received_events", "repos_url": "https://api.github.com/users/borgr/repos", "site_admin": false, "starred_url": "https://api.github.com/users/borgr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/borgr/subscriptions", "type": "User", "url": "https://api.github.com/users/borgr" }
[]
null
null
CONTRIBUTOR
2022-10-04T13:49:25Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5034.diff", "html_url": "https://github.com/huggingface/datasets/pull/5034", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5034.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5034" }
PR_kwDODunzps4_wJCu
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5034). All of your documentation changes will be reflected on that endpoint.", "Thanks, @borgr. We have removed all dataset scripts from this repo. Subsequent PRs should be opened directly on the Hugging Face Hub.", "Do you m...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5034/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5034
https://github.com/huggingface/datasets/pull/5034
true
1,388,842,236
https://api.github.com/repos/huggingface/datasets/issues/5033/labels{/name}
This PR removes some redundant code introduced by mistake after a refactoring in: - #4576
2022-09-28T16:57:51Z
5,033
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-28T07:06:26Z
https://api.github.com/repos/huggingface/datasets/issues/5033/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5033/timeline
Remove redundant code from some dataset module factories
https://api.github.com/repos/huggingface/datasets/issues/5033/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-28T16:55:12Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5033.diff", "html_url": "https://github.com/huggingface/datasets/pull/5033", "merged_at": "2022-09-28T16:55:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/5033.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5033" }
PR_kwDODunzps4_wGSE
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5033/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5033
https://github.com/huggingface/datasets/pull/5033
true
1,388,270,935
https://api.github.com/repos/huggingface/datasets/issues/5032/labels{/name}
**Is your feature request related to a problem? Please describe.** In my research, I am dealing with multi-modal (audio+text+frame sequence) video classification. It would be great if the datasets library supported generating multi-modal batches from a video dataset. **Describe the solution you'd like** Assume I have video files having single/multiple labels. I want to train a single/multi-label video classification model. I want datasets to support generating multi-modal batches (audio+frame sequence) from video files. Audio waveform and frame sequence can be extracted from each video clip then I can use any audio, image and video model from transformers library to extract features which will be fed into my model. **Describe alternatives you've considered** Currently, I am using https://github.com/facebookresearch/pytorchvideo dataloaders. There seems to be not much alternative. **Additional context** I am wiling to open a PR but don't know where to start.
2022-11-02T19:10:13Z
5,032
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-27T19:40:11Z
https://api.github.com/repos/huggingface/datasets/issues/5032/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5032/timeline
new dataset type: single-label and multi-label video classification
https://api.github.com/repos/huggingface/datasets/issues/5032/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/34196005?v=4", "events_url": "https://api.github.com/users/fcakyon/events{/privacy}", "followers_url": "https://api.github.com/users/fcakyon/followers", "following_url": "https://api.github.com/users/fcakyon/following{/other_user}", "gists_url": "https://api.github.com/users/fcakyon/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/fcakyon", "id": 34196005, "login": "fcakyon", "node_id": "MDQ6VXNlcjM0MTk2MDA1", "organizations_url": "https://api.github.com/users/fcakyon/orgs", "received_events_url": "https://api.github.com/users/fcakyon/received_events", "repos_url": "https://api.github.com/users/fcakyon/repos", "site_admin": false, "starred_url": "https://api.github.com/users/fcakyon/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fcakyon/subscriptions", "type": "User", "url": "https://api.github.com/users/fcakyon" }
[]
null
null
NONE
null
null
I_kwDODunzps5Sv1VX
[ "Hi ! You can in the `features` folder how we implemented the audio and image feature types.\r\n\r\nWe can have something similar to videos. What we need to decide:\r\n- the video loading library to use\r\n- the output format when a user accesses a video type object\r\n- what parameters a `Video()` feature type nee...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 1, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5032/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5032
https://github.com/huggingface/datasets/issues/5032
false
1,388,201,146
https://api.github.com/repos/huggingface/datasets/issues/5031/labels{/name}
In huggingface-hub 0.10 the `token` parameter is deprecated for dataset_info and list_repo_files in favor of use_auth_token. Moreover if use_auth_token=None then the user's token is used implicitly. I took those two changes into account Close https://github.com/huggingface/datasets/issues/4990 TODO: - [x] fix tests We should wait hfh 0.10 to be relased first to make sure it works correctly before merging
2022-09-30T09:18:24Z
5,031
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-27T18:37:49Z
https://api.github.com/repos/huggingface/datasets/issues/5031/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5031/timeline
Support hfh 0.10 implicit auth
https://api.github.com/repos/huggingface/datasets/issues/5031/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-30T09:15:59Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5031.diff", "html_url": "https://github.com/huggingface/datasets/pull/5031", "merged_at": "2022-09-30T09:15:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/5031.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5031" }
PR_kwDODunzps4_t82_
[ "_The documentation is not available anymore as the PR was closed or merged._", "@lhoestq it is now released so you can move forward with it :) ", "I took your comments into account @Wauplin :)\r\nI also bumped the requirement to 0.2.0 because we're using `set_access_token`\r\n\r\ncc @albertvillanova WDYT ? I e...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5031/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5031
https://github.com/huggingface/datasets/pull/5031
true
1,388,061,340
https://api.github.com/repos/huggingface/datasets/issues/5030/labels{/name}
Use `pa.Table.to_reader` to make iteration over examples/batches faster in `Dataset.{__iter__, map}` TODO: * [x] benchmarking (the only benchmark for now - iterating over (single) examples of `bookcorpus` (75 mil examples) in Colab is approx. 2.3x faster) * [x] check if iterating over bigger chunks + slicing to fetch individual examples in `_iter` yields better performance
2022-09-29T15:50:44Z
5,030
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-27T16:44:51Z
https://api.github.com/repos/huggingface/datasets/issues/5030/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5030/timeline
Fast dataset iter
https://api.github.com/repos/huggingface/datasets/issues/5030/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-09-29T15:48:17Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5030.diff", "html_url": "https://github.com/huggingface/datasets/pull/5030", "merged_at": "2022-09-29T15:48:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/5030.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5030" }
PR_kwDODunzps4_tfBO
[ "_The documentation is not available anymore as the PR was closed or merged._", "I ran some benchmarks (focused on the data fetching part of `__iter__`) and it seems like the combination `table.to_reader(batch_size)` + `RecordBatch.slice` performs the best ([script](https://gist.github.com/mariosasko/0248288a2e3a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5030/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5030
https://github.com/huggingface/datasets/pull/5030
true
1,387,600,960
https://api.github.com/repos/huggingface/datasets/issues/5029/labels{/name}
This PR addresses a super-simple fix: adding a missing `import` to the `ClassLabel` docstring example, as it was formatted as `from datasets Features`, so it's been fixed to `from datasets import Features`.
2022-09-27T14:03:24Z
5,029
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-27T11:35:29Z
https://api.github.com/repos/huggingface/datasets/issues/5029/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5029/timeline
Fix import in `ClassLabel` docstring example
https://api.github.com/repos/huggingface/datasets/issues/5029/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
[]
null
null
CONTRIBUTOR
2022-09-27T12:27:50Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5029.diff", "html_url": "https://github.com/huggingface/datasets/pull/5029", "merged_at": "2022-09-27T12:27:50Z", "patch_url": "https://github.com/huggingface/datasets/pull/5029.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5029" }
PR_kwDODunzps4_r8-j
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5029/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5029
https://github.com/huggingface/datasets/pull/5029
true
1,386,272,533
https://api.github.com/repos/huggingface/datasets/issues/5028/labels{/name}
Big thanks for providing dataset creation via a generator. I want to ask whether there is any way that parameters can be passed to the method Dataset.from_generator() method, like as follows. ``` from datasets import Dataset def gen(param1): for idx in len(custom_dataset): yield custom_dataset[idx] + param1 ds = Dataset.from_generator(gen(param1)) ```
2022-10-03T13:00:00Z
5,028
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-26T15:20:06Z
https://api.github.com/repos/huggingface/datasets/issues/5028/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5028/timeline
passing parameters to the method passed to Dataset.from_generator()
https://api.github.com/repos/huggingface/datasets/issues/5028/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/64276129?v=4", "events_url": "https://api.github.com/users/Basir-mahmood/events{/privacy}", "followers_url": "https://api.github.com/users/Basir-mahmood/followers", "following_url": "https://api.github.com/users/Basir-mahmood/following{/other_user}", "gists_url": "https://api.github.com/users/Basir-mahmood/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Basir-mahmood", "id": 64276129, "login": "Basir-mahmood", "node_id": "MDQ6VXNlcjY0Mjc2MTI5", "organizations_url": "https://api.github.com/users/Basir-mahmood/orgs", "received_events_url": "https://api.github.com/users/Basir-mahmood/received_events", "repos_url": "https://api.github.com/users/Basir-mahmood/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Basir-mahmood/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Basir-mahmood/subscriptions", "type": "User", "url": "https://api.github.com/users/Basir-mahmood" }
[]
null
completed
NONE
2022-10-03T13:00:00Z
null
I_kwDODunzps5SoNcV
[ "Hi! Yes, you can either use the `gen_kwargs` param in `Dataset.from_generator` (`ds = Dataset.from_generator(gen, gen_kwargs={\"param1\": val})`) or wrap the generator function with `functools.partial`\r\n(`ds = Dataset.from_generator(functools.partial(gen, param1=\"val\"))`) to pass custom parameters to it.\r\n" ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5028/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5028
https://github.com/huggingface/datasets/issues/5028
false
1,386,153,072
https://api.github.com/repos/huggingface/datasets/issues/5027/labels{/name}
null
2022-09-27T12:28:03Z
5,027
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-26T14:10:09Z
https://api.github.com/repos/huggingface/datasets/issues/5027/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5027/timeline
Fix typo in error message
https://api.github.com/repos/huggingface/datasets/issues/5027/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[]
null
null
CONTRIBUTOR
2022-09-27T12:26:02Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5027.diff", "html_url": "https://github.com/huggingface/datasets/pull/5027", "merged_at": "2022-09-27T12:26:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/5027.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5027" }
PR_kwDODunzps4_nFUE
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5027/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5027
https://github.com/huggingface/datasets/pull/5027
true
1,386,071,154
https://api.github.com/repos/huggingface/datasets/issues/5026/labels{/name}
Should fix the tests for `huggingface_hub==0.10.0rc0` prerelease (see [failed CI](https://github.com/huggingface/datasets/actions/runs/3127805250/jobs/5074879144)). Related to [this thread](https://huggingface.slack.com/archives/C02V5EA0A95/p1664195165294559) (internal link). Note: this should be a backward compatible fix (e.g. works also with previous versions of `huggingface_hub`) I am not sure where to put the changes so feel free to cherry-pick the commit and close this one without merging. cc @lhoestq
2022-09-26T14:30:55Z
5,026
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-26T13:19:01Z
https://api.github.com/repos/huggingface/datasets/issues/5026/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5026/timeline
patch CI_HUB_TOKEN_PATH with Path instead of str
https://api.github.com/repos/huggingface/datasets/issues/5026/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/11801849?v=4", "events_url": "https://api.github.com/users/Wauplin/events{/privacy}", "followers_url": "https://api.github.com/users/Wauplin/followers", "following_url": "https://api.github.com/users/Wauplin/following{/other_user}", "gists_url": "https://api.github.com/users/Wauplin/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Wauplin", "id": 11801849, "login": "Wauplin", "node_id": "MDQ6VXNlcjExODAxODQ5", "organizations_url": "https://api.github.com/users/Wauplin/orgs", "received_events_url": "https://api.github.com/users/Wauplin/received_events", "repos_url": "https://api.github.com/users/Wauplin/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Wauplin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wauplin/subscriptions", "type": "User", "url": "https://api.github.com/users/Wauplin" }
[]
null
null
CONTRIBUTOR
2022-09-26T14:28:45Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5026.diff", "html_url": "https://github.com/huggingface/datasets/pull/5026", "merged_at": "2022-09-26T14:28:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/5026.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5026" }
PR_kwDODunzps4_mz1w
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5026/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5026
https://github.com/huggingface/datasets/pull/5026
true
1,386,011,239
https://api.github.com/repos/huggingface/datasets/issues/5025/labels{/name}
## Describe the bug A clear and concise description of what the bug is. I tried to create my custom dataset using below code ``` from datasets import Features, Sequence, ClassLabel, Value, Array2D, Array3D from torchvision import transforms from transformers import AutoProcessor # we'll use the Auto API here - it will load LayoutLMv3Processor behind the scenes, # based on the checkpoint we provide from the hub from datasets import load_dataset def prepare_examples(examples): #Some preporcessing for each image and text as all my data saved in cloud #For this reason I couldn't set the batch to True. encoding = processor(img_as_tensor, words, boxes=boxes, word_labels=labels, truncation=True, padding="max_length") # encoding['pixel_values']=np.array(encoding['pixel_values']) return encoding dataset = load_dataset("json", data_files='issues.jsonl') processor = AutoProcessor.from_pretrained("microsoft/layoutlmv3-base", apply_ocr=False) features = dataset["train"].features column_names = dataset["train"].column_names # we need to define custom features for `set_format` (used later on) to work properly features = Features({ 'pixel_values': Array3D(dtype="float32", shape=(3, 224, 224)), 'input_ids': Sequence(feature=Value(dtype='int64')), 'attention_mask': Sequence(Value(dtype='int64')), 'bbox': Array2D(dtype="int64", shape=(512, 4)), 'labels': Sequence(feature=Value(dtype='int64')), }) train_dataset = dataset["train"].map( prepare_examples, batched=False, remove_columns=column_names, features=features ) ``` It throws below error. ``` /opt/conda/lib/python3.7/site-packages/datasets/arrow_writer.py in __arrow_array__(self, type) 172 storage = to_pyarrow_listarray(data, pa_type) --> 173 return pa.ExtensionArray.from_storage(pa_type, storage) 174 /opt/conda/lib/python3.7/site-packages/pyarrow/array.pxi in pyarrow.lib.ExtensionArray.from_storage() TypeError: Incompatible storage type list<item: list<item: list<item: list<item: float>>>> for extension type extension<arrow.py_extension_type<Array3DExtensionType>> ``` ## Steps to reproduce the bug ```python # Sample code to reproduce the bug ``` rom datasets import Features, Sequence, ClassLabel, Value, Array2D, Array3D from torchvision import transforms from transformers import AutoProcessor # we'll use the Auto API here - it will load LayoutLMv3Processor behind the scenes, # based on the checkpoint we provide from the hub from datasets import load_dataset def prepare_examples(examples): #Some preporcessing for each image and text as all my data saved in cloud encoding = processor(img_as_tensor, words, boxes=boxes, word_labels=labels, truncation=True, padding="max_length") # encoding['pixel_values']=np.array(encoding['pixel_values']) return encoding dataset = load_dataset("json", data_files='issues.jsonl') processor = AutoProcessor.from_pretrained("microsoft/layoutlmv3-base", apply_ocr=False) features = dataset["train"].features column_names = dataset["train"].column_names # we need to define custom features for `set_format` (used later on) to work properly features = Features({ 'pixel_values': Array3D(dtype="float32", shape=(3, 224, 224)), 'input_ids': Sequence(feature=Value(dtype='int64')), 'attention_mask': Sequence(Value(dtype='int64')), 'bbox': Array2D(dtype="int64", shape=(512, 4)), 'labels': Sequence(feature=Value(dtype='int64')), }) train_dataset = dataset["train"].map( prepare_examples, batched=False, remove_columns=column_names, features=features ) ## Expected results A clear and concise description of the expected results. Expected would be similar to all the otherdatasets with no error. ## Actual results Specify the actual results or traceback. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - Platform: Unix - Python version: 3.9 - PyArrow version: 9.0.0
2022-09-27T19:50:00Z
5,025
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-26T12:38:39Z
https://api.github.com/repos/huggingface/datasets/issues/5025/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5025/timeline
Custom Json Dataset Throwing Error when batch is False
https://api.github.com/repos/huggingface/datasets/issues/5025/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/21245519?v=4", "events_url": "https://api.github.com/users/jmandivarapu1/events{/privacy}", "followers_url": "https://api.github.com/users/jmandivarapu1/followers", "following_url": "https://api.github.com/users/jmandivarapu1/following{/other_user}", "gists_url": "https://api.github.com/users/jmandivarapu1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jmandivarapu1", "id": 21245519, "login": "jmandivarapu1", "node_id": "MDQ6VXNlcjIxMjQ1NTE5", "organizations_url": "https://api.github.com/users/jmandivarapu1/orgs", "received_events_url": "https://api.github.com/users/jmandivarapu1/received_events", "repos_url": "https://api.github.com/users/jmandivarapu1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jmandivarapu1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmandivarapu1/subscriptions", "type": "User", "url": "https://api.github.com/users/jmandivarapu1" }
[]
null
completed
NONE
2022-09-27T19:50:00Z
null
I_kwDODunzps5SnNpn
[ "Hi! Our processors are meant to be used in `batched` mode, so if `batched` is `False`, you need to drop the batch dimension (the error message warns you that the array has an extra dimension meaning it's 4D instead of 3D) to avoid the error:\r\n```python\r\ndef prepare_examples(examples):\r\n #Some preporcessin...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5025/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5025
https://github.com/huggingface/datasets/issues/5025
false
1,385,947,624
https://api.github.com/repos/huggingface/datasets/issues/5024/labels{/name}
This PR fixes string features of `xcsr` dataset to avoid character splitting. Fix #5023. CC: @yangxqiao, @yuchenlin
2022-09-28T07:56:18Z
5,024
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-26T11:55:36Z
https://api.github.com/repos/huggingface/datasets/issues/5024/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5024/timeline
Fix string features of xcsr dataset
https://api.github.com/repos/huggingface/datasets/issues/5024/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-28T07:54:19Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5024.diff", "html_url": "https://github.com/huggingface/datasets/pull/5024", "merged_at": "2022-09-28T07:54:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/5024.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5024" }
PR_kwDODunzps4_mZ3J
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5024/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5024
https://github.com/huggingface/datasets/pull/5024
true
1,385,881,112
https://api.github.com/repos/huggingface/datasets/issues/5023/labels{/name}
## Describe the bug Text strings are split into lists of characters. Example for "X-CSQA-en": ``` {'id': 'd3845adc08414fda', 'lang': 'en', 'question': {'stem': ['T', 'h', 'e', ' ', 'd', 'e', 'n', 't', 'a', 'l', ' ', 'o', 'f', 'f', 'i', 'c', 'e', ' ', 'h', 'a', 'n', 'd', 'l', 'e', 'd', ' ', 'a', ' ', 'l', 'o', 't', ' ', 'o', 'f', ' ', 'p', 'a', 't', 'i', 'e', 'n', 't', 's', ' ', 'w', 'h', 'o', ' ', 'e', 'x', 'p', 'e', 'r', 'i', 'e', 'n', 'c', 'e', 'd', ' ', 't', 'r', 'a', 'u', 'm', 'a', 't', 'i', 'c', ' ', 'm', 'o', 'u', 't', 'h', ' ', 'i', 'n', 'j', 'u', 'r', 'y', ',', ' ', 'w', 'h', 'e', 'r', 'e', ' ', 'w', 'e', 'r', 'e', ' ', 't', 'h', 'e', 's', 'e', ' ', 'p', 'a', 't', 'i', 'e', 'n', 't', 's', ' ', 'c', 'o', 'm', 'i', 'n', 'g', ' ', 'f', 'r', 'o', 'm', '?'], 'choices': [{'label': ['A'], 'text': ['t', 'o', 'w', 'n']}, {'label': ['B'], 'text': ['m', 'i', 'c', 'h', 'i', 'g', 'a', 'n']}, {'label': ['C'], 'text': ['h', 'o', 's', 'p', 'i', 't', 'a', 'l']}, {'label': ['D'], 'text': ['s', 'c', 'h', 'o', 'o', 'l', 's']}, {'label': ['E'], 'text': ['o', 'f', 'f', 'i', 'c', 'e', ' ', 'b', 'u', 'i', 'l', 'd', 'i', 'n', 'g']}]}, 'answerKey': 'C'} ## Steps to reproduce the bug ```python ds = load_dataset("datasets/xcsr", "X-CSQA-en", split="validation", streaming=True) item = next(iter(ds)) item ``` ## Expected results ``` {'id': 'd3845adc08414fda', 'lang': 'en', 'question': {'stem': 'The dental office handled a lot of patients who experienced traumatic mouth injury, where were these patients coming from?', 'choices': {'label': ['A', 'B', 'C', 'D', 'E'], 'text': ['town', 'michigan', 'hospital', 'schools', 'office building']}}, 'answerKey': 'C'} ```
2022-09-28T07:54:20Z
5,023
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
2022-09-26T11:11:50Z
https://api.github.com/repos/huggingface/datasets/issues/5023/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5023/timeline
Text strings are split into lists of characters in xcsr dataset
https://api.github.com/repos/huggingface/datasets/issues/5023/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
MEMBER
2022-09-28T07:54:20Z
null
I_kwDODunzps5Smt4Y
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5023/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5023
https://github.com/huggingface/datasets/issues/5023
false
1,385,432,859
https://api.github.com/repos/huggingface/datasets/issues/5022/labels{/name}
Fix #5017. CC: @yangxqiao, @yuchenlin
2022-09-26T12:27:20Z
5,022
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-26T05:13:39Z
https://api.github.com/repos/huggingface/datasets/issues/5022/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5022/timeline
Fix languages of X-CSQA configs in xcsr dataset
https://api.github.com/repos/huggingface/datasets/issues/5022/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-26T10:57:30Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5022.diff", "html_url": "https://github.com/huggingface/datasets/pull/5022", "merged_at": "2022-09-26T10:57:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/5022.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5022" }
PR_kwDODunzps4_kxYe
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks @lhoestq, I had missed that... ", "thx for the super fast work @albertvillanova ! any estimate for when the relevant release will happen?\r\n\r\nThanks again ", "@thesofakillers after a recent change in our library (see #4...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5022/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5022
https://github.com/huggingface/datasets/pull/5022
true
1,385,351,250
https://api.github.com/repos/huggingface/datasets/issues/5021/labels{/name}
## Describe the bug Including the strings "test" or "train" anywhere in a filename causes `datasets` to infer the split and silently ignore all other files. This behavior is documented for directory names but not filenames: https://huggingface.co/docs/datasets/image_dataset#imagefolder ## Steps to reproduce the bug `metadata.jsonl` ```json {"file_name": "photo of a cat.jpg", "text": "a photo of a cat"} {"file_name": "photo of a dog.jpg", "text": "a photo of a dog"} {"file_name": "photo of a train.jpg", "text": "a photo of a train"} {"file_name": "photo of test tubes.jpg", "text": "a photo of test tubes"} ``` `bug.py` ```python from datasets import load_dataset dataset = load_dataset("dataset") print(dataset) # DatasetDict({ # train: Dataset({ # features: ['image', 'text'], # num_rows: 1 # }) # test: Dataset({ # features: ['image', 'text'], # num_rows: 1 # }) # }) for split in dataset: for n in dataset[split]: print(n['text']) # a photo of a train # a photo of test tubes ``` ## Expected results One single dataset with all four images / a warning for unused files / documentation of this behavior ## Actual results Only the images with "test" or "train" in the name are loaded ## Environment info - `datasets` version: 2.5.1 - Platform: macOS-12.5.1-x86_64-i386-64bit - Python version: 3.10.4 - PyArrow version: 9.0.0 - Pandas version: 1.5.0
2022-09-29T08:07:50Z
5,021
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "cfd3d7", "default": true, "descript...
2022-09-26T03:22:14Z
https://api.github.com/repos/huggingface/datasets/issues/5021/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5021/timeline
Split is inferred from filename and overrides metadata.jsonl
https://api.github.com/repos/huggingface/datasets/issues/5021/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/102226344?v=4", "events_url": "https://api.github.com/users/float-trip/events{/privacy}", "followers_url": "https://api.github.com/users/float-trip/followers", "following_url": "https://api.github.com/users/float-trip/following{/other_user}", "gists_url": "https://api.github.com/users/float-trip/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/float-trip", "id": 102226344, "login": "float-trip", "node_id": "U_kgDOBhfZqA", "organizations_url": "https://api.github.com/users/float-trip/orgs", "received_events_url": "https://api.github.com/users/float-trip/received_events", "repos_url": "https://api.github.com/users/float-trip/repos", "site_admin": false, "starred_url": "https://api.github.com/users/float-trip/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/float-trip/subscriptions", "type": "User", "url": "https://api.github.com/users/float-trip" }
[]
null
completed
NONE
2022-09-29T08:07:50Z
null
I_kwDODunzps5SkshS
[ "Hi! What's the structure of your image folder? `datasets` by default tries to infer to what split each file belongs based on directory/file names. If it's OK to load all the images inside the `dataset` folder in the `train` split, you can do the following:\r\n```python\r\ndataset = load_dataset(\"imagefolder\", da...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5021/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5021
https://github.com/huggingface/datasets/issues/5021
false
1,384,684,078
https://api.github.com/repos/huggingface/datasets/issues/5020/labels{/name}
Forbidden You don't have permission to access /~vicente/sbucaptions/sbu-captions-all.tar.gz on this server. Additionally, a 403 Forbidden error was encountered while trying to use an ErrorDocument to handle the request. Apache/2.4.6 (Red Hat Enterprise Linux) OpenSSL/1.0.2k-fips PHP/5.4.16 mod_fcgid/2.3.9 mod_wsgi/3.4 Python/2.7.5 mod_perl/2.0.11 Perl/v5.16.3 Server at [www.cs.virginia.edu](mailto:csroot@virginia.edu) Port 443
2022-09-28T07:20:20Z
5,020
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-24T14:00:33Z
https://api.github.com/repos/huggingface/datasets/issues/5020/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5020/timeline
Fix URLs of sbu_captions dataset
https://api.github.com/repos/huggingface/datasets/issues/5020/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1070872?v=4", "events_url": "https://api.github.com/users/donglixp/events{/privacy}", "followers_url": "https://api.github.com/users/donglixp/followers", "following_url": "https://api.github.com/users/donglixp/following{/other_user}", "gists_url": "https://api.github.com/users/donglixp/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/donglixp", "id": 1070872, "login": "donglixp", "node_id": "MDQ6VXNlcjEwNzA4NzI=", "organizations_url": "https://api.github.com/users/donglixp/orgs", "received_events_url": "https://api.github.com/users/donglixp/received_events", "repos_url": "https://api.github.com/users/donglixp/repos", "site_admin": false, "starred_url": "https://api.github.com/users/donglixp/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/donglixp/subscriptions", "type": "User", "url": "https://api.github.com/users/donglixp" }
[]
null
null
CONTRIBUTOR
2022-09-28T07:18:23Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5020.diff", "html_url": "https://github.com/huggingface/datasets/pull/5020", "merged_at": "2022-09-28T07:18:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/5020.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5020" }
PR_kwDODunzps4_istJ
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5020/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5020
https://github.com/huggingface/datasets/pull/5020
true
1,384,673,718
https://api.github.com/repos/huggingface/datasets/issues/5019/labels{/name}
Hi, I updated the dataset to include additional data made available recently. When I test it locally, it seems to work. However, I get the following error with the dummy data creation: `Dummy data generation done but dummy data test failed since splits ['train', 'validation', 'test'] have 0 examples for config 'fr'`. Do you know why this could be the case? Cheers, Joel
2022-09-28T07:13:39Z
5,019
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-24T13:28:57Z
https://api.github.com/repos/huggingface/datasets/issues/5019/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5019/timeline
Update swiss judgment prediction
https://api.github.com/repos/huggingface/datasets/issues/5019/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3775944?v=4", "events_url": "https://api.github.com/users/JoelNiklaus/events{/privacy}", "followers_url": "https://api.github.com/users/JoelNiklaus/followers", "following_url": "https://api.github.com/users/JoelNiklaus/following{/other_user}", "gists_url": "https://api.github.com/users/JoelNiklaus/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JoelNiklaus", "id": 3775944, "login": "JoelNiklaus", "node_id": "MDQ6VXNlcjM3NzU5NDQ=", "organizations_url": "https://api.github.com/users/JoelNiklaus/orgs", "received_events_url": "https://api.github.com/users/JoelNiklaus/received_events", "repos_url": "https://api.github.com/users/JoelNiklaus/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JoelNiklaus/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JoelNiklaus/subscriptions", "type": "User", "url": "https://api.github.com/users/JoelNiklaus" }
[]
null
null
CONTRIBUTOR
2022-09-28T05:48:50Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5019.diff", "html_url": "https://github.com/huggingface/datasets/pull/5019", "merged_at": "2022-09-28T05:48:50Z", "patch_url": "https://github.com/huggingface/datasets/pull/5019.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5019" }
PR_kwDODunzps4_iq9b
[ "Thank you very much for the detailed review @albertvillanova!\r\n\r\nI updated the PR with the requested changes. ", "At the end, I had to manually fix the conflict, so that CI tests are launched.\r\n\r\nPLEASE NOTE: you should first pull to incorporate the previous commit\r\n```shell\r\ngit pull\r\n```", "_Th...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5019/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5019
https://github.com/huggingface/datasets/pull/5019
true
1,384,146,585
https://api.github.com/repos/huggingface/datasets/issues/5018/labels{/name}
Following https://github.com/huggingface/datasets/pull/4926 Creates all the `dataset_info` YAML fields in the dataset cards The JSON are also updated using the simplified backward compatible format added in https://github.com/huggingface/datasets/pull/4926 Needs https://github.com/huggingface/datasets/pull/4926 to be merged first
2023-09-24T09:33:21Z
5,018
null
https://api.github.com/repos/huggingface/datasets
true
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-23T18:08:15Z
https://api.github.com/repos/huggingface/datasets/issues/5018/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5018/timeline
Create all YAML dataset_info
https://api.github.com/repos/huggingface/datasets/issues/5018/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-10-03T17:08:05Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5018.diff", "html_url": "https://github.com/huggingface/datasets/pull/5018", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5018.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5018" }
PR_kwDODunzps4_hA0V
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5018). All of your documentation changes will be reflected on that endpoint.", "Closing since https://github.com/huggingface/datasets/pull/4974 removed all the datasets scripts.\r\n\r\nIndividual PRs must be opened on the Huggi...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5018/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5018
https://github.com/huggingface/datasets/pull/5018
true
1,384,022,463
https://api.github.com/repos/huggingface/datasets/issues/5017/labels{/name}
## Describe the bug All the alleged non-english subcollections for the X-CSQA task in the [xcsr benchmark dataset ](https://huggingface.co/datasets/xcsr) seem to be copies of the english subcollection, rather than translations. This is in contrast to the data description: > we automatically translate the original CSQA and CODAH datasets, which only have English versions, to 15 other languages, forming development and test sets for studying X-CSR ## Steps to reproduce the bug ```python # let's say you want to load the french X-CSQA subcollection french = datasets.load_dataset("xcsr", "X-CSQA-fr") # for good measure, let's load english too english = datasets.load_dataset("xcsr", "X-CSQA-en") # let's inspect "".join(english['test'][0]['question']['stem']) # output: 'The people wanted to stop the parade, so what did they set up to thwart it?' "".join(french['test'][0]['question']['stem']) # output: 'The people wanted to stop the parade, so what did they set up to thwart it?' # what? Why are they both in english? # I've checked this for validation and train splits too, across many datapoints. It's all the same english dataset # maybe i need to look better? french['test'].unique('lang') # output: ['en'] # no, it's all english ``` ## Expected results Accessing a subcollection in language X should return a subcollection containg samples in language X ## Actual results Accessing a subcollection in language X returns a subcollection containing samples in English. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.5.1 - Platform: macOS-10.15.7-x86_64-i386-64bit - Python version: 3.8.13 - PyArrow version: 9.0.0 - Pandas version: 1.4.3
2022-09-26T10:57:31Z
5,017
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
2022-09-23T16:11:54Z
https://api.github.com/repos/huggingface/datasets/issues/5017/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5017/timeline
xcsr: X-CSQA simply uses english for all alleged non-english data
https://api.github.com/repos/huggingface/datasets/issues/5017/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/26286291?v=4", "events_url": "https://api.github.com/users/thesofakillers/events{/privacy}", "followers_url": "https://api.github.com/users/thesofakillers/followers", "following_url": "https://api.github.com/users/thesofakillers/following{/other_user}", "gists_url": "https://api.github.com/users/thesofakillers/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thesofakillers", "id": 26286291, "login": "thesofakillers", "node_id": "MDQ6VXNlcjI2Mjg2Mjkx", "organizations_url": "https://api.github.com/users/thesofakillers/orgs", "received_events_url": "https://api.github.com/users/thesofakillers/received_events", "repos_url": "https://api.github.com/users/thesofakillers/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thesofakillers/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thesofakillers/subscriptions", "type": "User", "url": "https://api.github.com/users/thesofakillers" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
NONE
2022-09-26T10:57:31Z
null
I_kwDODunzps5SfoG_
[ "Thanks for reporting, @thesofakillers. Good catch. We are fixing this. " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5017/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5017
https://github.com/huggingface/datasets/issues/5017
false
1,383,883,058
https://api.github.com/repos/huggingface/datasets/issues/5016/labels{/name}
Fix for CVE-2007-4559 Description: Directory traversal vulnerability in the (1) extract and (2) extractall functions in the tarfile module in Python allows user-assisted remote attackers to overwrite arbitrary files via a .. (dot dot) sequence in filenames in a TAR archive, a related issue to CVE-2001-1267. I fixed it by using the solution proposed in https://stackoverflow.com/questions/10060069/safely-extract-zip-or-tar-using-python It blocks extraction of files with an absolute path or double dots and symlinks.
2022-09-29T12:42:26Z
5,016
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-23T14:22:21Z
https://api.github.com/repos/huggingface/datasets/issues/5016/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5016/timeline
Fix tar extraction vuln
https://api.github.com/repos/huggingface/datasets/issues/5016/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-29T12:40:28Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5016.diff", "html_url": "https://github.com/huggingface/datasets/pull/5016", "merged_at": "2022-09-29T12:40:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/5016.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5016" }
PR_kwDODunzps4_gKny
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5016/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5016
https://github.com/huggingface/datasets/pull/5016
true
1,383,485,558
https://api.github.com/repos/huggingface/datasets/issues/5015/labels{/name}
Before merging: - #4974 TODO: - [x] Create label: ["dataset contribution"](https://github.com/huggingface/datasets/pulls?q=label%3A%22dataset+contribution%22) - [x] Create project: [Datasets: Transfer datasets to Hub](https://github.com/orgs/huggingface/projects/22/) - [x] PRs: - [x] Add dataset: we should recommend transfer all additions of datasets to the Hub, under the appropriate namespace; no more additions of datasets on GitHub - [x] Update dataset: in general, we should merge bug fixes; enhancements should be considered on a case-by-case basis, depending on whether there is a more suitable namespace on the Hub - [ ] Issues Finally: - [x] #4974 Let me know what you think! :hugs:
2022-10-05T07:15:57Z
5,015
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-09-23T08:48:10Z
https://api.github.com/repos/huggingface/datasets/issues/5015/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5015/timeline
Transfer dataset scripts to Hub
https://api.github.com/repos/huggingface/datasets/issues/5015/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
MEMBER
2022-10-05T07:15:57Z
null
I_kwDODunzps5SdlB2
[ "Sounds good ! Can I help with anything ?" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5015/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5015
https://github.com/huggingface/datasets/issues/5015
false
1,383,422,639
https://api.github.com/repos/huggingface/datasets/issues/5014/labels{/name}
I need to read the custom dataset in conll format
2022-11-02T11:57:15Z
5,014
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-23T07:49:42Z
https://api.github.com/repos/huggingface/datasets/issues/5014/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5014/timeline
I need to read the custom dataset in conll format
https://api.github.com/repos/huggingface/datasets/issues/5014/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/39985245?v=4", "events_url": "https://api.github.com/users/shell-nlp/events{/privacy}", "followers_url": "https://api.github.com/users/shell-nlp/followers", "following_url": "https://api.github.com/users/shell-nlp/following{/other_user}", "gists_url": "https://api.github.com/users/shell-nlp/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/shell-nlp", "id": 39985245, "login": "shell-nlp", "node_id": "MDQ6VXNlcjM5OTg1MjQ1", "organizations_url": "https://api.github.com/users/shell-nlp/orgs", "received_events_url": "https://api.github.com/users/shell-nlp/received_events", "repos_url": "https://api.github.com/users/shell-nlp/repos", "site_admin": false, "starred_url": "https://api.github.com/users/shell-nlp/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shell-nlp/subscriptions", "type": "User", "url": "https://api.github.com/users/shell-nlp" }
[]
null
reopened
NONE
null
null
I_kwDODunzps5SdVqv
[ "Hi! We don't currently have a builder for parsing custom `conll` datasets, but I guess we could add one as a packaged module (similarly to what [TFDS](https://github.com/tensorflow/datasets/blob/master/tensorflow_datasets/core/dataset_builders/conll/conll_dataset_builder.py) did). @lhoestq @albertvillanova WDYT?\r...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5014/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5014
https://github.com/huggingface/datasets/issues/5014
false
1,383,415,971
https://api.github.com/repos/huggingface/datasets/issues/5013/labels{/name}
HI: I use cpp env libtorch, I like use hugggingface ,but huggingface not cpp binding, would you like publish cpp binding for it. thanks
2023-02-24T16:20:57Z
5,013
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "ffffff", "default": true, "description": "This will not be worked on", "id": 1935892913, "name": "wontfix", "node_id": "MDU6TGFiZWwxOTM1ODkyOTEz", "url": "https://api.github.com/repos/huggingface/datasets/labels/wontfix" } ]
2022-09-23T07:42:49Z
https://api.github.com/repos/huggingface/datasets/issues/5013/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5013/timeline
would huggingface like publish cpp binding for datasets package ?
https://api.github.com/repos/huggingface/datasets/issues/5013/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6143404?v=4", "events_url": "https://api.github.com/users/mullerhai/events{/privacy}", "followers_url": "https://api.github.com/users/mullerhai/followers", "following_url": "https://api.github.com/users/mullerhai/following{/other_user}", "gists_url": "https://api.github.com/users/mullerhai/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mullerhai", "id": 6143404, "login": "mullerhai", "node_id": "MDQ6VXNlcjYxNDM0MDQ=", "organizations_url": "https://api.github.com/users/mullerhai/orgs", "received_events_url": "https://api.github.com/users/mullerhai/received_events", "repos_url": "https://api.github.com/users/mullerhai/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mullerhai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mullerhai/subscriptions", "type": "User", "url": "https://api.github.com/users/mullerhai" }
[]
null
completed
NONE
2023-02-24T16:20:57Z
null
I_kwDODunzps5SdUCj
[ "Hi ! Can you share more information about your use case ? How could it help you to have cpp bindings versus using the python libraries ?", "> Hi ! Can you share more information about your use case ? How could it help you to have cpp bindings versus using the python libraries ?\r\n\r\nfor example ,the huggingfac...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5013/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5013
https://github.com/huggingface/datasets/issues/5013
false
1,382,851,096
https://api.github.com/repos/huggingface/datasets/issues/5012/labels{/name}
I have a file on S3 created by Data Version Control, it looks like `s3://dvc/ac/badff5b134382a0f25248f1b45d7b2` but contains a json file. If I run ```python dataset = load_dataset( "json", data_files='s3://dvc/ac/badff5b134382a0f25248f1b45d7b2' ) ``` It gives me ``` InvalidSchema: No connection adapters were found for 's3://dvc/ac/badff5b134382a0f25248f1b45d7b2' ``` However, I cannot go ahead and change the names of the s3 file. Is there a way to "force" load a S3 url with certain decoder (JSON, CSV, etc.) regardless of s3 URL naming?
2023-08-16T09:58:36Z
5,012
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-22T18:28:15Z
https://api.github.com/repos/huggingface/datasets/issues/5012/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5012/timeline
Force JSON format regardless of file naming on S3
https://api.github.com/repos/huggingface/datasets/issues/5012/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/112650299?v=4", "events_url": "https://api.github.com/users/junwang-wish/events{/privacy}", "followers_url": "https://api.github.com/users/junwang-wish/followers", "following_url": "https://api.github.com/users/junwang-wish/following{/other_user}", "gists_url": "https://api.github.com/users/junwang-wish/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/junwang-wish", "id": 112650299, "login": "junwang-wish", "node_id": "U_kgDOBrboOw", "organizations_url": "https://api.github.com/users/junwang-wish/orgs", "received_events_url": "https://api.github.com/users/junwang-wish/received_events", "repos_url": "https://api.github.com/users/junwang-wish/repos", "site_admin": false, "starred_url": "https://api.github.com/users/junwang-wish/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/junwang-wish/subscriptions", "type": "User", "url": "https://api.github.com/users/junwang-wish" }
[]
null
completed
NONE
2023-08-16T09:58:36Z
null
I_kwDODunzps5SbKIY
[ "Hi ! Support for URIs like `s3://...` is not implemented yet in `data_files=`. You can use the HTTP URL instead if your data is public in the meantime", "Hi,\r\nI want to make sure I understand this response. I have a set of files on S3 that are private for security reasons. Because they are not public files I ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5012/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5012
https://github.com/huggingface/datasets/issues/5012
false
1,382,609,587
https://api.github.com/repos/huggingface/datasets/issues/5011/labels{/name}
## Describe the bug Loading the dataset [earnings-22](https://huggingface.co/datasets/sanchit-gandhi/earnings22_split) from the Hub yields an Index Error. I created this dataset locally and then pushed to hub at the specified URL. Thus, I expect the dataset should work out-of-the-box! Indeed, the dataset viewer functions correctly, and there were no issues when I had the dataset locally. Don't think it's a sound file bug as the version matches what worked previously. Update: the bug appeared for me on a GPU, mysteriously on a TPU I can't repro and it downloads correctly... ## Steps to reproduce the bug ```python from datasets import load_dataset earnings22 = load_dataset("sanchit-gandhi/earnings22_split") ``` ## Expected results ``` >>> earnings22 DatasetDict({ validation: Dataset({ features: ['source_id', 'audio', 'segment_id', 'sentence', 'start_ts', 'end_ts', 'id'], num_rows: 2650 }) train: Dataset({ features: ['source_id', 'audio', 'segment_id', 'sentence', 'start_ts', 'end_ts', 'id'], num_rows: 52006 }) test: Dataset({ features: ['source_id', 'audio', 'segment_id', 'sentence', 'start_ts', 'end_ts', 'id'], num_rows: 2735 }) }) ``` ## Actual results ``` Traceback (most recent call last): File "/opt/conda/envs/hf/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 2764, in _map_single writer.write(example) File "/opt/conda/envs/hf/lib/python3.8/site-packages/datasets/arrow_writer.py", line 451, in write self.write_examples_on_file() File "/opt/conda/envs/hf/lib/python3.8/site-packages/datasets/arrow_writer.py", line 409, in write_examples_on_file self.write_batch(batch_examples=batch_examples) File "/opt/conda/envs/hf/lib/python3.8/site-packages/datasets/arrow_writer.py", line 508, in write_batch arrays.append(pa.array(typed_sequence)) File "pyarrow/array.pxi", line 231, in pyarrow.lib.array File "pyarrow/array.pxi", line 110, in pyarrow.lib._handle_arrow_array_protocol File "/opt/conda/envs/hf/lib/python3.8/site-packages/datasets/arrow_writer.py", line 197, in __arrow_array__ out = cast_array_to_feature(out, type, allow_number_to_str=not self.trying_type) File "/opt/conda/envs/hf/lib/python3.8/site-packages/datasets/table.py", line 1683, in wrapper return func(array, *args, **kwargs) File "/opt/conda/envs/hf/lib/python3.8/site-packages/datasets/table.py", line 1795, in cast_array_to_feature return feature.cast_storage(array) File "/opt/conda/envs/hf/lib/python3.8/site-packages/datasets/features/audio.py", line 190, in cast_storage storage = pa.array([Audio().encode_example(x) if x is not None else None for x in storage.to_pylist()]) File "/opt/conda/envs/hf/lib/python3.8/site-packages/datasets/features/audio.py", line 190, in <listcomp> storage = pa.array([Audio().encode_example(x) if x is not None else None for x in storage.to_pylist()]) File "/opt/conda/envs/hf/lib/python3.8/site-packages/datasets/features/audio.py", line 92, in encode_example sf.write(buffer, value["array"], value["sampling_rate"], format="wav") File "/opt/conda/envs/hf/lib/python3.8/site-packages/soundfile.py", line 313, in write channels = data.shape[1] IndexError: tuple index out of range ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.4.0 - Platform: Linux-4.19.0-21-cloud-amd64-x86_64-with-glibc2.10 - Python version: 3.8.13 - PyArrow version: 9.0.0 - Pandas version: 1.4.3 Plus: - SoundFile version: 0.10.3.post1 cc @lhoestq @polinaeterna
2022-09-23T09:05:18Z
5,011
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-22T15:07:27Z
https://api.github.com/repos/huggingface/datasets/issues/5011/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5011/timeline
Audio: `encode_example` fails with IndexError
https://api.github.com/repos/huggingface/datasets/issues/5011/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/93869735?v=4", "events_url": "https://api.github.com/users/sanchit-gandhi/events{/privacy}", "followers_url": "https://api.github.com/users/sanchit-gandhi/followers", "following_url": "https://api.github.com/users/sanchit-gandhi/following{/other_user}", "gists_url": "https://api.github.com/users/sanchit-gandhi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sanchit-gandhi", "id": 93869735, "login": "sanchit-gandhi", "node_id": "U_kgDOBZhWpw", "organizations_url": "https://api.github.com/users/sanchit-gandhi/orgs", "received_events_url": "https://api.github.com/users/sanchit-gandhi/received_events", "repos_url": "https://api.github.com/users/sanchit-gandhi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sanchit-gandhi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sanchit-gandhi/subscriptions", "type": "User", "url": "https://api.github.com/users/sanchit-gandhi" }
[]
null
completed
CONTRIBUTOR
2022-09-23T09:05:18Z
null
I_kwDODunzps5SaPKz
[ "Sorry bug on my part πŸ˜… Closing " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5011/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5011
https://github.com/huggingface/datasets/issues/5011
false
1,382,308,799
https://api.github.com/repos/huggingface/datasets/issues/5010/labels{/name}
Besides the current deprecation warning in the script of `multilingual_librispeech`, this PR adds a deprecation warning to its dataset card as well. The format of the deprecation warning is aligned with the one in the library documentation when docstrings contain the `<Deprecated/>` tag. Related to: - #4060
2022-09-23T12:04:37Z
5,010
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-22T11:41:59Z
https://api.github.com/repos/huggingface/datasets/issues/5010/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5010/timeline
Add deprecation warning to multilingual_librispeech dataset card
https://api.github.com/repos/huggingface/datasets/issues/5010/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-23T12:02:45Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5010.diff", "html_url": "https://github.com/huggingface/datasets/pull/5010", "merged_at": "2022-09-23T12:02:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/5010.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5010" }
PR_kwDODunzps4_bB3q
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5010/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5010
https://github.com/huggingface/datasets/pull/5010
true
1,381,194,067
https://api.github.com/repos/huggingface/datasets/issues/5009/labels{/name}
## Describe the bug I have added a new dataset with the identifier `StonyBrookNLP/tellmewhy` to the hub. When I load the individual files using my local copy using `dataset = datasets.load_dataset("json", data_files="data/train.jsonl")`, it loads the dataset correctly. However, when I try to load it from the hub, I get an error (pasted below). Additionally, `dataset = datasets.load_dataset("json", data_dir="data/")` throws the same error. ## Steps to reproduce the bug ```python dataset = datasets.load_dataset('StonyBrookNLP/tellmewhy') ``` ## Expected results Successfully load the `StonyBrookNLP/tellmewhy` dataset. ## Actual results ``` Using custom data configuration StonyBrookNLP--tellmewhy-82712924092694ff Downloading and preparing dataset json/StonyBrookNLP--tellmewhy to /home/yklal95/.cache/huggingface/datasets/StonyBrookNLP___json/StonyBrookNLP--tellmewhy-82712924092694ff/0.0.0/a3e658c4731e59120d44081ac10bf85dc7e1388126b92338344ce9661907f253... Downloading data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3/3 [00:00<00:00, 957.46it/s] Extracting data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3/3 [00:00<00:00, 299.14it/s] Traceback (most recent call last): File "/home/yklal95/tmw-generalization/src/load_datasets.py", line 17, in <module> main(args) File "/home/yklal95/tmw-generalization/src/load_datasets.py", line 11, in main dataset = datasets.load_dataset(args.dataset_name) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/load.py", line 1746, in load_dataset builder_instance.download_and_prepare( File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/builder.py", line 704, in download_and_prepare self._download_and_prepare( File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/builder.py", line 793, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/builder.py", line 1277, in _prepare_split writer.write_table(table) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/arrow_writer.py", line 524, in write_table pa_table = table_cast(pa_table, self._schema) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/table.py", line 2005, in table_cast return cast_table_to_schema(table, schema) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/table.py", line 1969, in cast_table_to_schema arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/table.py", line 1969, in <listcomp> arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/table.py", line 1681, in wrapper return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/table.py", line 1681, in <listcomp> return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/table.py", line 1822, in cast_array_to_feature casted_values = _c(array.values, feature.feature) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/table.py", line 1683, in wrapper return func(array, *args, **kwargs) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/table.py", line 1853, in cast_array_to_feature return array_cast(array, feature(), allow_number_to_str=allow_number_to_str) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/table.py", line 1683, in wrapper return func(array, *args, **kwargs) File "/home/yklal95/anaconda3/envs/tmw-generalization/lib/python3.9/site-packages/datasets/table.py", line 1761, in array_cast raise TypeError(f"Couldn't cast array of type {array.type} to {pa_type}") TypeError: Couldn't cast array of type int64 to null ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.4.0 - Platform: Linux-4.15.0-121-generic-x86_64-with-glibc2.27 - Python version: 3.9.13 - PyArrow version: 9.0.0 - Pandas version: 1.5.0
2022-09-29T13:07:29Z
5,009
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-21T16:23:06Z
https://api.github.com/repos/huggingface/datasets/issues/5009/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5009/timeline
Error loading StonyBrookNLP/tellmewhy dataset from hub even though local copy loads correctly
https://api.github.com/repos/huggingface/datasets/issues/5009/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/4996184?v=4", "events_url": "https://api.github.com/users/ykl7/events{/privacy}", "followers_url": "https://api.github.com/users/ykl7/followers", "following_url": "https://api.github.com/users/ykl7/following{/other_user}", "gists_url": "https://api.github.com/users/ykl7/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ykl7", "id": 4996184, "login": "ykl7", "node_id": "MDQ6VXNlcjQ5OTYxODQ=", "organizations_url": "https://api.github.com/users/ykl7/orgs", "received_events_url": "https://api.github.com/users/ykl7/received_events", "repos_url": "https://api.github.com/users/ykl7/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ykl7/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ykl7/subscriptions", "type": "User", "url": "https://api.github.com/users/ykl7" }
[]
null
completed
NONE
2022-09-29T13:07:29Z
null
I_kwDODunzps5SU1lT
[ "I think this is because some columns are mostly empty lists. In particular the train and validation splits only have empty lists for `val_ann`. Therefore the type inference doesn't know which type is inside (or it would have to scan the other splits first before knowing).\r\n\r\nYou can fix that by specifying the ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5009/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5009
https://github.com/huggingface/datasets/issues/5009
false
1,381,090,903
https://api.github.com/repos/huggingface/datasets/issues/5008/labels{/name}
Fixes the `filter` + `input_columns` combination, which is used in the `transformers` examples for instance. Revert #5006 (which in turn reverts #4971) Fix https://github.com/huggingface/datasets/issues/4858
2022-09-22T13:57:36Z
5,008
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-21T15:09:01Z
https://api.github.com/repos/huggingface/datasets/issues/5008/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5008/timeline
Re-apply input columns change
https://api.github.com/repos/huggingface/datasets/issues/5008/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-09-22T13:55:23Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5008.diff", "html_url": "https://github.com/huggingface/datasets/pull/5008", "merged_at": "2022-09-22T13:55:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/5008.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5008" }
PR_kwDODunzps4_XAc5
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5008/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5008
https://github.com/huggingface/datasets/pull/5008
true
1,381,007,607
https://api.github.com/repos/huggingface/datasets/issues/5007/labels{/name}
null
2022-09-22T10:16:14Z
5,007
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-21T14:14:25Z
https://api.github.com/repos/huggingface/datasets/issues/5007/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5007/timeline
Add some note about running the transformers ci before a release
https://api.github.com/repos/huggingface/datasets/issues/5007/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-22T10:14:06Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5007.diff", "html_url": "https://github.com/huggingface/datasets/pull/5007", "merged_at": "2022-09-22T10:14:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/5007.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5007" }
PR_kwDODunzps4_WvFQ
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5007/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5007
https://github.com/huggingface/datasets/pull/5007
true
1,380,968,395
https://api.github.com/repos/huggingface/datasets/issues/5006/labels{/name}
Revert https://github.com/huggingface/datasets/pull/4971 Fix https://github.com/huggingface/datasets/issues/5005
2022-09-21T14:14:33Z
5,006
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-21T13:49:20Z
https://api.github.com/repos/huggingface/datasets/issues/5006/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5006/timeline
Revert input_columns change
https://api.github.com/repos/huggingface/datasets/issues/5006/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-21T14:11:57Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5006.diff", "html_url": "https://github.com/huggingface/datasets/pull/5006", "merged_at": "2022-09-21T14:11:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/5006.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5006" }
PR_kwDODunzps4_Wm8z
[ "_The documentation is not available anymore as the PR was closed or merged._", "Merging this one and I'll check if it fixes the `transformers` CI before doing a patch release" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5006/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5006
https://github.com/huggingface/datasets/pull/5006
true
1,380,952,960
https://api.github.com/repos/huggingface/datasets/issues/5005/labels{/name}
## Describe the bug As reported by @lhoestq: > see https://app.circleci.com/pipelines/github/huggingface/transformers/47634/workflows/b491886b-e66e-4edb-af96-8b459e72aa25/jobs/564563 this is used here: [https://github.com/huggingface/transformers/blob/3b19c0317b6909e2d7f11b5053895ac55[…]torch/speech-pretraining/run_wav2vec2_pretraining_no_trainer.py](https://github.com/huggingface/transformers/blob/3b19c0317b6909e2d7f11b5053895ac55250e7da/examples/pytorch/speech-pretraining/run_wav2vec2_pretraining_no_trainer.py#L482-L488)
2022-09-21T14:11:57Z
5,005
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-21T13:39:19Z
https://api.github.com/repos/huggingface/datasets/issues/5005/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
https://api.github.com/repos/huggingface/datasets/issues/5005/timeline
Release 2.5.0 breaks transformers CI
https://api.github.com/repos/huggingface/datasets/issues/5005/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
completed
MEMBER
2022-09-21T14:11:57Z
null
I_kwDODunzps5ST6uA
[ "Shall we revert https://github.com/huggingface/datasets/pull/4971 @mariosasko ?\r\n\r\nAnd for consistency we can update IterableDataset.map later" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5005/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5005
https://github.com/huggingface/datasets/issues/5005
false
1,380,860,606
https://api.github.com/repos/huggingface/datasets/issues/5004/labels{/name}
As requested, we are removing the validation of the licenses from `datasets` because this is done on the Hub. Fix #4994. Related to: - #4926, which is removing all the validation from `datasets`
2022-09-22T11:47:41Z
5,004
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-21T12:35:14Z
https://api.github.com/repos/huggingface/datasets/issues/5004/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5004/timeline
Remove license tag file and validation
https://api.github.com/repos/huggingface/datasets/issues/5004/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-22T11:45:46Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5004.diff", "html_url": "https://github.com/huggingface/datasets/pull/5004", "merged_at": "2022-09-22T11:45:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/5004.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5004" }
PR_kwDODunzps4_WQck
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5004/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5004
https://github.com/huggingface/datasets/pull/5004
true
1,380,617,353
https://api.github.com/repos/huggingface/datasets/issues/5003/labels{/name}
This PRs fixes docstrings: - adds the missing `use_auth_token` param - updates syntax of param types - adds params to docstrings without them - fixes return/yield types - fixes syntax
2022-09-21T16:24:01Z
5,003
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-21T09:27:03Z
https://api.github.com/repos/huggingface/datasets/issues/5003/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5003/timeline
Fix missing use_auth_token in streaming docstrings
https://api.github.com/repos/huggingface/datasets/issues/5003/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-21T16:20:59Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5003.diff", "html_url": "https://github.com/huggingface/datasets/pull/5003", "merged_at": "2022-09-21T16:20:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/5003.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5003" }
PR_kwDODunzps4_Vdko
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5003/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5003
https://github.com/huggingface/datasets/pull/5003
true
1,380,589,402
https://api.github.com/repos/huggingface/datasets/issues/5002/labels{/name}
### Link https://huggingface.co/datasets/loubnabnl/humaneval-x/viewer/ ### Description The dataset has subsets but the viewer gets stuck in the default subset even when I select another one (the data loading of the subsets works fine) ### Owner Yes
2022-09-21T11:49:49Z
5,002
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
2022-09-21T09:06:17Z
https://api.github.com/repos/huggingface/datasets/issues/5002/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
https://api.github.com/repos/huggingface/datasets/issues/5002/timeline
Dataset Viewer issue for loubnabnl/humaneval-x
https://api.github.com/repos/huggingface/datasets/issues/5002/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/44069155?v=4", "events_url": "https://api.github.com/users/loubnabnl/events{/privacy}", "followers_url": "https://api.github.com/users/loubnabnl/followers", "following_url": "https://api.github.com/users/loubnabnl/following{/other_user}", "gists_url": "https://api.github.com/users/loubnabnl/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/loubnabnl", "id": 44069155, "login": "loubnabnl", "node_id": "MDQ6VXNlcjQ0MDY5MTU1", "organizations_url": "https://api.github.com/users/loubnabnl/orgs", "received_events_url": "https://api.github.com/users/loubnabnl/received_events", "repos_url": "https://api.github.com/users/loubnabnl/repos", "site_admin": false, "starred_url": "https://api.github.com/users/loubnabnl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/loubnabnl/subscriptions", "type": "User", "url": "https://api.github.com/users/loubnabnl" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url...
null
completed
NONE
2022-09-21T11:49:49Z
null
I_kwDODunzps5SSh9a
[ "It's a bug! Thanks for reporting, I'm looking at it", "Fixed." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 1, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5002/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5002
https://github.com/huggingface/datasets/issues/5002
false
1,379,844,820
https://api.github.com/repos/huggingface/datasets/issues/5001/labels{/name}
CC: @davanstrien
2022-11-01T12:44:42Z
5,001
null
https://api.github.com/repos/huggingface/datasets
true
[]
2022-09-20T18:42:58Z
https://api.github.com/repos/huggingface/datasets/issues/5001/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5001/timeline
Support loading XML datasets
https://api.github.com/repos/huggingface/datasets/issues/5001/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/5001.diff", "html_url": "https://github.com/huggingface/datasets/pull/5001", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5001.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5001" }
PR_kwDODunzps4_TBWa
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5001). All of your documentation changes will be reflected on that endpoint.", "> CC: @davanstrien\r\n\r\nI should have some time to look at this on Friday :) ", "@albertvillanova I've tried this with a few different XML data...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 2, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/5001/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5001
https://github.com/huggingface/datasets/pull/5001
true
1,379,709,398
https://api.github.com/repos/huggingface/datasets/issues/5000/labels{/name}
### Link https://huggingface.co/datasets/asapp/slue/viewer/ ### Description Hi, I wonder how to get the dataset viewer of our slue dataset to work. Best, Felix ### Owner Yes
2022-09-27T07:04:03Z
5,000
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-09-20T16:45:45Z
https://api.github.com/repos/huggingface/datasets/issues/5000/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5000/timeline
Dataset Viewer issue for asapp/slue
https://api.github.com/repos/huggingface/datasets/issues/5000/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/56092571?v=4", "events_url": "https://api.github.com/users/fwu-asapp/events{/privacy}", "followers_url": "https://api.github.com/users/fwu-asapp/followers", "following_url": "https://api.github.com/users/fwu-asapp/following{/other_user}", "gists_url": "https://api.github.com/users/fwu-asapp/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/fwu-asapp", "id": 56092571, "login": "fwu-asapp", "node_id": "MDQ6VXNlcjU2MDkyNTcx", "organizations_url": "https://api.github.com/users/fwu-asapp/orgs", "received_events_url": "https://api.github.com/users/fwu-asapp/received_events", "repos_url": "https://api.github.com/users/fwu-asapp/repos", "site_admin": false, "starred_url": "https://api.github.com/users/fwu-asapp/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fwu-asapp/subscriptions", "type": "User", "url": "https://api.github.com/users/fwu-asapp" }
[]
null
completed
NONE
2022-09-21T07:24:07Z
null
I_kwDODunzps5SPLHW
[ "<img width=\"519\" alt=\"Capture d’écran 2022-09-20 aΜ€ 22 33 47\" src=\"https://user-images.githubusercontent.com/1676121/191358952-1220cb7d-745a-4203-a66b-3c707b25038f.png\">\r\n\r\n```\r\nNot found.\r\n\r\nError code: SplitsResponseNotFound\r\n```\r\n\r\nhttps://datasets-server.huggingface.co/splits?dataset=a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5000/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5000
https://github.com/huggingface/datasets/issues/5000
false
1,379,610,030
https://api.github.com/repos/huggingface/datasets/issues/4999/labels{/name}
examples: from the hub: ```python Traceback (most recent call last): File "playground/ttest.py", line 3, in <module> print(load_dataset("lhoestq/empty")) File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/load.py", line 1686, in load_dataset **config_kwargs, File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/load.py", line 1458, in load_dataset_builder data_files=data_files, File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/load.py", line 1171, in dataset_module_factory raise e1 from None File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/load.py", line 1162, in dataset_module_factory download_mode=download_mode, File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/load.py", line 760, in get_module else get_data_patterns_in_dataset_repository(hfh_dataset_info, self.data_dir) File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/data_files.py", line 678, in get_data_patterns_in_dataset_repository ) from None datasets.data_files.EmptyDatasetError: The dataset repository at 'lhoestq/empty' doesn't contain any data file. ``` from local directory: ```python Traceback (most recent call last): File "playground/ttest.py", line 3, in <module> print(load_dataset("playground/empty")) File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/load.py", line 1686, in load_dataset **config_kwargs, File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/load.py", line 1458, in load_dataset_builder data_files=data_files, File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/load.py", line 1107, in dataset_module_factory path, data_dir=data_dir, data_files=data_files, download_mode=download_mode File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/load.py", line 625, in get_module else get_data_patterns_locally(base_path) File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/data_files.py", line 460, in get_data_patterns_locally raise EmptyDatasetError(f"The directory at {base_path} doesn't contain any data file") from None datasets.data_files.EmptyDatasetError: The directory at playground/empty doesn't contain any data file ``` Close https://github.com/huggingface/datasets/issues/4995
2022-09-21T12:23:43Z
4,999
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-20T15:28:05Z
https://api.github.com/repos/huggingface/datasets/issues/4999/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4999/timeline
Add EmptyDatasetError
https://api.github.com/repos/huggingface/datasets/issues/4999/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-21T12:21:24Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4999.diff", "html_url": "https://github.com/huggingface/datasets/pull/4999", "merged_at": "2022-09-21T12:21:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/4999.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4999" }
PR_kwDODunzps4_SQxL
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4999/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4999
https://github.com/huggingface/datasets/pull/4999
true
1,379,466,717
https://api.github.com/repos/huggingface/datasets/issues/4998/labels{/name}
Datasets with no namespace on the Hub have tags to redirect to the version of datasets where they come from. I’m about to remove them all because I think it looks bad/unexpected in the UI and it’s not actually useful Therefore I'm also disabling tagging. Note that the CI job will be completely removed in https://github.com/huggingface/datasets/pull/4974 anyway
2022-09-20T14:11:46Z
4,998
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-20T13:54:57Z
https://api.github.com/repos/huggingface/datasets/issues/4998/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4998/timeline
Don't add a tag on the Hub on release
https://api.github.com/repos/huggingface/datasets/issues/4998/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-20T14:08:54Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4998.diff", "html_url": "https://github.com/huggingface/datasets/pull/4998", "merged_at": "2022-09-20T14:08:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/4998.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4998" }
PR_kwDODunzps4_Ryp3
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4998/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4998
https://github.com/huggingface/datasets/pull/4998
true
1,379,430,711
https://api.github.com/repos/huggingface/datasets/issues/4997/labels{/name}
Support parsing JSON files in the array form (top-level object is an array). For simplicity, `json.load` is used for decoding. This means the entire file is loaded into memory. If requested, we can optimize this by introducing a param similar to `lines` in [`pandas.read_json`](https://pandas.pydata.org/docs/reference/api/pandas.read_json.html), which, if set to `True`, would allow us to read in chunks. Fixes https://github.com/huggingface/datasets/issues/4963
2022-09-20T15:42:40Z
4,997
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-20T13:31:26Z
https://api.github.com/repos/huggingface/datasets/issues/4997/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4997/timeline
Add support for parsing JSON files in array form
https://api.github.com/repos/huggingface/datasets/issues/4997/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-09-20T15:40:06Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4997.diff", "html_url": "https://github.com/huggingface/datasets/pull/4997", "merged_at": "2022-09-20T15:40:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/4997.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4997" }
PR_kwDODunzps4_RrBU
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4997/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4997
https://github.com/huggingface/datasets/pull/4997
true
1,379,345,161
https://api.github.com/repos/huggingface/datasets/issues/4996/labels{/name}
### Link https://huggingface.co/datasets/Jean-Baptiste/wikiner_fr ### Description ``` Error code: StreamingRowsError Exception: FileNotFoundError Message: [Errno 2] No such file or directory: 'zip:/data/train::https:/huggingface.co/datasets/Jean-Baptiste/wikiner_fr/resolve/main/data.zip/state.json' Traceback: Traceback (most recent call last): File "/src/services/worker/src/worker/responses/first_rows.py", line 337, in get_first_rows_response rows = get_rows(dataset, config, split, streaming=True, rows_max_number=rows_max_number, hf_token=hf_token) File "/src/services/worker/src/worker/utils.py", line 123, in decorator return func(*args, **kwargs) File "/src/services/worker/src/worker/responses/first_rows.py", line 77, in get_rows rows_plus_one = list(itertools.islice(ds, rows_max_number + 1)) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 718, in __iter__ for key, example in self._iter(): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 708, in _iter yield from ex_iterable File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 112, in __iter__ yield from self.generate_examples_fn(**self.kwargs) File "/tmp/modules-cache/datasets_modules/datasets/Jean-Baptiste--wikiner_fr/683a580ba6ec769d508f7dfc603a651667b0ed3817b1ae5bfd45f97cc024923f/wikiner_fr.py", line 165, in _generate_examples dataset = Dataset.load_from_disk(filepath) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 1210, in load_from_disk with open(Path(dataset_path, config.DATASET_STATE_JSON_FILENAME).as_posix(), encoding="utf-8") as state_file: FileNotFoundError: [Errno 2] No such file or directory: 'zip:/data/train::https:/huggingface.co/datasets/Jean-Baptiste/wikiner_fr/resolve/main/data.zip/state.json' ``` Is it an error with the dataset script, or the data itself, @huggingface/datasets? https://huggingface.co/datasets/Jean-Baptiste/wikiner_fr/tree/main ### Owner No
2022-09-27T12:35:44Z
4,996
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-09-20T12:32:07Z
https://api.github.com/repos/huggingface/datasets/issues/4996/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4996/timeline
Dataset Viewer issue for Jean-Baptiste/wikiner_fr
https://api.github.com/repos/huggingface/datasets/issues/4996/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[]
null
completed
CONTRIBUTOR
2022-09-27T12:35:44Z
null
I_kwDODunzps5SNyMJ
[ "The script uses `Dataset.load_from_disk`, which as you can expect, doesn't work in streaming mode.\r\n\r\nIt would probably be more practical to load the dataset locally using `Dataset.load_from_disk` first and then `push_to_hub` to upload it in Parquet on the Hub", "I've transferred this issue to the Hub repo: ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4996/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4996
https://github.com/huggingface/datasets/issues/4996
false
1,379,108,482
https://api.github.com/repos/huggingface/datasets/issues/4995/labels{/name}
In the dataset viewer on the Hub (https://huggingface.co/datasets/glue/viewer), we would like (https://github.com/huggingface/moon-landing/issues/3882) to show a specific message when the repository lacks any data files. In that case, instead of showing a complex traceback, we want to show a call to action to help the user upload data. To do that, it would be very helpful to know for sure that the repository is missing any (supported) data files. It could be done by raising a custom exception, for example, `NoDataError`.
2022-09-21T12:21:25Z
4,995
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "E5583E", "default": fals...
2022-09-20T09:31:59Z
https://api.github.com/repos/huggingface/datasets/issues/4995/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
https://api.github.com/repos/huggingface/datasets/issues/4995/timeline
Get a specific Exception when the dataset has no data
https://api.github.com/repos/huggingface/datasets/issues/4995/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
completed
CONTRIBUTOR
2022-09-21T12:21:25Z
null
I_kwDODunzps5SM4aC
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4995/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4995
https://github.com/huggingface/datasets/issues/4995
false
1,379,084,015
https://api.github.com/repos/huggingface/datasets/issues/4994/labels{/name}
> Feel free to delete the license list in `datasets` [...] > > Also FYI in #4926 I also removed all the validation steps anyway (language, license, types etc.) _Originally posted by @lhoestq in https://github.com/huggingface/datasets/issues/4930#issuecomment-1238401662_ > [...], in my opinion we can just delete this file from `datasets`, the validation is happening hub-side anyways now? _Originally posted by @julien-c in https://github.com/huggingface/datasets/issues/4930#issuecomment-1238390659_
2022-09-22T11:45:47Z
4,994
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-09-20T09:14:41Z
https://api.github.com/repos/huggingface/datasets/issues/4994/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4994/timeline
delete the hardcoded license list in `datasets`
https://api.github.com/repos/huggingface/datasets/issues/4994/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/326577?v=4", "events_url": "https://api.github.com/users/julien-c/events{/privacy}", "followers_url": "https://api.github.com/users/julien-c/followers", "following_url": "https://api.github.com/users/julien-c/following{/other_user}", "gists_url": "https://api.github.com/users/julien-c/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/julien-c", "id": 326577, "login": "julien-c", "node_id": "MDQ6VXNlcjMyNjU3Nw==", "organizations_url": "https://api.github.com/users/julien-c/orgs", "received_events_url": "https://api.github.com/users/julien-c/received_events", "repos_url": "https://api.github.com/users/julien-c/repos", "site_admin": false, "starred_url": "https://api.github.com/users/julien-c/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/julien-c/subscriptions", "type": "User", "url": "https://api.github.com/users/julien-c" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
MEMBER
2022-09-22T11:45:47Z
null
I_kwDODunzps5SMybv
[]
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4994/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4994
https://github.com/huggingface/datasets/issues/4994
false
1,379,044,435
https://api.github.com/repos/huggingface/datasets/issues/4993/labels{/name}
This PR updates features.py to avoid casting tuples to lists when reading the results of Dataset.map as suggested by @lhoestq [here](https://github.com/huggingface/datasets/issues/4676#issuecomment-1187371367) in https://github.com/huggingface/datasets/issues/4676.
2022-09-20T16:11:27Z
4,993
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-20T08:45:16Z
https://api.github.com/repos/huggingface/datasets/issues/4993/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4993/timeline
fix: avoid casting tuples after Dataset.map
https://api.github.com/repos/huggingface/datasets/issues/4993/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/5697926?v=4", "events_url": "https://api.github.com/users/szmoro/events{/privacy}", "followers_url": "https://api.github.com/users/szmoro/followers", "following_url": "https://api.github.com/users/szmoro/following{/other_user}", "gists_url": "https://api.github.com/users/szmoro/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/szmoro", "id": 5697926, "login": "szmoro", "node_id": "MDQ6VXNlcjU2OTc5MjY=", "organizations_url": "https://api.github.com/users/szmoro/orgs", "received_events_url": "https://api.github.com/users/szmoro/received_events", "repos_url": "https://api.github.com/users/szmoro/repos", "site_admin": false, "starred_url": "https://api.github.com/users/szmoro/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/szmoro/subscriptions", "type": "User", "url": "https://api.github.com/users/szmoro" }
[]
null
null
CONTRIBUTOR
2022-09-20T13:08:29Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4993.diff", "html_url": "https://github.com/huggingface/datasets/pull/4993", "merged_at": "2022-09-20T13:08:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/4993.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4993" }
PR_kwDODunzps4_QYas
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4993/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4993
https://github.com/huggingface/datasets/pull/4993
true
1,379,031,842
https://api.github.com/repos/huggingface/datasets/issues/4992/labels{/name}
Support streaming iwslt2017 dataset. Once this PR is merged: - [x] Remove old ".tgz" data files from the Hub.
2022-09-20T09:27:55Z
4,992
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-20T08:35:41Z
https://api.github.com/repos/huggingface/datasets/issues/4992/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4992/timeline
Support streaming iwslt2017 dataset
https://api.github.com/repos/huggingface/datasets/issues/4992/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-20T09:15:24Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4992.diff", "html_url": "https://github.com/huggingface/datasets/pull/4992", "merged_at": "2022-09-20T09:15:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/4992.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4992" }
PR_kwDODunzps4_QVw4
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4992/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4992
https://github.com/huggingface/datasets/pull/4992
true
1,378,898,752
https://api.github.com/repos/huggingface/datasets/issues/4991/labels{/name}
Fix missing tags in dataset cards: - aeslc - empathetic_dialogues - event2Mind - gap - iwslt2017 - newsgroup - qa4mre - scicite This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833 - #4891 - #4896 - #4908 - #4921 - #4931 - #4979
2022-09-22T12:25:32Z
4,991
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-20T06:42:07Z
https://api.github.com/repos/huggingface/datasets/issues/4991/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4991/timeline
Fix missing tags in dataset cards
https://api.github.com/repos/huggingface/datasets/issues/4991/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-20T07:37:30Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4991.diff", "html_url": "https://github.com/huggingface/datasets/pull/4991", "merged_at": "2022-09-20T07:37:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/4991.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4991" }
PR_kwDODunzps4_P5hI
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4991/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4991
https://github.com/huggingface/datasets/pull/4991
true
1,378,120,806
https://api.github.com/repos/huggingface/datasets/issues/4990/labels{/name}
## Describe the bug In the 2 lines listed below, a token is passed to `huggingface_hub` to get information from a dataset. If no token is provided, a "no-token" string is passed. What is the purpose of it ? If no real, I would prefer if the `None` value could be sent directly to be handle by `huggingface_hub`. I feel that here it is working because we assume the token will never be validated. https://github.com/huggingface/datasets/blob/5b23f58535f14cc4dd7649485bce1ccc836e7bca/src/datasets/load.py#L753 https://github.com/huggingface/datasets/blob/5b23f58535f14cc4dd7649485bce1ccc836e7bca/src/datasets/load.py#L1121 ## Expected results Pass `token=None` to `huggingface_hub`. ## Actual results `token="no-token"` is passed. ## Environment info `huggingface_hub v0.10.0dev`
2022-09-30T09:16:00Z
4,990
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-19T15:14:40Z
https://api.github.com/repos/huggingface/datasets/issues/4990/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
https://api.github.com/repos/huggingface/datasets/issues/4990/timeline
"no-token" is passed to `huggingface_hub` when token is `None`
https://api.github.com/repos/huggingface/datasets/issues/4990/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/11801849?v=4", "events_url": "https://api.github.com/users/Wauplin/events{/privacy}", "followers_url": "https://api.github.com/users/Wauplin/followers", "following_url": "https://api.github.com/users/Wauplin/following{/other_user}", "gists_url": "https://api.github.com/users/Wauplin/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Wauplin", "id": 11801849, "login": "Wauplin", "node_id": "MDQ6VXNlcjExODAxODQ5", "organizations_url": "https://api.github.com/users/Wauplin/orgs", "received_events_url": "https://api.github.com/users/Wauplin/received_events", "repos_url": "https://api.github.com/users/Wauplin/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Wauplin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wauplin/subscriptions", "type": "User", "url": "https://api.github.com/users/Wauplin" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
completed
CONTRIBUTOR
2022-09-30T09:16:00Z
null
I_kwDODunzps5SJHRm
[ "Hi @Wauplin, thanks for raising this potential issue.\r\n\r\nThe choice of passing `\"no-token\"` instead of `None` was made in this PR:\r\n- #4536 \r\n\r\nAccording to the PR description, the reason why it is passed is to avoid that `HfApi.dataset_info` uses the local token when no token should be used.", "Hi @...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4990/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4990
https://github.com/huggingface/datasets/issues/4990
false
1,376,832,233
https://api.github.com/repos/huggingface/datasets/issues/4989/labels{/name}
I have a dataset that contains a column ("foo") that is a sequence type of length 4. So when I run .to_pandas() on it, the resulting dataframe correctly contains 4 columns - foo_0, foo_1, foo_2, foo_3. So the 1st row of the dataframe might look like: ds = load_dataset(...) df = ds.to_pandas() df: foo_0 | foo_1 | foo_2 | foo_3 0.0 | 1.0 | 2.0 | 3.0 If I run .add_column("new_col", data) on the dataset, and then .to_pandas() on the resulting new dataset, the resulting dataframe contains only 2 columns - foo, new_col. The values in column foo are lists of length 4, the 4 elements that should have been split into separate columns. Dataframe 1st row would be: ds = load_dataset(...) new_ds = ds.add_column("new_col", data) df = new_ds.to_pandas() df: foo | new_col [0.0, 1.0, 2.0, 3.0] | new_val I've explored the 2 datasets in a debugger and haven't noticed any changes to any attributes related to the foo column, but I can't determine why the dataframes are so different.
2022-09-19T12:54:54Z
4,989
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-17T17:42:05Z
https://api.github.com/repos/huggingface/datasets/issues/4989/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4989/timeline
Running add_column() seems to corrupt existing sequence-type column info
https://api.github.com/repos/huggingface/datasets/issues/4989/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/93728165?v=4", "events_url": "https://api.github.com/users/derek-rocheleau/events{/privacy}", "followers_url": "https://api.github.com/users/derek-rocheleau/followers", "following_url": "https://api.github.com/users/derek-rocheleau/following{/other_user}", "gists_url": "https://api.github.com/users/derek-rocheleau/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/derek-rocheleau", "id": 93728165, "login": "derek-rocheleau", "node_id": "U_kgDOBZYtpQ", "organizations_url": "https://api.github.com/users/derek-rocheleau/orgs", "received_events_url": "https://api.github.com/users/derek-rocheleau/received_events", "repos_url": "https://api.github.com/users/derek-rocheleau/repos", "site_admin": false, "starred_url": "https://api.github.com/users/derek-rocheleau/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/derek-rocheleau/subscriptions", "type": "User", "url": "https://api.github.com/users/derek-rocheleau" }
[]
null
completed
NONE
2022-09-19T12:54:54Z
null
I_kwDODunzps5SEMrp
[ "Nevermind, I was incorrect." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4989/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4989
https://github.com/huggingface/datasets/issues/4989
false
1,376,096,584
https://api.github.com/repos/huggingface/datasets/issues/4988/labels{/name}
We've just added `Dataset.from_generator` to the API. It would also be cool to add `IterableDataset.from_generator` to support creating an iterable dataset from a generator. cc @lhoestq
2022-10-05T12:10:49Z
4,988
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "7057ff", "default": true...
2022-09-16T15:19:41Z
https://api.github.com/repos/huggingface/datasets/issues/4988/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/56002455?v=4", "events_url": "https://api.github.com/users/hamid-vakilzadeh/events{/privacy}", "followers_url": "https://api.github.com/users/hamid-vakilzadeh/followers", "following_url": "https://api.github.com/users/hamid-vakilzadeh/following{/other_user}", "gists_url": "https://api.github.com/users/hamid-vakilzadeh/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hamid-vakilzadeh", "id": 56002455, "login": "hamid-vakilzadeh", "node_id": "MDQ6VXNlcjU2MDAyNDU1", "organizations_url": "https://api.github.com/users/hamid-vakilzadeh/orgs", "received_events_url": "https://api.github.com/users/hamid-vakilzadeh/received_events", "repos_url": "https://api.github.com/users/hamid-vakilzadeh/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hamid-vakilzadeh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hamid-vakilzadeh/subscriptions", "type": "User", "url": "https://api.github.com/users/hamid-vakilzadeh" }
https://api.github.com/repos/huggingface/datasets/issues/4988/timeline
Add `IterableDataset.from_generator` to the API
https://api.github.com/repos/huggingface/datasets/issues/4988/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/56002455?v=4", "events_url": "https://api.github.com/users/hamid-vakilzadeh/events{/privacy}", "followers_url": "https://api.github.com/users/hamid-vakilzadeh/followers", "following_url": "https://api.github.com/users/hamid-vakilzadeh/followin...
null
completed
CONTRIBUTOR
2022-10-05T12:10:49Z
null
I_kwDODunzps5SBZFI
[ "#take", "Thanks @hamid-vakilzadeh ! Let us know if you have some questions or if we can help", "Thank you! I certainly will reach out if I need any help." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4988/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4988
https://github.com/huggingface/datasets/issues/4988
false
1,376,006,477
https://api.github.com/repos/huggingface/datasets/issues/4987/labels{/name}
Embed the bytes of the image or audio files in the Parquet files directly, instead of having a "path" that points to a local file. Indeed Parquet files are often used to share data or to be used by workers that may not have access to the local files.
2022-09-16T16:24:47Z
4,987
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-16T14:09:27Z
https://api.github.com/repos/huggingface/datasets/issues/4987/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4987/timeline
Embed image/audio data in dl_and_prepare parquet
https://api.github.com/repos/huggingface/datasets/issues/4987/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-16T16:22:35Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4987.diff", "html_url": "https://github.com/huggingface/datasets/pull/4987", "merged_at": "2022-09-16T16:22:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/4987.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4987" }
PR_kwDODunzps4_GlIu
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4987/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4987
https://github.com/huggingface/datasets/pull/4987
true