url
stringlengths
61
61
repository_url
stringclasses
1 value
labels_url
stringlengths
75
75
comments_url
stringlengths
70
70
events_url
stringlengths
68
68
html_url
stringlengths
49
51
id
int64
758M
1.95B
node_id
stringlengths
18
32
number
int64
1.2k
6.31k
title
stringlengths
1
290
user
dict
labels
listlengths
0
3
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
listlengths
0
30
created_at
timestamp[ns, tz=UTC]
updated_at
timestamp[ns, tz=UTC]
closed_at
timestamp[ns, tz=UTC]
author_association
stringclasses
3 values
active_lock_reason
float64
draft
float64
0
1
pull_request
dict
body
stringlengths
0
36.2k
reactions
dict
timeline_url
stringlengths
70
70
performed_via_github_app
float64
state_reason
stringclasses
3 values
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/4253
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4253/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4253/comments
https://api.github.com/repos/huggingface/datasets/issues/4253/events
https://github.com/huggingface/datasets/pull/4253
1,219,286,408
PR_kwDODunzps42-c8Q
4,253
Create metric cards for mean IOU
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-04-28T20:58:27Z
2022-04-29T17:44:47Z
2022-04-29T17:38:06Z
NONE
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4253.diff", "html_url": "https://github.com/huggingface/datasets/pull/4253", "merged_at": "2022-04-29T17:38:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/4253.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4253" }
Proposing a metric card for mIoU :rocket: sorry for spamming you with review requests, @albertvillanova ! :hugs:
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4253/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4253/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4809
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4809/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4809/comments
https://api.github.com/repos/huggingface/datasets/issues/4809/events
https://github.com/huggingface/datasets/pull/4809
1,332,842,747
PR_kwDODunzps483Y4h
4,809
Complete the mlqa dataset card
{ "avatar_url": "https://avatars.githubusercontent.com/u/7940237?v=4", "events_url": "https://api.github.com/users/el2e10/events{/privacy}", "followers_url": "https://api.github.com/users/el2e10/followers", "following_url": "https://api.github.com/users/el2e10/following{/other_user}", "gists_url": "https://api.github.com/users/el2e10/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/el2e10", "id": 7940237, "login": "el2e10", "node_id": "MDQ6VXNlcjc5NDAyMzc=", "organizations_url": "https://api.github.com/users/el2e10/orgs", "received_events_url": "https://api.github.com/users/el2e10/received_events", "repos_url": "https://api.github.com/users/el2e10/repos", "site_admin": false, "starred_url": "https://api.github.com/users/el2e10/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/el2e10/subscriptions", "type": "User", "url": "https://api.github.com/users/el2e10" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "> Thanks for your contribution, @eldhoittangeorge.\r\n> \r\n> The CI error message: https://github.com/huggingface/datasets/runs/7743526624?check_suite_focus=true\r\n> \r\n> ```\r\n> E ValueError: The following issues have ...
2022-08-09T07:38:06Z
2022-08-09T16:26:21Z
2022-08-09T13:26:43Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4809.diff", "html_url": "https://github.com/huggingface/datasets/pull/4809", "merged_at": "2022-08-09T13:26:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/4809.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4809" }
I fixed the issue #4808 Details of PR: - Added languages included in the dataset. - Added task id and task category. - Updated the citation information. Fix #4808.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4809/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4809/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3760
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3760/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3760/comments
https://api.github.com/repos/huggingface/datasets/issues/3760/events
https://github.com/huggingface/datasets/issues/3760
1,144,804,558
I_kwDODunzps5EPFTO
3,760
Unable to view the Gradio flagged call back dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/36753484?v=4", "events_url": "https://api.github.com/users/kingabzpro/events{/privacy}", "followers_url": "https://api.github.com/users/kingabzpro/followers", "following_url": "https://api.github.com/users/kingabzpro/following{/other_user}", "gists_url": "https://api.github.com/users/kingabzpro/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kingabzpro", "id": 36753484, "login": "kingabzpro", "node_id": "MDQ6VXNlcjM2NzUzNDg0", "organizations_url": "https://api.github.com/users/kingabzpro/orgs", "received_events_url": "https://api.github.com/users/kingabzpro/received_events", "repos_url": "https://api.github.com/users/kingabzpro/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kingabzpro/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kingabzpro/subscriptions", "type": "User", "url": "https://api.github.com/users/kingabzpro" }
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
closed
false
null
[]
null
[ "Hi @kingabzpro.\r\n\r\nI think you need to create a loading script that creates the dataset from the CSV file and the image paths.\r\n\r\nAs example, you could have a look at the Food-101 dataset: https://huggingface.co/datasets/food101\r\n- Loading script: https://huggingface.co/datasets/food101/blob/main/food101...
2022-02-19T17:45:08Z
2022-03-22T07:12:11Z
2022-03-22T07:12:11Z
NONE
null
null
null
## Dataset viewer issue for '*savtadepth-flags*' **Link:** *[savtadepth-flags](https://huggingface.co/datasets/kingabzpro/savtadepth-flags)* *with the Gradio 2.8.1 the dataset viers stopped working. I tried to add values manually but its not working. The dataset is also not showing the link with the app https://huggingface.co/spaces/kingabzpro/savtadepth.* Am I the one who added this dataset ? Yes
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3760/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3760/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/1948
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1948/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1948/comments
https://api.github.com/repos/huggingface/datasets/issues/1948/events
https://github.com/huggingface/datasets/issues/1948
816,689,329
MDU6SXNzdWU4MTY2ODkzMjk=
1,948
dataset loading logger level
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stas00", "id": 10676103, "login": "stas00", "node_id": "MDQ6VXNlcjEwNjc2MTAz", "organizations_url": "https://api.github.com/users/stas00/orgs", "received_events_url": "https://api.github.com/users/stas00/received_events", "repos_url": "https://api.github.com/users/stas00/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "type": "User", "url": "https://api.github.com/users/stas00" }
[]
closed
false
null
[]
null
[ "These warnings are showed when there's a call to `.map` to say to the user that a dataset is reloaded from the cache instead of being recomputed.\r\nThey are warnings since we want to make sure the users know that it's not recomputed.", "Thank you for explaining the intention, @lhoestq \r\n\r\n1. Could it be the...
2021-02-25T18:33:37Z
2023-07-12T17:19:30Z
2023-07-12T17:19:30Z
CONTRIBUTOR
null
null
null
on master I get this with `--dataset_name wmt16 --dataset_config ro-en`: ``` WARNING:datasets.arrow_dataset:Loading cached processed dataset at /home/stas/.cache/huggingface/datasets/wmt16/ro-en/1.0.0/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f/cache-2e01bead8cf42e26.arrow WARNING:datasets.arrow_dataset:Loading cached processed dataset at /home/stas/.cache/huggingface/datasets/wmt16/ro-en/1.0.0/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f/cache-ac3bebaf4f91f776.arrow WARNING:datasets.arrow_dataset:Loading cached processed dataset at /home/stas/.cache/huggingface/datasets/wmt16/ro-en/1.0.0/9dc00622c30446e99c4c63d12a484ea4fb653f2f37c867d6edcec839d7eae50f/cache-810c3e61259d73a9.arrow ``` why are those WARNINGs? Should be INFO, no? warnings should only be used when a user needs to pay attention to something, this is just informative - I'd even say it should be DEBUG, but definitely not WARNING. Thank you.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1948/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1948/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/5276
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5276/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5276/comments
https://api.github.com/repos/huggingface/datasets/issues/5276/events
https://github.com/huggingface/datasets/issues/5276
1,459,363,442
I_kwDODunzps5W_B5y
5,276
Bug in downloading common_voice data and snall chunk of it to one's own hub
{ "avatar_url": "https://avatars.githubusercontent.com/u/48530104?v=4", "events_url": "https://api.github.com/users/capsabogdan/events{/privacy}", "followers_url": "https://api.github.com/users/capsabogdan/followers", "following_url": "https://api.github.com/users/capsabogdan/following{/other_user}", "gists_url": "https://api.github.com/users/capsabogdan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/capsabogdan", "id": 48530104, "login": "capsabogdan", "node_id": "MDQ6VXNlcjQ4NTMwMTA0", "organizations_url": "https://api.github.com/users/capsabogdan/orgs", "received_events_url": "https://api.github.com/users/capsabogdan/received_events", "repos_url": "https://api.github.com/users/capsabogdan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/capsabogdan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/capsabogdan/subscriptions", "type": "User", "url": "https://api.github.com/users/capsabogdan" }
[]
closed
false
null
[]
null
[ "Sounds like one of the file is not a valid one, can you make sure you uploaded valid mp3 files ?", "Well I just sharded the original commonVoice dataset and pushed a small chunk of it in a private rep\n\nWhat did go wrong?\n\nHolen Sie sich Outlook für iOS<https://aka.ms/o0ukef>\n________________________________...
2022-11-22T08:17:53Z
2023-07-21T14:33:10Z
2023-07-21T14:33:10Z
NONE
null
null
null
### Describe the bug I'm trying to load the common voice dataset. Currently there is no implementation to download just par tof the data, and I need just one part of it, without downloading the entire dataset Help please? ![image](https://user-images.githubusercontent.com/48530104/203260511-26df766f-6013-4eaf-be26-8aa13794def2.png) ### Steps to reproduce the bug So here is what I have done: 1. Download common_voice data 2. Trim part of it and publish it to my own repo. 3. Download data from my own repo, but am getting this error. ### Expected behavior There shouldn't be an error in downloading part of the data and publishing it to one's own repo ### Environment info common_voice 11
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5276/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5276/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/4692
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4692/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4692/comments
https://api.github.com/repos/huggingface/datasets/issues/4692/events
https://github.com/huggingface/datasets/issues/4692
1,306,609,680
I_kwDODunzps5N4UgQ
4,692
Unable to cast a column with `Image()` by using the `cast_column()` feature
{ "avatar_url": "https://avatars.githubusercontent.com/u/28833916?v=4", "events_url": "https://api.github.com/users/skrishnan99/events{/privacy}", "followers_url": "https://api.github.com/users/skrishnan99/followers", "following_url": "https://api.github.com/users/skrishnan99/following{/other_user}", "gists_url": "https://api.github.com/users/skrishnan99/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/skrishnan99", "id": 28833916, "login": "skrishnan99", "node_id": "MDQ6VXNlcjI4ODMzOTE2", "organizations_url": "https://api.github.com/users/skrishnan99/orgs", "received_events_url": "https://api.github.com/users/skrishnan99/received_events", "repos_url": "https://api.github.com/users/skrishnan99/repos", "site_admin": false, "starred_url": "https://api.github.com/users/skrishnan99/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/skrishnan99/subscriptions", "type": "User", "url": "https://api.github.com/users/skrishnan99" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[ "Hi, thanks for reporting! A PR (https://github.com/huggingface/datasets/pull/4614) has already been opened to address this issue." ]
2022-07-15T22:56:03Z
2022-07-19T13:36:24Z
2022-07-19T13:36:24Z
NONE
null
null
null
## Describe the bug A clear and concise description of what the bug is. When I create a dataset, then add a column to the created dataset through the `dataset.add_column` feature and then try to cast a column of the dataset (this column contains image paths) with `Image()` by using the `cast_column()` feature, I get the following error - ``` TypeError: Couldn't cast array of type string to {'bytes': Value(dtype='binary', id=None), 'path': Value(dtype='string', id=None)} ``` When I try and cast the same column, but without doing the `add_column` in the previous step, it works as expected. ## Steps to reproduce the bug ```python from datasets import Dataset, Image data_dict = { "img_path": ["https://picsum.photos/200/300"] } dataset = Dataset.from_dict(data_dict) #NOTE Comment out this line and use cast_column and it works properly dataset = dataset.add_column("yeet", [1]) #NOTE This line fails to execute properly if `add_column` is called before dataset = dataset.cast_column("img_path", Image()) # #NOTE This is my current workaround. This seems to work fine with/without `add_column`. While # # running this, make sure to comment out the `cast_column` line # new_features = dataset.features.copy() # new_features["img_path"] = Image() # dataset = dataset.cast(new_features) print(dataset) print(dataset.features) print(dataset[0]) ``` ## Expected results A clear and concise description of the expected results. Able to successfully use `cast_column` to cast a column containing img_paths to now be Image() features after modifying the dataset using `add_column` in a previous step ## Actual results Specify the actual results or traceback. ``` Traceback (most recent call last): File "/home/surya/Desktop/hf_bug_test.py", line 14, in <module> dataset = dataset.cast_column("img_path", Image()) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/fingerprint.py", line 458, in wrapper out = func(self, *args, **kwargs) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 1580, in cast_column dataset._data = dataset._data.cast(dataset.features.arrow_schema) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1487, in cast new_tables.append(subtable.cast(subschema, *args, **kwargs)) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 834, in cast return InMemoryTable(table_cast(self.table, *args, **kwargs)) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1897, in table_cast return cast_table_to_schema(table, schema) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1880, in cast_table_to_schema arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1880, in <listcomp> arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1673, in wrapper return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1673, in <listcomp> return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "/home/surya/anaconda3/envs/snap_test/lib/python3.9/site-packages/datasets/table.py", line 1846, in cast_array_to_feature raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{feature}") TypeError: Couldn't cast array of type string to {'bytes': Value(dtype='binary', id=None), 'path': Value(dtype='string', id=None)} ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.3.2 - Platform: Ubuntu 20.04.3 LTS - Python version: 3.9.7 - PyArrow version: 7.0.0
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4692/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4692/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/2066
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2066/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2066/comments
https://api.github.com/repos/huggingface/datasets/issues/2066/events
https://github.com/huggingface/datasets/pull/2066
833,480,551
MDExOlB1bGxSZXF1ZXN0NTk0NDcwMjEz
2,066
Fix docstring rendering of Dataset/DatasetDict.from_csv args
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2021-03-17T07:23:10Z
2021-03-17T09:21:21Z
2021-03-17T09:21:21Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/2066.diff", "html_url": "https://github.com/huggingface/datasets/pull/2066", "merged_at": "2021-03-17T09:21:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/2066.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2066" }
Fix the docstring rendering of Dataset/DatasetDict.from_csv args.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2066/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2066/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4500
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4500/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4500/comments
https://api.github.com/repos/huggingface/datasets/issues/4500/events
https://github.com/huggingface/datasets/pull/4500
1,272,281,992
PR_kwDODunzps45tdxk
4,500
Add `concatenate_datasets` for iterable datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks ! I addressed your comments :)\r\n\r\n> There is a slight difference in concatenate_datasets between the version for map-style datasets and the one for iterable datasets\r\n\r\nIndeed, here is what I did to fix this:\r\n\r\n- ...
2022-06-15T13:58:50Z
2022-06-28T21:25:39Z
2022-06-28T21:15:04Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4500.diff", "html_url": "https://github.com/huggingface/datasets/pull/4500", "merged_at": "2022-06-28T21:15:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/4500.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4500" }
`concatenate_datasets` currently only supports lists of `datasets.Dataset`, not lists of `datasets.IterableDataset` like `interleave_datasets` Fix https://github.com/huggingface/datasets/issues/2564 I also moved `_interleave_map_style_datasets` from combine.py to arrow_dataset.py, since the logic depends a lot on the `Dataset` object internals And I moved `concatenate_datasets` from arrow_dataset.py to combine.py to have it with `interleave_datasets` (though it's also copied in arrow_dataset module for backward compatibility for now)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4500/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4500/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2097
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2097/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2097/comments
https://api.github.com/repos/huggingface/datasets/issues/2097/events
https://github.com/huggingface/datasets/pull/2097
838,105,289
MDExOlB1bGxSZXF1ZXN0NTk4MzM4MTA3
2,097
fixes issue #1110 by descending further if `obj["_type"]` is a dict
{ "avatar_url": "https://avatars.githubusercontent.com/u/15979778?v=4", "events_url": "https://api.github.com/users/dcfidalgo/events{/privacy}", "followers_url": "https://api.github.com/users/dcfidalgo/followers", "following_url": "https://api.github.com/users/dcfidalgo/following{/other_user}", "gists_url": "https://api.github.com/users/dcfidalgo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dcfidalgo", "id": 15979778, "login": "dcfidalgo", "node_id": "MDQ6VXNlcjE1OTc5Nzc4", "organizations_url": "https://api.github.com/users/dcfidalgo/orgs", "received_events_url": "https://api.github.com/users/dcfidalgo/received_events", "repos_url": "https://api.github.com/users/dcfidalgo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dcfidalgo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dcfidalgo/subscriptions", "type": "User", "url": "https://api.github.com/users/dcfidalgo" }
[]
closed
false
null
[]
null
[]
2021-03-22T21:00:55Z
2021-03-22T21:01:11Z
2021-03-22T21:01:11Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/2097.diff", "html_url": "https://github.com/huggingface/datasets/pull/2097", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2097.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2097" }
Check metrics
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2097/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2097/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4071
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4071/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4071/comments
https://api.github.com/repos/huggingface/datasets/issues/4071/events
https://github.com/huggingface/datasets/issues/4071
1,187,587,683
I_kwDODunzps5GySZj
4,071
Loading issue for xuyeliu/notebookCDG dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/46160972?v=4", "events_url": "https://api.github.com/users/Jun-jie-Huang/events{/privacy}", "followers_url": "https://api.github.com/users/Jun-jie-Huang/followers", "following_url": "https://api.github.com/users/Jun-jie-Huang/following{/other_user}", "gists_url": "https://api.github.com/users/Jun-jie-Huang/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Jun-jie-Huang", "id": 46160972, "login": "Jun-jie-Huang", "node_id": "MDQ6VXNlcjQ2MTYwOTcy", "organizations_url": "https://api.github.com/users/Jun-jie-Huang/orgs", "received_events_url": "https://api.github.com/users/Jun-jie-Huang/received_events", "repos_url": "https://api.github.com/users/Jun-jie-Huang/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Jun-jie-Huang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Jun-jie-Huang/subscriptions", "type": "User", "url": "https://api.github.com/users/Jun-jie-Huang" }
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
[]
null
[ "Hi @Jun-jie-Huang,\r\n\r\nAs the error message says, \".pkl\" data files are not supported.\r\n\r\nIf you would like to share your dataset on the Hub, you would need:\r\n- either to create a Python loading script, that loads the data in any format\r\n- or to transform your data files to one of the supported format...
2022-03-31T06:36:29Z
2022-03-31T08:17:01Z
2022-03-31T08:16:16Z
NONE
null
null
null
## Dataset viewer issue for '*xuyeliu/notebookCDG*' **Link:** *[link to the dataset viewer page](https://huggingface.co/datasets/xuyeliu/notebookCDG)* *Couldn't load the xuyeliu/notebookCDG with provided scripts: * ``` from datasets import load_dataset dataset = load_dataset("xuyeliu/notebookCDG/dataset_notebook.pkl") ``` I get an error message as follows: FileNotFoundError: Couldn't find a dataset script at /home/code_documentation/code/xuyeliu/notebookCDG/notebookCDG.py or any data file in the same directory. Couldn't find 'xuyeliu/notebookCDG' on the Hugging Face Hub either: FileNotFoundError: Unable to resolve any data file that matches ['**train*'] in dataset repository xuyeliu/notebookCDG with any supported extension ['csv', 'tsv', 'json', 'jsonl', 'parquet', 'txt', 'blp', 'bmp', 'dib', 'bufr', 'cur', 'pcx', 'dcx', 'dds', 'ps', 'eps', 'fit', 'fits', 'fli', 'flc', 'ftc', 'ftu', 'gbr', 'gif', 'grib', 'h5', 'hdf', 'png', 'apng', 'jp2', 'j2k', 'jpc', 'jpf', 'jpx', 'j2c', 'icns', 'ico', 'im', 'iim', 'tif', 'tiff', 'jfif', 'jpe', 'jpg', 'jpeg', 'mpg', 'mpeg', 'msp', 'pcd', 'pxr', 'pbm', 'pgm', 'ppm', 'pnm', 'psd', 'bw', 'rgb', 'rgba', 'sgi', 'ras', 'tga', 'icb', 'vda', 'vst', 'webp', 'wmf', 'emf', 'xbm', 'xpm', 'zip'] Am I the one who added this dataset ? No
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4071/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4071/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/2522
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2522/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2522/comments
https://api.github.com/repos/huggingface/datasets/issues/2522/events
https://github.com/huggingface/datasets/issues/2522
925,334,379
MDU6SXNzdWU5MjUzMzQzNzk=
2,522
Documentation Mistakes in Dataset: emotion
{ "avatar_url": "https://avatars.githubusercontent.com/u/62606251?v=4", "events_url": "https://api.github.com/users/GDGauravDutta/events{/privacy}", "followers_url": "https://api.github.com/users/GDGauravDutta/followers", "following_url": "https://api.github.com/users/GDGauravDutta/following{/other_user}", "gists_url": "https://api.github.com/users/GDGauravDutta/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/GDGauravDutta", "id": 62606251, "login": "GDGauravDutta", "node_id": "MDQ6VXNlcjYyNjA2MjUx", "organizations_url": "https://api.github.com/users/GDGauravDutta/orgs", "received_events_url": "https://api.github.com/users/GDGauravDutta/received_events", "repos_url": "https://api.github.com/users/GDGauravDutta/repos", "site_admin": false, "starred_url": "https://api.github.com/users/GDGauravDutta/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/GDGauravDutta/subscriptions", "type": "User", "url": "https://api.github.com/users/GDGauravDutta" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[ "Hi,\r\n\r\nthis issue has been already reported in the dataset repo (https://github.com/dair-ai/emotion_dataset/issues/2), so this is a bug on their side.", "The documentation has another bug in the dataset card [here](https://huggingface.co/datasets/emotion). \r\n\r\nIn the dataset summary **six** emotions are ...
2021-06-19T07:08:57Z
2023-01-02T12:04:58Z
2023-01-02T12:04:58Z
NONE
null
null
null
As per documentation, Dataset: emotion Homepage: https://github.com/dair-ai/emotion_dataset Dataset: https://github.com/huggingface/datasets/blob/master/datasets/emotion/emotion.py Permalink: https://huggingface.co/datasets/viewer/?dataset=emotion Emotion is a dataset of English Twitter messages with eight basic emotions: anger, anticipation, disgust, fear, joy, sadness, surprise, and trust. For more detailed information please refer to the paper. But when we view the data, there are only 6 emotions, anger, fear, joy, sadness, surprise, and trust.
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2522/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2522/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/5198
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5198/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5198/comments
https://api.github.com/repos/huggingface/datasets/issues/5198/events
https://github.com/huggingface/datasets/pull/5198
1,434,699,165
PR_kwDODunzps5CI49J
5,198
Add note about the name of a dataset script
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_use...
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-11-03T13:51:32Z
2022-11-04T12:47:59Z
2022-11-04T12:46:01Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/5198.diff", "html_url": "https://github.com/huggingface/datasets/pull/5198", "merged_at": "2022-11-04T12:46:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/5198.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5198" }
Add note that a dataset script should has the same name as a repo/dir, a bit related to this issue https://github.com/huggingface/datasets/issues/5193 also fixed two minor issues in audio docs (broken links)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5198/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5198/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/5221
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5221/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5221/comments
https://api.github.com/repos/huggingface/datasets/issues/5221/events
https://github.com/huggingface/datasets/issues/5221
1,442,309,094
I_kwDODunzps5V9-Pm
5,221
Cannot push
{ "avatar_url": "https://avatars.githubusercontent.com/u/43239645?v=4", "events_url": "https://api.github.com/users/bayartsogt-ya/events{/privacy}", "followers_url": "https://api.github.com/users/bayartsogt-ya/followers", "following_url": "https://api.github.com/users/bayartsogt-ya/following{/other_user}", "gists_url": "https://api.github.com/users/bayartsogt-ya/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bayartsogt-ya", "id": 43239645, "login": "bayartsogt-ya", "node_id": "MDQ6VXNlcjQzMjM5NjQ1", "organizations_url": "https://api.github.com/users/bayartsogt-ya/orgs", "received_events_url": "https://api.github.com/users/bayartsogt-ya/received_events", "repos_url": "https://api.github.com/users/bayartsogt-ya/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bayartsogt-ya/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bayartsogt-ya/subscriptions", "type": "User", "url": "https://api.github.com/users/bayartsogt-ya" }
[]
closed
false
null
[]
null
[ "Did you run `huggingface-cli lfs-enable-largefiles` before committing or before adding ? Maybe you can try before adding\r\n\r\nAnyway I'd encourage you to split your data into several TAR archives if possible, this way the dataset can loaded faster using multiprocessing (by giving each process a subset of shards ...
2022-11-09T15:32:05Z
2022-11-10T18:11:21Z
2022-11-10T18:11:11Z
NONE
null
null
null
### Describe the bug I am facing the issue when I try to push the tar.gz file around 11G to HUB. ``` (venv) ╭─laptop@laptop ~/PersonalProjects/data/ulaanbal_v0 ‹main●› ╰─$ du -sh * 4.0K README.md 13G data 516K test.jsonl 18M train.jsonl 4.0K ulaanbal_v0.py 11G ulaanbal_v0.tar.gz 452K validation.jsonl (venv) ╭─laptop@laptop~/PersonalProjects/data/ulaanbal_v0 ‹main●› ╰─$ git add ulaanbal_v0.tar.gz && git commit -m 'large version' (venv) ╭─laptop@laptop ~/PersonalProjects/data/ulaanbal_v0 ‹main●› ╰─$ git push EOFoading LFS objects: 0% (0/1), 0 B | 0 B/s Uploading LFS objects: 0% (0/1), 0 B | 0 B/s, done. error: failed to push some refs to 'https://huggingface.co/datasets/bayartsogt/ulaanbal_v0' ``` I have already tried pushing a small version of this and it was working fine. So my guess it is probably because of the big file. Following I run before the commit: ``` ╰─$ git lfs install ╰─$ huggingface-cli lfs-enable-largefiles . ``` ### Steps to reproduce the bug Create a private dataset on huggingface and push 12G tar.gz file ### Expected behavior To be pushed with no issue ### Environment info - `datasets` version: 2.6.1 - Platform: Darwin-21.6.0-x86_64-i386-64bit - Python version: 3.7.11 - PyArrow version: 10.0.0 - Pandas version: 1.3.5
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5221/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5221/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/4248
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4248/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4248/comments
https://api.github.com/repos/huggingface/datasets/issues/4248/events
https://github.com/huggingface/datasets/issues/4248
1,218,460,444
I_kwDODunzps5IoDsc
4,248
conll2003 dataset loads original data.
{ "avatar_url": "https://avatars.githubusercontent.com/u/26458611?v=4", "events_url": "https://api.github.com/users/sue991/events{/privacy}", "followers_url": "https://api.github.com/users/sue991/followers", "following_url": "https://api.github.com/users/sue991/following{/other_user}", "gists_url": "https://api.github.com/users/sue991/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sue991", "id": 26458611, "login": "sue991", "node_id": "MDQ6VXNlcjI2NDU4NjEx", "organizations_url": "https://api.github.com/users/sue991/orgs", "received_events_url": "https://api.github.com/users/sue991/received_events", "repos_url": "https://api.github.com/users/sue991/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sue991/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sue991/subscriptions", "type": "User", "url": "https://api.github.com/users/sue991" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
[ "Thanks for reporting @sue99.\r\n\r\nUnfortunately. I'm not able to reproduce your problem:\r\n```python\r\nIn [1]: import datasets\r\n ...: from datasets import load_dataset\r\n ...: dataset = load_dataset(\"conll2003\")\r\n\r\nIn [2]: dataset\r\nOut[2]: \r\nDatasetDict({\r\n train: Dataset({\r\n fea...
2022-04-28T09:33:31Z
2022-07-18T07:15:48Z
2022-07-18T07:15:48Z
NONE
null
null
null
## Describe the bug I load `conll2003` dataset to use refined data like [this](https://huggingface.co/datasets/conll2003/viewer/conll2003/train) preview, but it is original data that contains `'-DOCSTART- -X- -X- O'` text. Is this a bug or should I use another dataset_name like `lhoestq/conll2003` ? ## Steps to reproduce the bug ```python import datasets from datasets import load_dataset dataset = load_dataset("conll2003") ``` ## Expected results { "chunk_tags": [11, 12, 12, 21, 13, 11, 11, 21, 13, 11, 12, 13, 11, 21, 22, 11, 12, 17, 11, 21, 17, 11, 12, 12, 21, 22, 22, 13, 11, 0], "id": "0", "ner_tags": [0, 3, 4, 0, 0, 0, 0, 0, 0, 7, 0, 0, 0, 0, 0, 7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], "pos_tags": [12, 22, 22, 38, 15, 22, 28, 38, 15, 16, 21, 35, 24, 35, 37, 16, 21, 15, 24, 41, 15, 16, 21, 21, 20, 37, 40, 35, 21, 7], "tokens": ["The", "European", "Commission", "said", "on", "Thursday", "it", "disagreed", "with", "German", "advice", "to", "consumers", "to", "shun", "British", "lamb", "until", "scientists", "determine", "whether", "mad", "cow", "disease", "can", "be", "transmitted", "to", "sheep", "."] } ## Actual results ```python print(dataset) DatasetDict({ train: Dataset({ features: ['text'], num_rows: 219554 }) test: Dataset({ features: ['text'], num_rows: 50350 }) validation: Dataset({ features: ['text'], num_rows: 55044 }) }) ``` ```python for i in range(20): print(dataset['train'][i]) {'text': '-DOCSTART- -X- -X- O'} {'text': ''} {'text': 'EU NNP B-NP B-ORG'} {'text': 'rejects VBZ B-VP O'} {'text': 'German JJ B-NP B-MISC'} {'text': 'call NN I-NP O'} {'text': 'to TO B-VP O'} {'text': 'boycott VB I-VP O'} {'text': 'British JJ B-NP B-MISC'} {'text': 'lamb NN I-NP O'} {'text': '. . O O'} {'text': ''} {'text': 'Peter NNP B-NP B-PER'} {'text': 'Blackburn NNP I-NP I-PER'} {'text': ''} {'text': 'BRUSSELS NNP B-NP B-LOC'} {'text': '1996-08-22 CD I-NP O'} {'text': ''} {'text': 'The DT B-NP O'} {'text': 'European NNP I-NP B-ORG'} ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4248/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4248/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/2038
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2038/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2038/comments
https://api.github.com/repos/huggingface/datasets/issues/2038/events
https://github.com/huggingface/datasets/issues/2038
830,036,875
MDU6SXNzdWU4MzAwMzY4NzU=
2,038
outdated dataset_infos.json might fail verifications
{ "avatar_url": "https://avatars.githubusercontent.com/u/2062185?v=4", "events_url": "https://api.github.com/users/songfeng/events{/privacy}", "followers_url": "https://api.github.com/users/songfeng/followers", "following_url": "https://api.github.com/users/songfeng/following{/other_user}", "gists_url": "https://api.github.com/users/songfeng/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/songfeng", "id": 2062185, "login": "songfeng", "node_id": "MDQ6VXNlcjIwNjIxODU=", "organizations_url": "https://api.github.com/users/songfeng/orgs", "received_events_url": "https://api.github.com/users/songfeng/received_events", "repos_url": "https://api.github.com/users/songfeng/repos", "site_admin": false, "starred_url": "https://api.github.com/users/songfeng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/songfeng/subscriptions", "type": "User", "url": "https://api.github.com/users/songfeng" }
[]
closed
false
null
[]
null
[ "Hi ! Thanks for reporting.\r\n\r\nTo update the dataset_infos.json you can run:\r\n```\r\ndatasets-cli test ./datasets/doc2dial --all_configs --save_infos --ignore_verifications\r\n```", "Fixed by #2041, thanks again @songfeng !" ]
2021-03-12T11:41:54Z
2021-03-16T16:27:40Z
2021-03-16T16:27:40Z
CONTRIBUTOR
null
null
null
The [doc2dial/dataset_infos.json](https://github.com/huggingface/datasets/blob/master/datasets/doc2dial/dataset_infos.json) is outdated. It would fail data_loader when verifying download checksum etc.. Could you please update this file or point me how to update this file? Thank you.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2038/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2038/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/4622
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4622/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4622/comments
https://api.github.com/repos/huggingface/datasets/issues/4622/events
https://github.com/huggingface/datasets/pull/4622
1,293,031,939
PR_kwDODunzps46ynmT
4,622
Fix ImageFolder with parameters drop_metadata=True and drop_labels=False (when metadata.jsonl is present)
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "@lhoestq @mariosasko pls take a look at https://github.com/huggingface/datasets/pull/4622/commits/769e4c046a5bd5e3a4dbd09cfad1f4cf60677869. I modified `_generate_examples()` according to the same logic too: removed checking if `metad...
2022-07-04T11:23:20Z
2022-07-15T14:37:23Z
2022-07-15T14:24:24Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4622.diff", "html_url": "https://github.com/huggingface/datasets/pull/4622", "merged_at": "2022-07-15T14:24:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/4622.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4622" }
Will fix #4621 ImageFolder raises `KeyError: 'label'` with params `drop_metadata=True` and `drop_labels=False` (if there is at least one metadata.jsonl file a data directory). This happens because metadata files are collected inside `analyze()` function regardless of `drop_metadata` value. And then the following condition doesn't pass: https://github.com/huggingface/datasets/blob/master/src/datasets/packaged_modules/imagefolder/imagefolder.py#L167 So I suggest to double check it inside `analyze()` not to collect metadata files if they are not needed. (and labels too, to be consistent) --- Also, I added a test to check if labels are inferred correctly from directories names in general (because we didn't have it) :)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4622/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4622/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/1305
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1305/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1305/comments
https://api.github.com/repos/huggingface/datasets/issues/1305/events
https://github.com/huggingface/datasets/pull/1305
759,446,665
MDExOlB1bGxSZXF1ZXN0NTM0NDUxNzEx
1,305
[README] Added Windows command to enable slow tests
{ "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "events_url": "https://api.github.com/users/TevenLeScao/events{/privacy}", "followers_url": "https://api.github.com/users/TevenLeScao/followers", "following_url": "https://api.github.com/users/TevenLeScao/following{/other_user}", "gists_url": "https://api.github.com/users/TevenLeScao/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/TevenLeScao", "id": 26709476, "login": "TevenLeScao", "node_id": "MDQ6VXNlcjI2NzA5NDc2", "organizations_url": "https://api.github.com/users/TevenLeScao/orgs", "received_events_url": "https://api.github.com/users/TevenLeScao/received_events", "repos_url": "https://api.github.com/users/TevenLeScao/repos", "site_admin": false, "starred_url": "https://api.github.com/users/TevenLeScao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TevenLeScao/subscriptions", "type": "User", "url": "https://api.github.com/users/TevenLeScao" }
[]
closed
false
null
[]
null
[]
2020-12-08T13:29:04Z
2020-12-08T13:56:33Z
2020-12-08T13:56:32Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1305.diff", "html_url": "https://github.com/huggingface/datasets/pull/1305", "merged_at": "2020-12-08T13:56:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/1305.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1305" }
The Windows command to run slow tests has caused issues, so this adds a functional Windows command.
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/1305/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1305/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4488
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4488/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4488/comments
https://api.github.com/repos/huggingface/datasets/issues/4488/events
https://github.com/huggingface/datasets/pull/4488
1,270,613,857
PR_kwDODunzps45n6Ja
4,488
Update PASS dataset version
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-06-14T10:47:14Z
2022-06-14T16:41:55Z
2022-06-14T16:32:28Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4488.diff", "html_url": "https://github.com/huggingface/datasets/pull/4488", "merged_at": "2022-06-14T16:32:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/4488.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4488" }
Update the PASS dataset to version v3 (the newest one) from the [version history](https://github.com/yukimasano/PASS/blob/main/version_history.txt). PS: The older versions are not exposed as configs in the script because v1 was removed from Zenodo, and the same thing will probably happen to v2.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4488/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4488/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4925
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4925/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4925/comments
https://api.github.com/repos/huggingface/datasets/issues/4925/events
https://github.com/huggingface/datasets/pull/4925
1,360,007,616
PR_kwDODunzps4-RbP5
4,925
Add note about loading image / audio files to docs
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4925). All of your documentation changes will be reflected on that endpoint.", "Thanks for the feedback @polinaeterna ! I've reworded the docs a bit to integrate your comments and this should be ready for another review :)", ...
2022-09-02T10:31:58Z
2022-09-26T12:21:30Z
2022-09-23T13:59:07Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4925.diff", "html_url": "https://github.com/huggingface/datasets/pull/4925", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4925.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4925" }
This PR adds a small note about how to load image / audio datasets that have multiple splits in their dataset structure. Related forum thread: https://discuss.huggingface.co/t/loading-train-and-test-splits-with-audiofolder/22447 cc @NielsRogge
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4925/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4925/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/1220
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1220/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1220/comments
https://api.github.com/repos/huggingface/datasets/issues/1220/events
https://github.com/huggingface/datasets/pull/1220
758,015,894
MDExOlB1bGxSZXF1ZXN0NTMzMjYxNTgw
1,220
add Korean HateSpeech dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
closed
false
null
[]
null
[ "It looks like you forgot to `make style` (I forget it a lot too 🤦 )\r\n+ add dummy data", "hi @cceyda 👋, thanks for the hint! it looks like i've run into some other errors though in `_split_generators` or `_generate_examples`. do you have any idea of what's wrong here? 😅", "I get the same errors on another...
2020-12-06T20:31:29Z
2020-12-08T15:21:09Z
2020-12-08T11:05:42Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1220.diff", "html_url": "https://github.com/huggingface/datasets/pull/1220", "merged_at": "2020-12-08T11:05:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/1220.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1220" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1220/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1220/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2866
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2866/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2866/comments
https://api.github.com/repos/huggingface/datasets/issues/2866/events
https://github.com/huggingface/datasets/issues/2866
986,706,676
MDU6SXNzdWU5ODY3MDY2NzY=
2,866
"counter" dataset raises an error in normal mode, but not in streaming mode
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
[ "Hi @severo, thanks for reporting.\r\n\r\nJust note that currently not all canonical datasets support streaming mode: this is one case!\r\n\r\nAll datasets that use `pathlib` joins (using `/`) instead of `os.path.join` (as in this dataset) do not support streaming mode yet.", "OK. Do you think it's possible to de...
2021-09-02T13:10:53Z
2021-10-14T09:24:09Z
2021-10-14T09:24:09Z
CONTRIBUTOR
null
null
null
## Describe the bug `counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode. ## Steps to reproduce the bug ```python >>> import datasets as ds >>> a = ds.load_dataset('counter', split="train", streaming=False) Using custom data configuration default Downloading and preparing dataset counter/default (download: 1.29 MiB, generated: 2.48 MiB, post-processed: Unknown size, total: 3.77 MiB) to /home/slesage/.cache/huggingface/datasets/counter/default/1.0.0/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9... Traceback (most recent call last): File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.8/site-packages/datasets/builder.py", line 726, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.8/site-packages/datasets/builder.py", line 1124, in _prepare_split for key, record in utils.tqdm( File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.8/site-packages/tqdm/std.py", line 1185, in __iter__ for obj in iterable: File "/home/slesage/.cache/huggingface/modules/datasets_modules/datasets/counter/9f84962fa0f35bec5a34fe0bdff8681838d497008c457f7856c48654476ec0e9/counter.py", line 161, in _generate_examples with derived_file.open(encoding="utf-8") as f: File "/home/slesage/.pyenv/versions/3.8.11/lib/python3.8/pathlib.py", line 1222, in open return io.open(self, mode, buffering, encoding, errors, newline, File "/home/slesage/.pyenv/versions/3.8.11/lib/python3.8/pathlib.py", line 1078, in _opener return self._accessor.open(self, flags, mode) FileNotFoundError: [Errno 2] No such file or directory: '/home/slesage/.cache/huggingface/datasets/downloads/extracted/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211/COUNTER/0032p.xml' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.8/site-packages/datasets/load.py", line 1112, in load_dataset builder_instance.download_and_prepare( File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.8/site-packages/datasets/builder.py", line 636, in download_and_prepare self._download_and_prepare( File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.8/site-packages/datasets/builder.py", line 728, in _download_and_prepare raise OSError( OSError: Cannot find data file. Original error: [Errno 2] No such file or directory: '/home/slesage/.cache/huggingface/datasets/downloads/extracted/b57aa6db5601a738e57b95c1fd8cced54ff28fc540efcdaf0f6c4f1bb5dfe211/COUNTER/0032p.xml' ``` ```python >>> import datasets as ds >>> b = ds.load_dataset('counter', split="train", streaming=True) Using custom data configuration default >>> list(b) [] ``` ## Expected results An exception should be raised in streaming mode ## Actual results No exception is raised in streaming mode: there is no way to tell if something has broken or if the dataset is simply empty. ## Environment info - `datasets` version: 1.11.1.dev0 - Platform: Linux-5.11.0-1016-aws-x86_64-with-glibc2.29 - Python version: 3.8.11 - PyArrow version: 4.0.1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2866/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2866/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/5905
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5905/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5905/comments
https://api.github.com/repos/huggingface/datasets/issues/5905/events
https://github.com/huggingface/datasets/issues/5905
1,727,541,392
I_kwDODunzps5m-DCQ
5,905
Offer an alternative to Iterable Dataset that allows lazy loading and processing while skipping batches efficiently
{ "avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4", "events_url": "https://api.github.com/users/Hubert-Bonisseur/events{/privacy}", "followers_url": "https://api.github.com/users/Hubert-Bonisseur/followers", "following_url": "https://api.github.com/users/Hubert-Bonisseur/following{/other_user}", "gists_url": "https://api.github.com/users/Hubert-Bonisseur/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Hubert-Bonisseur", "id": 48770768, "login": "Hubert-Bonisseur", "node_id": "MDQ6VXNlcjQ4NzcwNzY4", "organizations_url": "https://api.github.com/users/Hubert-Bonisseur/orgs", "received_events_url": "https://api.github.com/users/Hubert-Bonisseur/received_events", "repos_url": "https://api.github.com/users/Hubert-Bonisseur/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Hubert-Bonisseur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hubert-Bonisseur/subscriptions", "type": "User", "url": "https://api.github.com/users/Hubert-Bonisseur" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[ "We plan to improve this eventually (see https://github.com/huggingface/datasets/issues/5454 and https://github.com/huggingface/datasets/issues/5380).\r\n\r\n> Is it possible to lazily load samples of a mapped dataset ? I'm used to [dataset scripts](https://huggingface.co/docs/datasets/dataset_script), maybe someth...
2023-05-26T12:33:02Z
2023-06-15T13:34:18Z
null
CONTRIBUTOR
null
null
null
### Feature request I would like a way to resume training from a checkpoint without waiting for a very long time when using an iterable dataset. ### Motivation I am training models on the speech-recognition task. I have very large datasets that I can't comfortably store on a disk and also quite computationally intensive audio processing to do. As a result I want to load data from my remote when it is needed and perform all processing on the fly. I am currently using the iterable dataset feature of _datasets_. It does everything I need with one exception. My issue is that when resuming training at a step n, we have to download all the data and perform the processing of steps < n, just to get the iterable at the right step. In my case it takes almost as long as training for the same steps, which make resuming training from a checkpoint useless in practice. I understand that the nature of iterators make it probably nearly impossible to quickly resume training. I thought about a possible solution nonetheless : I could in fact index my large dataset and make it a mapped dataset. Then I could use set_transform to perform the processing on the fly. Finally, if I'm not mistaken, the _accelerate_ package allows to [skip steps efficiently](https://github.com/huggingface/accelerate/blob/a73898027a211c3f6dc4460351b0ec246aa824aa/src/accelerate/data_loader.py#L827) for a mapped dataset. Is it possible to lazily load samples of a mapped dataset ? I'm used to [dataset scripts](https://huggingface.co/docs/datasets/dataset_script), maybe something can be done there. If not, I could do it using a plain _Pytorch_ dataset. Then I would need to convert it to a _datasets_' dataset to get all the features of _datasets_. Is it something possible ? ### Your contribution I could provide a PR to allow lazy loading of mapped dataset or the conversion of a mapped _Pytorch_ dataset into a _Datasets_ dataset if you think it is an useful new feature.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5905/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5905/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2432
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2432/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2432/comments
https://api.github.com/repos/huggingface/datasets/issues/2432/events
https://github.com/huggingface/datasets/pull/2432
907,462,881
MDExOlB1bGxSZXF1ZXN0NjU4MzA3MTE1
2,432
Fix CI six installation on linux
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[]
2021-05-31T13:15:36Z
2021-05-31T13:17:07Z
2021-05-31T13:17:06Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/2432.diff", "html_url": "https://github.com/huggingface/datasets/pull/2432", "merged_at": "2021-05-31T13:17:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/2432.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2432" }
For some reason we end up with this error in the linux CI when running pip install .[tests] ``` pip._vendor.resolvelib.resolvers.InconsistentCandidate: Provided candidate AlreadyInstalledCandidate(six 1.16.0 (/usr/local/lib/python3.6/site-packages)) does not satisfy SpecifierRequirement('six>1.9'), SpecifierRequirement('six>1.9'), SpecifierRequirement('six>=1.11'), SpecifierRequirement('six~=1.15'), SpecifierRequirement('six'), SpecifierRequirement('six>=1.5.2'), SpecifierRequirement('six>=1.9.0'), SpecifierRequirement('six>=1.11.0'), SpecifierRequirement('six'), SpecifierRequirement('six>=1.6.1'), SpecifierRequirement('six>=1.9'), SpecifierRequirement('six>=1.5'), SpecifierRequirement('six<2.0'), SpecifierRequirement('six<2.0'), SpecifierRequirement('six'), SpecifierRequirement('six'), SpecifierRequirement('six~=1.15.0'), SpecifierRequirement('six'), SpecifierRequirement('six<2.0,>=1.6.1'), SpecifierRequirement('six'), SpecifierRequirement('six>=1.5.2'), SpecifierRequirement('six>=1.9.0') ``` example CI failure here: https://app.circleci.com/pipelines/github/huggingface/datasets/6200/workflows/b64fdec9-f9e6-431c-acd7-e9f2c440c568/jobs/38247 The main version requirement comes from tensorflow: `six~=1.15.0` So I pinned the six version to this.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2432/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2432/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3976
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3976/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3976/comments
https://api.github.com/repos/huggingface/datasets/issues/3976/events
https://github.com/huggingface/datasets/pull/3976
1,175,043,780
PR_kwDODunzps40uOY6
3,976
Fix main classes reference in docs
{ "avatar_url": "https://avatars.githubusercontent.com/u/24835382?v=4", "events_url": "https://api.github.com/users/qqaatw/events{/privacy}", "followers_url": "https://api.github.com/users/qqaatw/followers", "following_url": "https://api.github.com/users/qqaatw/following{/other_user}", "gists_url": "https://api.github.com/users/qqaatw/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/qqaatw", "id": 24835382, "login": "qqaatw", "node_id": "MDQ6VXNlcjI0ODM1Mzgy", "organizations_url": "https://api.github.com/users/qqaatw/orgs", "received_events_url": "https://api.github.com/users/qqaatw/received_events", "repos_url": "https://api.github.com/users/qqaatw/repos", "site_admin": false, "starred_url": "https://api.github.com/users/qqaatw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/qqaatw/subscriptions", "type": "User", "url": "https://api.github.com/users/qqaatw" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_3976). All of your documentation changes will be reflected on that endpoint.", "Not sure why some section titles end with `[[datasets.xxx]]`, like this: https://moon-ci-docs.huggingface.co/docs/datasets/pr_3976/en/package_refer...
2022-03-21T08:19:46Z
2022-04-12T14:19:39Z
2022-04-12T14:19:38Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3976.diff", "html_url": "https://github.com/huggingface/datasets/pull/3976", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/3976.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3976" }
Currently the section index (on the page's right side) of the [main classes reference](https://huggingface.co/docs/datasets/master/en/package_reference/main_classes) incorrectly displays `Tensor returned:`, this PR fixes this issue by wrapping code examples in this page with markdown code block. There are other examples in datasets library having this issue.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3976/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3976/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/1552
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1552/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1552/comments
https://api.github.com/repos/huggingface/datasets/issues/1552/events
https://github.com/huggingface/datasets/pull/1552
765,664,411
MDExOlB1bGxSZXF1ZXN0NTM5MDI2MzAx
1,552
Added OPUS ParaCrawl
{ "avatar_url": "https://avatars.githubusercontent.com/u/22396042?v=4", "events_url": "https://api.github.com/users/rkc007/events{/privacy}", "followers_url": "https://api.github.com/users/rkc007/followers", "following_url": "https://api.github.com/users/rkc007/following{/other_user}", "gists_url": "https://api.github.com/users/rkc007/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/rkc007", "id": 22396042, "login": "rkc007", "node_id": "MDQ6VXNlcjIyMzk2MDQy", "organizations_url": "https://api.github.com/users/rkc007/orgs", "received_events_url": "https://api.github.com/users/rkc007/received_events", "repos_url": "https://api.github.com/users/rkc007/repos", "site_admin": false, "starred_url": "https://api.github.com/users/rkc007/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rkc007/subscriptions", "type": "User", "url": "https://api.github.com/users/rkc007" }
[]
closed
false
null
[]
null
[ "@lhoestq I saw some common changes you made on the other PR's (Similar Opus Datasets). I fixed those changes here. Can you please review it once ? \r\nThanks.", "@rkc007 @lhoestq just noticed a dataset named para_crawl has been added a long time ago: #91.", "They're not exactly the same so it's ok to have both...
2020-12-13T21:44:29Z
2020-12-21T09:50:26Z
2020-12-21T09:50:25Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1552.diff", "html_url": "https://github.com/huggingface/datasets/pull/1552", "merged_at": "2020-12-21T09:50:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/1552.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1552" }
Dataset : http://opus.nlpl.eu/ParaCrawl.php
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1552/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1552/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6239
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6239/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6239/comments
https://api.github.com/repos/huggingface/datasets/issues/6239/events
https://github.com/huggingface/datasets/issues/6239
1,895,349,382
I_kwDODunzps5w-LyG
6,239
Load local audio data doesn't work
{ "avatar_url": "https://avatars.githubusercontent.com/u/554032?v=4", "events_url": "https://api.github.com/users/abodacs/events{/privacy}", "followers_url": "https://api.github.com/users/abodacs/followers", "following_url": "https://api.github.com/users/abodacs/following{/other_user}", "gists_url": "https://api.github.com/users/abodacs/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/abodacs", "id": 554032, "login": "abodacs", "node_id": "MDQ6VXNlcjU1NDAzMg==", "organizations_url": "https://api.github.com/users/abodacs/orgs", "received_events_url": "https://api.github.com/users/abodacs/received_events", "repos_url": "https://api.github.com/users/abodacs/repos", "site_admin": false, "starred_url": "https://api.github.com/users/abodacs/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/abodacs/subscriptions", "type": "User", "url": "https://api.github.com/users/abodacs" }
[]
closed
false
null
[]
null
[ "I think this is the same issue as https://github.com/huggingface/datasets/issues/4776. Maybe installing `ffmpeg` can fix it:\r\n```python\r\nadd-apt-repository -y ppa:savoury1/ffmpeg4\r\napt-get -qq install -y ffmpeg\r\n```\r\n\r\nHowever, the best solution is to use a newer version of `datasets`. In the recent re...
2023-09-13T22:30:01Z
2023-09-15T14:32:10Z
2023-09-15T14:32:10Z
NONE
null
null
null
### Describe the bug I get a RuntimeError from the following code: ```python audio_dataset = Dataset.from_dict({"audio": ["/kaggle/input/bengaliai-speech/train_mp3s/000005f3362c.mp3"]}).cast_column("audio", Audio()) audio_dataset[0] ``` ### Traceback <details> ```python RuntimeError Traceback (most recent call last) Cell In[33], line 1 ----> 1 train_dataset[0] File /opt/conda/lib/python3.10/site-packages/datasets/arrow_dataset.py:1764, in Dataset.__getitem__(self, key) 1762 def __getitem__(self, key): # noqa: F811 1763 """Can be used to index columns (by string names) or rows (by integer index or iterable of indices or bools).""" -> 1764 return self._getitem( 1765 key, 1766 ) File /opt/conda/lib/python3.10/site-packages/datasets/arrow_dataset.py:1749, in Dataset._getitem(self, key, decoded, **kwargs) 1747 formatter = get_formatter(format_type, features=self.features, decoded=decoded, **format_kwargs) 1748 pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None) -> 1749 formatted_output = format_table( 1750 pa_subtable, key, formatter=formatter, format_columns=format_columns, output_all_columns=output_all_columns 1751 ) 1752 return formatted_output File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:532, in format_table(table, key, formatter, format_columns, output_all_columns) 530 python_formatter = PythonFormatter(features=None) 531 if format_columns is None: --> 532 return formatter(pa_table, query_type=query_type) 533 elif query_type == "column": 534 if key in format_columns: File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:281, in Formatter.__call__(self, pa_table, query_type) 279 def __call__(self, pa_table: pa.Table, query_type: str) -> Union[RowFormat, ColumnFormat, BatchFormat]: 280 if query_type == "row": --> 281 return self.format_row(pa_table) 282 elif query_type == "column": 283 return self.format_column(pa_table) File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:312, in PythonFormatter.format_row(self, pa_table) 310 row = self.python_arrow_extractor().extract_row(pa_table) 311 if self.decoded: --> 312 row = self.python_features_decoder.decode_row(row) 313 return row File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:221, in PythonFeaturesDecoder.decode_row(self, row) 220 def decode_row(self, row: dict) -> dict: --> 221 return self.features.decode_example(row) if self.features else row File /opt/conda/lib/python3.10/site-packages/datasets/features/features.py:1386, in Features.decode_example(self, example) 1376 def decode_example(self, example: dict): 1377 """Decode example with custom feature decoding. 1378 1379 Args: (...) 1383 :obj:`dict[str, Any]` 1384 """ -> 1386 return { 1387 column_name: decode_nested_example(feature, value) 1388 if self._column_requires_decoding[column_name] 1389 else value 1390 for column_name, (feature, value) in zip_dict( 1391 {key: value for key, value in self.items() if key in example}, example 1392 ) 1393 } File /opt/conda/lib/python3.10/site-packages/datasets/features/features.py:1387, in <dictcomp>(.0) 1376 def decode_example(self, example: dict): 1377 """Decode example with custom feature decoding. 1378 1379 Args: (...) 1383 :obj:`dict[str, Any]` 1384 """ 1386 return { -> 1387 column_name: decode_nested_example(feature, value) 1388 if self._column_requires_decoding[column_name] 1389 else value 1390 for column_name, (feature, value) in zip_dict( 1391 {key: value for key, value in self.items() if key in example}, example 1392 ) 1393 } File /opt/conda/lib/python3.10/site-packages/datasets/features/features.py:1087, in decode_nested_example(schema, obj) 1085 # Object with special decoding: 1086 elif isinstance(schema, (Audio, Image)): -> 1087 return schema.decode_example(obj) if obj is not None else None 1088 return obj File /opt/conda/lib/python3.10/site-packages/datasets/features/audio.py:103, in Audio.decode_example(self, value) 101 raise ValueError(f"An audio sample should have one of 'path' or 'bytes' but both are None in {value}.") 102 elif path is not None and path.endswith("mp3"): --> 103 array, sampling_rate = self._decode_mp3(file if file else path) 104 elif path is not None and path.endswith("opus"): 105 if file: File /opt/conda/lib/python3.10/site-packages/datasets/features/audio.py:241, in Audio._decode_mp3(self, path_or_file) 238 except RuntimeError as err: 239 raise ImportError("To support decoding 'mp3' audio files, please install 'sox'.") from err --> 241 array, sampling_rate = torchaudio.load(path_or_file, format="mp3") 242 if self.sampling_rate and self.sampling_rate != sampling_rate: 243 if not hasattr(self, "_resampler") or self._resampler.orig_freq != sampling_rate: File /opt/conda/lib/python3.10/site-packages/torchaudio/backend/sox_io_backend.py:256, in load(filepath, frame_offset, num_frames, normalize, channels_first, format) 254 if ret is not None: 255 return ret --> 256 return _fallback_load(filepath, frame_offset, num_frames, normalize, channels_first, format) File /opt/conda/lib/python3.10/site-packages/torchaudio/backend/sox_io_backend.py:30, in _fail_load(filepath, frame_offset, num_frames, normalize, channels_first, format) 22 def _fail_load( 23 filepath: str, 24 frame_offset: int = 0, (...) 28 format: Optional[str] = None, 29 ) -> Tuple[torch.Tensor, int]: ---> 30 raise RuntimeError("Failed to load audio from {}".format(filepath)) RuntimeError: Failed to load audio from /kaggle/input/bengaliai-speech/train_mp3s/000005f3362c.mp3 ``` </details> ### Steps to reproduce the bug 1. - Create a custom dataset using Local files of type mp3. 3. - Try to read the first audio item. ### Expected behavior Expected output ```python audio_dataset[0]["audio"] {'array': array([ 0. , 0.00024414, -0.00024414, ..., -0.00024414, 0. , 0. ], dtype=float32), 'path': 'path/to/audio_1', 'sampling_rate': 16000} ``` ### Environment info N/A
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6239/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6239/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/3233
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3233/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3233/comments
https://api.github.com/repos/huggingface/datasets/issues/3233/events
https://github.com/huggingface/datasets/pull/3233
1,047,474,931
PR_kwDODunzps4uOl9-
3,233
Improve repository structure docs
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[]
2021-11-08T13:51:35Z
2021-11-09T10:02:18Z
2021-11-09T10:02:17Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3233.diff", "html_url": "https://github.com/huggingface/datasets/pull/3233", "merged_at": "2021-11-09T10:02:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/3233.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3233" }
Continuation of the documentation started in https://github.com/huggingface/datasets/pull/3221, taking into account @stevhliu 's comments
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3233/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3233/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2104
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2104/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2104/comments
https://api.github.com/repos/huggingface/datasets/issues/2104/events
https://github.com/huggingface/datasets/issues/2104
839,027,834
MDU6SXNzdWU4MzkwMjc4MzQ=
2,104
Trouble loading wiki_movies
{ "avatar_url": "https://avatars.githubusercontent.com/u/35391599?v=4", "events_url": "https://api.github.com/users/adityaarunsinghal/events{/privacy}", "followers_url": "https://api.github.com/users/adityaarunsinghal/followers", "following_url": "https://api.github.com/users/adityaarunsinghal/following{/other_user}", "gists_url": "https://api.github.com/users/adityaarunsinghal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/adityaarunsinghal", "id": 35391599, "login": "adityaarunsinghal", "node_id": "MDQ6VXNlcjM1MzkxNTk5", "organizations_url": "https://api.github.com/users/adityaarunsinghal/orgs", "received_events_url": "https://api.github.com/users/adityaarunsinghal/received_events", "repos_url": "https://api.github.com/users/adityaarunsinghal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/adityaarunsinghal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/adityaarunsinghal/subscriptions", "type": "User", "url": "https://api.github.com/users/adityaarunsinghal" }
[]
closed
false
null
[]
null
[ "Hi ! `wiki_movies` was added in `datasets==1.2.0`. However it looks like you have `datasets==1.1.2`.\r\n\r\nTo use `wiki_movies`, please update `datasets` with\r\n```\r\npip install --upgrade datasets\r\n```", "Thanks a lot! That solved it and I was able to upload a model trained on it as well :)" ]
2021-03-23T18:59:54Z
2022-03-30T08:22:58Z
2022-03-30T08:22:58Z
NONE
null
null
null
Hello, I am trying to load_dataset("wiki_movies") and it gives me this error - `FileNotFoundError: Couldn't find file locally at wiki_movies/wiki_movies.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.2/datasets/wiki_movies/wiki_movies.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/datasets/wiki_movies/wiki_movies.py` Trying to do `python run_mlm.py \ --model_name_or_path roberta-base \ --dataset_name wiki_movies \` also gives the same error. Is this something on my end? From what I can tell, this dataset was re-added by @lhoestq a few months ago. Thank you!
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2104/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2104/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/1538
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1538/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1538/comments
https://api.github.com/repos/huggingface/datasets/issues/1538/events
https://github.com/huggingface/datasets/pull/1538
765,139,739
MDExOlB1bGxSZXF1ZXN0NTM4ODkxOTE3
1,538
tweets_hate_speech_detection
{ "avatar_url": "https://avatars.githubusercontent.com/u/44197177?v=4", "events_url": "https://api.github.com/users/darshan-gandhi/events{/privacy}", "followers_url": "https://api.github.com/users/darshan-gandhi/followers", "following_url": "https://api.github.com/users/darshan-gandhi/following{/other_user}", "gists_url": "https://api.github.com/users/darshan-gandhi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/darshan-gandhi", "id": 44197177, "login": "darshan-gandhi", "node_id": "MDQ6VXNlcjQ0MTk3MTc3", "organizations_url": "https://api.github.com/users/darshan-gandhi/orgs", "received_events_url": "https://api.github.com/users/darshan-gandhi/received_events", "repos_url": "https://api.github.com/users/darshan-gandhi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/darshan-gandhi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/darshan-gandhi/subscriptions", "type": "User", "url": "https://api.github.com/users/darshan-gandhi" }
[]
closed
false
null
[]
null
[ "Hi @lhoestq I have added this new dataset for tweet's hate speech detection. \r\n\r\nPlease if u could review it. \r\n\r\nThank you", "Hi @darshan-gandhi have you add a chance to take a look at my suggestions ?\r\n\r\nFeel free to ping me when you're ready for the final review", "Closing in favor of #1607" ]
2020-12-13T07:37:53Z
2020-12-21T15:54:28Z
2020-12-21T15:54:27Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1538.diff", "html_url": "https://github.com/huggingface/datasets/pull/1538", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1538.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1538" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1538/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1538/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3284
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3284/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3284/comments
https://api.github.com/repos/huggingface/datasets/issues/3284/events
https://github.com/huggingface/datasets/issues/3284
1,055,502,909
I_kwDODunzps4-6bI9
3,284
Add VoxLingua107 dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/7246357?v=4", "events_url": "https://api.github.com/users/osanseviero/events{/privacy}", "followers_url": "https://api.github.com/users/osanseviero/followers", "following_url": "https://api.github.com/users/osanseviero/following{/other_user}", "gists_url": "https://api.github.com/users/osanseviero/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/osanseviero", "id": 7246357, "login": "osanseviero", "node_id": "MDQ6VXNlcjcyNDYzNTc=", "organizations_url": "https://api.github.com/users/osanseviero/orgs", "received_events_url": "https://api.github.com/users/osanseviero/received_events", "repos_url": "https://api.github.com/users/osanseviero/repos", "site_admin": false, "starred_url": "https://api.github.com/users/osanseviero/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/osanseviero/subscriptions", "type": "User", "url": "https://api.github.com/users/osanseviero" }
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" }, { "color": "d93f0b",...
open
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_use...
null
[ "#self-assign" ]
2021-11-16T22:44:08Z
2021-12-06T09:49:45Z
null
MEMBER
null
null
null
## Adding a Dataset - **Name:** VoxLingua107 - **Description:** VoxLingua107 is a speech dataset for training spoken language identification models. The dataset consists of short speech segments automatically extracted from YouTube videos and labeled according the language of the video title and description, with some post-processing steps to filter out false positives. - **Paper:** https://arxiv.org/abs/2011.12998 - **Data:** http://bark.phon.ioc.ee/voxlingua107/ - **Motivation:** Nice audio classification dataset cc @anton-l Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3284/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3284/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6284
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6284/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6284/comments
https://api.github.com/repos/huggingface/datasets/issues/6284/events
https://github.com/huggingface/datasets/issues/6284
1,929,551,712
I_kwDODunzps5zAp9g
6,284
Add Belebele multiple-choice machine reading comprehension (MRC) dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/64583161?v=4", "events_url": "https://api.github.com/users/rajveer43/events{/privacy}", "followers_url": "https://api.github.com/users/rajveer43/followers", "following_url": "https://api.github.com/users/rajveer43/following{/other_user}", "gists_url": "https://api.github.com/users/rajveer43/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/rajveer43", "id": 64583161, "login": "rajveer43", "node_id": "MDQ6VXNlcjY0NTgzMTYx", "organizations_url": "https://api.github.com/users/rajveer43/orgs", "received_events_url": "https://api.github.com/users/rajveer43/received_events", "repos_url": "https://api.github.com/users/rajveer43/repos", "site_admin": false, "starred_url": "https://api.github.com/users/rajveer43/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rajveer43/subscriptions", "type": "User", "url": "https://api.github.com/users/rajveer43" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[ "This dataset is already available on the Hub: https://huggingface.co/datasets/facebook/belebele.\r\n" ]
2023-10-06T06:58:03Z
2023-10-06T13:26:51Z
2023-10-06T13:26:51Z
NONE
null
null
null
### Feature request Belebele is a multiple-choice machine reading comprehension (MRC) dataset spanning 122 language variants. This dataset enables the evaluation of mono- and multi-lingual models in high-, medium-, and low-resource languages. Each question has four multiple-choice answers and is linked to a short passage from the [FLORES-200](https://github.com/facebookresearch/flores/tree/main/flores200) dataset. The human annotation procedure was carefully curated to create questions that discriminate between different levels of generalizable language comprehension and is reinforced by extensive quality checks. While all questions directly relate to the passage, the English dataset on its own proves difficult enough to challenge state-of-the-art language models. Being fully parallel, this dataset enables direct comparison of model performance across all languages. Belebele opens up new avenues for evaluating and analyzing the multilingual abilities of language models and NLP systems. Please refer to paper for more details, [The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language Variants](https://arxiv.org/abs/2308.16884). ## Composition - 900 questions per language variant - 488 distinct passages, there are 1-2 associated questions for each. - For each question, there is 4 multiple-choice answers, exactly 1 of which is correct. - 122 language/language variants (including English). - 900 x 122 = 109,800 total questions. ### Motivation official repo https://github.com/facebookresearch/belebele ### Your contribution -
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/6284/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6284/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/4180
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4180/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4180/comments
https://api.github.com/repos/huggingface/datasets/issues/4180/events
https://github.com/huggingface/datasets/issues/4180
1,208,042,320
I_kwDODunzps5IAUNQ
4,180
Add some iteration method on a dataset column (specific for inference)
{ "avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4", "events_url": "https://api.github.com/users/Narsil/events{/privacy}", "followers_url": "https://api.github.com/users/Narsil/followers", "following_url": "https://api.github.com/users/Narsil/following{/other_user}", "gists_url": "https://api.github.com/users/Narsil/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Narsil", "id": 204321, "login": "Narsil", "node_id": "MDQ6VXNlcjIwNDMyMQ==", "organizations_url": "https://api.github.com/users/Narsil/orgs", "received_events_url": "https://api.github.com/users/Narsil/received_events", "repos_url": "https://api.github.com/users/Narsil/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Narsil/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Narsil/subscriptions", "type": "User", "url": "https://api.github.com/users/Narsil" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[ "Thanks for the suggestion ! I agree it would be nice to have something directly in `datasets` to do something as simple as that\r\n\r\ncc @albertvillanova @mariosasko @polinaeterna What do you think if we have something similar to pandas `Series` that wouldn't bring everything in memory when doing `dataset[\"audio...
2022-04-19T09:15:45Z
2022-04-21T10:30:58Z
null
CONTRIBUTOR
null
null
null
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Currently, `dataset["audio"]` will load EVERY element in the dataset in RAM, which can be quite big for an audio dataset. Having an iterator (or sequence) type of object, would make inference with `transformers` 's `pipeline` easier to use and not so memory hungry. **Describe the solution you'd like** A clear and concise description of what you want to happen. For a non breaking change: ```python for audio in dataset.iterate("audio"): # {"array": np.array(...), "sampling_rate":...} ``` For a breaking change solution (not necessary), changing the type of `dataset["audio"]` to a sequence type so that ```python pipe = pipeline(model="...") for out in pipe(dataset["audio"]): # {"text":....} ``` could work **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. ```python def iterate(dataset, key): for item in dataset: yield dataset[key] for out in pipeline(iterate(dataset, "audio")): # {"array": ...} ``` This works but requires the helper function which feels slightly clunky. **Additional context** Add any other context about the feature request here. The context is actually to showcase better integration between `pipeline` and `datasets` in the Quicktour demo: https://github.com/huggingface/transformers/pull/16723/files @lhoestq
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4180/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4180/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/3672
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3672/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3672/comments
https://api.github.com/repos/huggingface/datasets/issues/3672/events
https://github.com/huggingface/datasets/pull/3672
1,122,980,556
PR_kwDODunzps4yBUrZ
3,672
Prioritize `module.builder_kwargs` over defaults in `TestCommand`
{ "avatar_url": "https://avatars.githubusercontent.com/u/8264887?v=4", "events_url": "https://api.github.com/users/lvwerra/events{/privacy}", "followers_url": "https://api.github.com/users/lvwerra/followers", "following_url": "https://api.github.com/users/lvwerra/following{/other_user}", "gists_url": "https://api.github.com/users/lvwerra/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lvwerra", "id": 8264887, "login": "lvwerra", "node_id": "MDQ6VXNlcjgyNjQ4ODc=", "organizations_url": "https://api.github.com/users/lvwerra/orgs", "received_events_url": "https://api.github.com/users/lvwerra/received_events", "repos_url": "https://api.github.com/users/lvwerra/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lvwerra/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lvwerra/subscriptions", "type": "User", "url": "https://api.github.com/users/lvwerra" }
[]
closed
false
null
[]
null
[]
2022-02-03T11:38:42Z
2022-02-04T12:37:20Z
2022-02-04T12:37:19Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3672.diff", "html_url": "https://github.com/huggingface/datasets/pull/3672", "merged_at": "2022-02-04T12:37:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/3672.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3672" }
This fixes a bug in the `TestCommand` where multiple kwargs for `name` were passed if it was set in both default and `module.builder_kwargs`. Example error: ```Python Traceback (most recent call last): File "create_metadata.py", line 96, in <module> main(**vars(args)) File "create_metadata.py", line 86, in main metadata_command.run() File "/opt/conda/lib/python3.7/site-packages/datasets/commands/test.py", line 144, in run for j, builder in enumerate(get_builders()): File "/opt/conda/lib/python3.7/site-packages/datasets/commands/test.py", line 141, in get_builders name=name, cache_dir=self._cache_dir, data_dir=self._data_dir, **module.builder_kwargs TypeError: type object got multiple values for keyword argument 'name' ``` Let me know what you think.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3672/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3672/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/1638
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1638/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1638/comments
https://api.github.com/repos/huggingface/datasets/issues/1638/events
https://github.com/huggingface/datasets/pull/1638
774,869,184
MDExOlB1bGxSZXF1ZXN0NTQ1Njg5ODQ5
1,638
Add id_puisi dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/31740013?v=4", "events_url": "https://api.github.com/users/ilhamfp/events{/privacy}", "followers_url": "https://api.github.com/users/ilhamfp/followers", "following_url": "https://api.github.com/users/ilhamfp/following{/other_user}", "gists_url": "https://api.github.com/users/ilhamfp/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ilhamfp", "id": 31740013, "login": "ilhamfp", "node_id": "MDQ6VXNlcjMxNzQwMDEz", "organizations_url": "https://api.github.com/users/ilhamfp/orgs", "received_events_url": "https://api.github.com/users/ilhamfp/received_events", "repos_url": "https://api.github.com/users/ilhamfp/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ilhamfp/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ilhamfp/subscriptions", "type": "User", "url": "https://api.github.com/users/ilhamfp" }
[]
closed
false
null
[]
null
[]
2020-12-26T12:41:55Z
2020-12-30T16:34:17Z
2020-12-30T16:34:17Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1638.diff", "html_url": "https://github.com/huggingface/datasets/pull/1638", "merged_at": "2020-12-30T16:34:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/1638.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1638" }
Puisi (poem) is an Indonesian poetic form. The dataset contains 7223 Indonesian puisi with its title and author. :)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1638/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1638/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2328
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2328/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2328/comments
https://api.github.com/repos/huggingface/datasets/issues/2328/events
https://github.com/huggingface/datasets/pull/2328
877,673,896
MDExOlB1bGxSZXF1ZXN0NjMxNTg2MzU2
2,328
Add Matthews/Pearson/Spearman correlation metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[]
2021-05-06T16:09:27Z
2021-05-06T16:58:10Z
2021-05-06T16:58:10Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/2328.diff", "html_url": "https://github.com/huggingface/datasets/pull/2328", "merged_at": "2021-05-06T16:58:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/2328.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2328" }
Added three metrics: - The Matthews correlation coefficient (from sklearn) - The Pearson correlation coefficient (from scipy) - The Spearman correlation coefficient (from scipy) cc @sgugger
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2328/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2328/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3871
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3871/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3871/comments
https://api.github.com/repos/huggingface/datasets/issues/3871/events
https://github.com/huggingface/datasets/pull/3871
1,163,714,113
PR_kwDODunzps40KRcM
3,871
add pandas to env command
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_3871). All of your documentation changes will be reflected on that endpoint.", "Think failures are unrelated - feel free to merge whenever you want :-)" ]
2022-03-09T09:48:51Z
2022-03-09T11:21:38Z
2022-03-09T11:21:37Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3871.diff", "html_url": "https://github.com/huggingface/datasets/pull/3871", "merged_at": "2022-03-09T11:21:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/3871.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3871" }
Pandas is a required packages and used quite a bit. I don't see any downside with adding its version to the `datasets-cli env` command.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3871/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3871/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/1556
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1556/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1556/comments
https://api.github.com/repos/huggingface/datasets/issues/1556/events
https://github.com/huggingface/datasets/pull/1556
765,689,730
MDExOlB1bGxSZXF1ZXN0NTM5MDM2OTYz
1,556
add bswac
{ "avatar_url": "https://avatars.githubusercontent.com/u/11391118?v=4", "events_url": "https://api.github.com/users/IvanZidov/events{/privacy}", "followers_url": "https://api.github.com/users/IvanZidov/followers", "following_url": "https://api.github.com/users/IvanZidov/following{/other_user}", "gists_url": "https://api.github.com/users/IvanZidov/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/IvanZidov", "id": 11391118, "login": "IvanZidov", "node_id": "MDQ6VXNlcjExMzkxMTE4", "organizations_url": "https://api.github.com/users/IvanZidov/orgs", "received_events_url": "https://api.github.com/users/IvanZidov/received_events", "repos_url": "https://api.github.com/users/IvanZidov/repos", "site_admin": false, "starred_url": "https://api.github.com/users/IvanZidov/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/IvanZidov/subscriptions", "type": "User", "url": "https://api.github.com/users/IvanZidov" }
[]
closed
false
null
[]
null
[ "merging since the CI is fixed on master" ]
2020-12-13T22:55:35Z
2020-12-18T15:14:28Z
2020-12-18T15:14:27Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1556.diff", "html_url": "https://github.com/huggingface/datasets/pull/1556", "merged_at": "2020-12-18T15:14:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/1556.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1556" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1556/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1556/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2343
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2343/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2343/comments
https://api.github.com/repos/huggingface/datasets/issues/2343/events
https://github.com/huggingface/datasets/issues/2343
883,208,539
MDU6SXNzdWU4ODMyMDg1Mzk=
2,343
Columns are removed before or after map function applied?
{ "avatar_url": "https://avatars.githubusercontent.com/u/8199406?v=4", "events_url": "https://api.github.com/users/taghizad3h/events{/privacy}", "followers_url": "https://api.github.com/users/taghizad3h/followers", "following_url": "https://api.github.com/users/taghizad3h/following{/other_user}", "gists_url": "https://api.github.com/users/taghizad3h/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/taghizad3h", "id": 8199406, "login": "taghizad3h", "node_id": "MDQ6VXNlcjgxOTk0MDY=", "organizations_url": "https://api.github.com/users/taghizad3h/orgs", "received_events_url": "https://api.github.com/users/taghizad3h/received_events", "repos_url": "https://api.github.com/users/taghizad3h/repos", "site_admin": false, "starred_url": "https://api.github.com/users/taghizad3h/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/taghizad3h/subscriptions", "type": "User", "url": "https://api.github.com/users/taghizad3h" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
open
false
null
[]
null
[ "Hi! Columns are removed **after** applying the function and **before** updating the examples with the function's output (as per the docs [here](https://huggingface.co/docs/datasets/package_reference/main_classes#datasets.Dataset.map.remove_columns)). I agree the docs on this should be more clear." ]
2021-05-10T02:36:20Z
2022-10-24T11:31:55Z
null
NONE
null
null
null
## Describe the bug According to the documentation when applying map function the [remove_columns ](https://huggingface.co/docs/datasets/processing.html#removing-columns) will be removed after they are passed to the function, but in the [source code](https://huggingface.co/docs/datasets/package_reference/main_classes.html#datasets.Dataset.map) it's documented that they are removed before applying function. I thinks the source code doc is more accurate, right?
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2343/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2343/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/5354
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5354/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5354/comments
https://api.github.com/repos/huggingface/datasets/issues/5354/events
https://github.com/huggingface/datasets/issues/5354
1,492,174,125
I_kwDODunzps5Y8MUt
5,354
Consider using "Sequence" instead of "List"
{ "avatar_url": "https://avatars.githubusercontent.com/u/15568078?v=4", "events_url": "https://api.github.com/users/tranhd95/events{/privacy}", "followers_url": "https://api.github.com/users/tranhd95/followers", "following_url": "https://api.github.com/users/tranhd95/following{/other_user}", "gists_url": "https://api.github.com/users/tranhd95/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/tranhd95", "id": 15568078, "login": "tranhd95", "node_id": "MDQ6VXNlcjE1NTY4MDc4", "organizations_url": "https://api.github.com/users/tranhd95/orgs", "received_events_url": "https://api.github.com/users/tranhd95/received_events", "repos_url": "https://api.github.com/users/tranhd95/repos", "site_admin": false, "starred_url": "https://api.github.com/users/tranhd95/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tranhd95/subscriptions", "type": "User", "url": "https://api.github.com/users/tranhd95" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "7057ff", "default": true...
open
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/22453634?v=4", "events_url": "https://api.github.com/users/avinashsai/events{/privacy}", "followers_url": "https://api.github.com/users/avinashsai/followers", "following_url": "https://api.github.com/users/avinashsai/following{/other_user}", "gists_url": "https://api.github.com/users/avinashsai/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/avinashsai", "id": 22453634, "login": "avinashsai", "node_id": "MDQ6VXNlcjIyNDUzNjM0", "organizations_url": "https://api.github.com/users/avinashsai/orgs", "received_events_url": "https://api.github.com/users/avinashsai/received_events", "repos_url": "https://api.github.com/users/avinashsai/repos", "site_admin": false, "starred_url": "https://api.github.com/users/avinashsai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/avinashsai/subscriptions", "type": "User", "url": "https://api.github.com/users/avinashsai" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/22453634?v=4", "events_url": "https://api.github.com/users/avinashsai/events{/privacy}", "followers_url": "https://api.github.com/users/avinashsai/followers", "following_url": "https://api.github.com/users/avinashsai/following{/other_user}", ...
null
[ "Hi! Linking a comment to provide more info on the issue: https://stackoverflow.com/a/39458225. This means we should replace all (most of) the occurrences of `List` with `Sequence` in function signatures.\r\n\r\n@tranhd95 Would you be interested in submitting a PR?", "Hi all! I tried to reproduce this issue and d...
2022-12-12T15:39:45Z
2023-07-26T16:25:51Z
null
NONE
null
null
null
### Feature request Hi, please consider using `Sequence` type annotation instead of `List` in function arguments such as in [`Dataset.from_parquet()`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L1088). It leads to type checking errors, see below. **How to reproduce** ```py list_of_filenames = ["foo.parquet", "bar.parquet"] ds = Dataset.from_parquet(list_of_filenames) ``` **Expected mypy output:** ``` Success: no issues found ``` **Actual mypy output:** ```py test.py:19: error: Argument 1 to "from_parquet" of "Dataset" has incompatible type "List[str]"; expected "Union[Union[str, bytes, PathLike[Any]], List[Union[str, bytes, PathLike[Any]]]]" [arg-type] test.py:19: note: "List" is invariant -- see https://mypy.readthedocs.io/en/stable/common_issues.html#variance test.py:19: note: Consider using "Sequence" instead, which is covariant ``` **Env:** mypy 0.991, Python 3.10.0, datasets 2.7.1
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5354/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5354/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/3363
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3363/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3363/comments
https://api.github.com/repos/huggingface/datasets/issues/3363/events
https://github.com/huggingface/datasets/pull/3363
1,068,824,340
PR_kwDODunzps4vRVCl
3,363
Update URL of Jeopardy! dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[ "Closing this PR in favor of #3266.", "I think you should also close this branch" ]
2021-12-01T20:08:10Z
2022-10-06T13:45:49Z
2021-12-03T12:35:01Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3363.diff", "html_url": "https://github.com/huggingface/datasets/pull/3363", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/3363.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3363" }
Updates the URL of the Jeopardy! dataset. Fix #3361
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3363/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3363/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4635
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4635/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4635/comments
https://api.github.com/repos/huggingface/datasets/issues/4635/events
https://github.com/huggingface/datasets/issues/4635
1,294,475,931
I_kwDODunzps5NKCKb
4,635
Dataset Viewer issue for vadis/sv-ident
{ "avatar_url": "https://avatars.githubusercontent.com/u/20404466?v=4", "events_url": "https://api.github.com/users/e-tornike/events{/privacy}", "followers_url": "https://api.github.com/users/e-tornike/followers", "following_url": "https://api.github.com/users/e-tornike/following{/other_user}", "gists_url": "https://api.github.com/users/e-tornike/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/e-tornike", "id": 20404466, "login": "e-tornike", "node_id": "MDQ6VXNlcjIwNDA0NDY2", "organizations_url": "https://api.github.com/users/e-tornike/orgs", "received_events_url": "https://api.github.com/users/e-tornike/received_events", "repos_url": "https://api.github.com/users/e-tornike/repos", "site_admin": false, "starred_url": "https://api.github.com/users/e-tornike/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/e-tornike/subscriptions", "type": "User", "url": "https://api.github.com/users/e-tornike" }
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
[ "Thanks for reporting, @e-tornike \r\n\r\nSome context:\r\n- #4527 \r\n\r\nThe dataset loads locally in streaming mode:\r\n```python\r\nIn [2]: from datasets import load_dataset; ds = load_dataset(\"vadis/sv-ident\", split=\"validation\", streaming=True); item = next(iter(ds)); item\r\nUsing custom data configurati...
2022-07-05T15:48:13Z
2022-07-06T07:13:33Z
2022-07-06T07:12:14Z
NONE
null
null
null
### Link https://huggingface.co/datasets/vadis/sv-ident/viewer/default/validation ### Description Error message when loading validation split in the viewer: ``` Status code: 400 Exception: Status400Error Message: The split cache is empty. ``` ### Owner _No response_
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4635/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4635/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/1450
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1450/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1450/comments
https://api.github.com/repos/huggingface/datasets/issues/1450/events
https://github.com/huggingface/datasets/pull/1450
761,102,429
MDExOlB1bGxSZXF1ZXN0NTM1ODIwNjg0
1,450
Fix version in bible_para
{ "avatar_url": "https://avatars.githubusercontent.com/u/1183441?v=4", "events_url": "https://api.github.com/users/abhishekkrthakur/events{/privacy}", "followers_url": "https://api.github.com/users/abhishekkrthakur/followers", "following_url": "https://api.github.com/users/abhishekkrthakur/following{/other_user}", "gists_url": "https://api.github.com/users/abhishekkrthakur/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/abhishekkrthakur", "id": 1183441, "login": "abhishekkrthakur", "node_id": "MDQ6VXNlcjExODM0NDE=", "organizations_url": "https://api.github.com/users/abhishekkrthakur/orgs", "received_events_url": "https://api.github.com/users/abhishekkrthakur/received_events", "repos_url": "https://api.github.com/users/abhishekkrthakur/repos", "site_admin": false, "starred_url": "https://api.github.com/users/abhishekkrthakur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/abhishekkrthakur/subscriptions", "type": "User", "url": "https://api.github.com/users/abhishekkrthakur" }
[]
closed
false
null
[]
null
[]
2020-12-10T10:13:55Z
2020-12-11T16:40:41Z
2020-12-11T16:40:40Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1450.diff", "html_url": "https://github.com/huggingface/datasets/pull/1450", "merged_at": "2020-12-11T16:40:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/1450.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1450" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1450/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1450/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4628
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4628/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4628/comments
https://api.github.com/repos/huggingface/datasets/issues/4628/events
https://github.com/huggingface/datasets/pull/4628
1,293,361,308
PR_kwDODunzps46zvFJ
4,628
Fix time type `_arrow_to_datasets_dtype` conversion
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-07-04T16:20:15Z
2022-07-07T14:08:38Z
2022-07-07T13:57:12Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4628.diff", "html_url": "https://github.com/huggingface/datasets/pull/4628", "merged_at": "2022-07-07T13:57:11Z", "patch_url": "https://github.com/huggingface/datasets/pull/4628.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4628" }
Fix #4620 The issue stems from the fact that `pa.array([time_data]).type` returns `DataType(time64[unit])`, which doesn't expose the `unit` attribute, instead of `Time64Type(time64[unit])`. I believe this is a bug in PyArrow. Luckily, the both types have the same `str()`, so in this PR I call `pa.type_for_alias(str(type))` to convert them both to the `Time64Type(time64[unit])` format. cc @severo
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4628/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4628/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/1373
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1373/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1373/comments
https://api.github.com/repos/huggingface/datasets/issues/1373/events
https://github.com/huggingface/datasets/pull/1373
760,280,869
MDExOlB1bGxSZXF1ZXN0NTM1MTM5MTY0
1,373
Add OPUS ECB Dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/1183441?v=4", "events_url": "https://api.github.com/users/abhishekkrthakur/events{/privacy}", "followers_url": "https://api.github.com/users/abhishekkrthakur/followers", "following_url": "https://api.github.com/users/abhishekkrthakur/following{/other_user}", "gists_url": "https://api.github.com/users/abhishekkrthakur/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/abhishekkrthakur", "id": 1183441, "login": "abhishekkrthakur", "node_id": "MDQ6VXNlcjExODM0NDE=", "organizations_url": "https://api.github.com/users/abhishekkrthakur/orgs", "received_events_url": "https://api.github.com/users/abhishekkrthakur/received_events", "repos_url": "https://api.github.com/users/abhishekkrthakur/repos", "site_admin": false, "starred_url": "https://api.github.com/users/abhishekkrthakur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/abhishekkrthakur/subscriptions", "type": "User", "url": "https://api.github.com/users/abhishekkrthakur" }
[]
closed
false
null
[]
null
[]
2020-12-09T12:18:22Z
2020-12-10T15:25:55Z
2020-12-10T15:25:54Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1373.diff", "html_url": "https://github.com/huggingface/datasets/pull/1373", "merged_at": "2020-12-10T15:25:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/1373.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1373" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1373/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1373/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2981
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2981/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2981/comments
https://api.github.com/repos/huggingface/datasets/issues/2981/events
https://github.com/huggingface/datasets/pull/2981
1,009,969,310
PR_kwDODunzps4sZ4ke
2,981
add wit dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4", "events_url": "https://api.github.com/users/nateraw/events{/privacy}", "followers_url": "https://api.github.com/users/nateraw/followers", "following_url": "https://api.github.com/users/nateraw/following{/other_user}", "gists_url": "https://api.github.com/users/nateraw/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nateraw", "id": 32437151, "login": "nateraw", "node_id": "MDQ6VXNlcjMyNDM3MTUx", "organizations_url": "https://api.github.com/users/nateraw/orgs", "received_events_url": "https://api.github.com/users/nateraw/received_events", "repos_url": "https://api.github.com/users/nateraw/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nateraw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nateraw/subscriptions", "type": "User", "url": "https://api.github.com/users/nateraw" }
[]
closed
false
null
[]
null
[ "Opening this up - Here's whats left to make it really shine:\r\n- [ ] Update dataset card (this is being worked on by the author)\r\n- [ ] Add `dataset_info.json`. Requires downloading the entire dataset. I believe @lhoestq mentioned he may have a machine he's using for this sort of thing.\r\n\r\nI think both of t...
2021-09-28T16:34:49Z
2022-05-05T14:26:41Z
2022-05-05T14:26:41Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/2981.diff", "html_url": "https://github.com/huggingface/datasets/pull/2981", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2981.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2981" }
Resolves #2902 based on conversation there - would also close #2810. Open to suggestions/help 😀 CC @hassiahk @lhoestq @yjernite
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 2, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2981/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2981/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2316
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2316/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2316/comments
https://api.github.com/repos/huggingface/datasets/issues/2316/events
https://github.com/huggingface/datasets/issues/2316
875,756,353
MDU6SXNzdWU4NzU3NTYzNTM=
2,316
Incorrect version specification for pyarrow
{ "avatar_url": "https://avatars.githubusercontent.com/u/32267027?v=4", "events_url": "https://api.github.com/users/cemilcengiz/events{/privacy}", "followers_url": "https://api.github.com/users/cemilcengiz/followers", "following_url": "https://api.github.com/users/cemilcengiz/following{/other_user}", "gists_url": "https://api.github.com/users/cemilcengiz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cemilcengiz", "id": 32267027, "login": "cemilcengiz", "node_id": "MDQ6VXNlcjMyMjY3MDI3", "organizations_url": "https://api.github.com/users/cemilcengiz/orgs", "received_events_url": "https://api.github.com/users/cemilcengiz/received_events", "repos_url": "https://api.github.com/users/cemilcengiz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cemilcengiz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cemilcengiz/subscriptions", "type": "User", "url": "https://api.github.com/users/cemilcengiz" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[ "Fixed by #2317." ]
2021-05-04T19:15:11Z
2021-05-05T10:10:03Z
2021-05-05T10:10:03Z
CONTRIBUTOR
null
null
null
## Describe the bug The pyarrow dependency is incorrectly specified in setup.py file, in [this line](https://github.com/huggingface/datasets/blob/3a3e5a4da20bfcd75f8b6a6869b240af8feccc12/setup.py#L77). Also as a snippet: ```python "pyarrow>=1.0.0<4.0.0", ``` ## Steps to reproduce the bug ```bash pip install "pyarrow>=1.0.0<4.0.0" ``` ## Expected results It is expected to get a pyarrow version between 1.0.0 (inclusive) and 4.0.0 (exclusive). ## Actual results pip ignores the specified versions since there is a missing comma between the lower and upper limits. Therefore, pip installs the latest pyarrow version from PYPI, which is 4.0.0. This is especially problematic since "conda env export" fails due to incorrect version specification. Here is the conda error as well: ```bash conda env export InvalidVersionSpec: Invalid version '1.0.0<4.0.0': invalid character(s) ``` ## Fix suggestion Put a comma between the version limits which means replacing the line in setup.py file with the following: ```python "pyarrow>=1.0.0,<4.0.0", ``` ## Versions Paste the output of the following code: ```python - Datasets: 1.6.2 - Python: 3.7.10 (default, Feb 26 2021, 18:47:35) [GCC 7.3.0] - Platform: Linux-5.4.0-42-generic-x86_64-with-debian-buster-sid ```
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2316/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2316/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/3704
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3704/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3704/comments
https://api.github.com/repos/huggingface/datasets/issues/3704/events
https://github.com/huggingface/datasets/issues/3704
1,132,042,631
I_kwDODunzps5DeZmH
3,704
OSCAR-2109 datasets are misaligned and truncated
{ "avatar_url": "https://avatars.githubusercontent.com/u/5794899?v=4", "events_url": "https://api.github.com/users/adrianeboyd/events{/privacy}", "followers_url": "https://api.github.com/users/adrianeboyd/followers", "following_url": "https://api.github.com/users/adrianeboyd/following{/other_user}", "gists_url": "https://api.github.com/users/adrianeboyd/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/adrianeboyd", "id": 5794899, "login": "adrianeboyd", "node_id": "MDQ6VXNlcjU3OTQ4OTk=", "organizations_url": "https://api.github.com/users/adrianeboyd/orgs", "received_events_url": "https://api.github.com/users/adrianeboyd/received_events", "repos_url": "https://api.github.com/users/adrianeboyd/repos", "site_admin": false, "starred_url": "https://api.github.com/users/adrianeboyd/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/adrianeboyd/subscriptions", "type": "User", "url": "https://api.github.com/users/adrianeboyd" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[ "Hi @adrianeboyd, thanks for reporting.\r\n\r\nThere is indeed a bug in that community dataset:\r\nLine:\r\n```python\r\nmetadata_and_text_files = list(zip(metadata_files, text_files))\r\n``` \r\nshould be replaced with\r\n```python\r\nmetadata_and_text_files = list(zip(sorted(metadata_files), sorted(text_files)))\...
2022-02-11T08:14:59Z
2022-03-17T18:01:04Z
2022-03-16T16:21:28Z
NONE
null
null
null
## Describe the bug The `oscar-corpus/OSCAR-2109` data appears to be misaligned and truncated by the dataset builder for subsets that contain more than one part and for cases where the texts contain non-unix newlines. ## Steps to reproduce the bug A few examples, although I'm not sure how deterministic the particular (mis)alignment is in various configurations: ```python from datasets import load_dataset dataset = load_dataset("oscar-corpus/OSCAR-2109", "deduplicated_fi", split="train", use_auth_token=True) entry = dataset[0] # entry["text"] is from fi_part_3.txt.gz # entry["meta"] is from fi_meta_part_2.jsonl.gz dataset = load_dataset("oscar-corpus/OSCAR-2109", "deduplicated_no", split="train", use_auth_token=True) entry = dataset[900000] # entry["text"] is from no_part_3.txt.gz and contains a blank line # entry["meta"] is from no_meta_part_1.jsonl.gz dataset = load_dataset("oscar-corpus/OSCAR-2109", "deduplicated_mk", split="train", streaming=True, use_auth_token=True) # 9088 texts in the dataset are empty ``` For `deduplicated_fi`, all exported raw texts from the dataset are 17GB rather than 20GB as reported in the data splits overview table. The token count with `wc -w` for the raw texts is 2,067,556,874 rather than the expected 2,357,264,196 from the data splits table. For `deduplicated_no` all exported raw texts contain 624,040,887 rather than the expected 776,354,517 tokens. For `deduplicated_mk` it is 122,236,936 rather than 134,544,934 tokens. I'm not expecting the `wc -w` counts to line up exactly with the data splits table, but for comparison the `wc -w` count for `deduplicated_mk` on the raw texts is 134,545,424. ## Issues * The meta / text files are not paired correctly when loading, so the extracted texts do not have the right offsets, the metadata is not associated with the correct text, and the text files may not be processed to the end or may be processed beyond the end (empty texts). * The line count offset is not reset per file so the texts aren't aligned to the right offsets in any parts beyond the first part, leading to truncation when in effect blank lines are not skipped. * Non-unix newline characters are treated as newlines when reading the text files while the metadata only counts unix newlines for its line offsets, leading to further misalignments between the metadata and the extracted texts, and which also results in truncation. ## Expected results All texts from the OSCAR release are extracted according to the metadata and aligned with the correct metadata. ## Fixes Not necessarily the exact fixes/checks you may want to use (I didn't test all languages or do any cross-platform testing, I'm not sure all the details are compatible with streaming), however to highlight the issues: ```diff diff --git a/OSCAR-2109.py b/OSCAR-2109.py index bbac1076..5eee8de7 100644 --- a/OSCAR-2109.py +++ b/OSCAR-2109.py @@ -20,6 +20,7 @@ import collections import gzip import json +import os import datasets @@ -387,9 +388,20 @@ class Oscar2109(datasets.GeneratorBasedBuilder): with open(checksum_file, encoding="utf-8") as f: data_filenames = [line.split()[1] for line in f if line] data_urls = [self.config.base_data_path + data_filename for data_filename in data_filenames] - text_files = dl_manager.download([url for url in data_urls if url.endswith(".txt.gz")]) - metadata_files = dl_manager.download([url for url in data_urls if url.endswith(".jsonl.gz")]) + # sort filenames so corresponding parts are aligned + text_files = sorted(dl_manager.download([url for url in data_urls if url.endswith(".txt.gz")])) + metadata_files = sorted(dl_manager.download([url for url in data_urls if url.endswith(".jsonl.gz")])) + assert len(text_files) == len(metadata_files) metadata_and_text_files = list(zip(metadata_files, text_files)) + for meta_path, text_path in metadata_and_text_files: + # check that meta/text part numbers are the same + if "part" in os.path.basename(text_path): + assert ( + os.path.basename(text_path).replace(".txt.gz", "").split("_")[-1] + == os.path.basename(meta_path).replace(".jsonl.gz", "").split("_")[-1] + ) + else: + assert len(metadata_and_text_files) == 1 return [ datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"metadata_and_text_files": metadata_and_text_files}), ] @@ -397,10 +409,14 @@ class Oscar2109(datasets.GeneratorBasedBuilder): def _generate_examples(self, metadata_and_text_files): """This function returns the examples in the raw (text) form by iterating on all the files.""" id_ = 0 - offset = 0 for meta_path, text_path in metadata_and_text_files: + # line offsets are per text file + offset = 0 logger.info("generating examples from = %s", text_path) - with gzip.open(open(text_path, "rb"), "rt", encoding="utf-8") as text_f: + # some texts contain non-Unix newlines that should not be + # interpreted as line breaks for the line counts in the metadata + # with readline() + with gzip.open(open(text_path, "rb"), "rt", encoding="utf-8", newline="\n") as text_f: with gzip.open(open(meta_path, "rb"), "rt", encoding="utf-8") as meta_f: for line in meta_f: # read meta @@ -411,7 +427,12 @@ class Oscar2109(datasets.GeneratorBasedBuilder): offset += 1 text_f.readline() # read text - text = "".join([text_f.readline() for _ in range(meta["nb_sentences"])]).rstrip() + text_lines = [text_f.readline() for _ in range(meta["nb_sentences"])] + # all lines contain text (no blank lines or EOF) + assert all(text_lines) + assert "\n" not in text_lines offset += meta["nb_sentences"] + # only strip the trailing newline + text = "".join(text_lines).rstrip("\n") yield id_, {"id": id_, "text": text, "meta": meta} id_ += 1 ``` I've tested this with a number of smaller deduplicated languages with 1-20 parts and the resulting datasets looked correct in terms of word count and size when compared to the data splits table and raw texts, and the text/metadata alignments were correct in all my spot checks. However, there are many many languages I didn't test and I'm not sure that there aren't any texts containing blank lines in the corpus, for instance. For the cases I tested, the assertions related to blank lines and EOF made it easier to verify that the text and metadata were aligned as intended, since there would be little chance of spurious alignments of variable-length texts across so much data.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3704/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3704/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/5137
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5137/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5137/comments
https://api.github.com/repos/huggingface/datasets/issues/5137/events
https://github.com/huggingface/datasets/issues/5137
1,414,642,723
I_kwDODunzps5UUbwj
5,137
Align task tags in dataset metadata
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
[ "I removed all the invalid task_ids in datasts without namespace, based on the <s>(internal)</s> types.ts", "(Types.ts is not internal it's public)", "I have opened PRs to fix the task_ids in all datasets within a namespace as well.\r\n\r\nWorking on task_categories...", "For future reference: this fix had so...
2022-10-19T09:41:42Z
2022-11-10T05:25:58Z
2022-10-25T06:17:00Z
MEMBER
null
null
null
## Describe Once we have agreed on a common naming for task tags for all open source projects, we should align on them. ## Steps - [x] Align task tags in canonical datasets - [x] task_categories: 4 datasets - [x] task_ids (by @lhoestq) - [x] Open PRs in community datasets - [x] task_categories: 451 datasets - [x] task_ids: 556 datasets
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5137/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5137/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6100
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6100/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6100/comments
https://api.github.com/repos/huggingface/datasets/issues/6100/events
https://github.com/huggingface/datasets/issues/6100
1,828,118,930
I_kwDODunzps5s9uGS
6,100
TypeError when loading from GCP bucket
{ "avatar_url": "https://avatars.githubusercontent.com/u/16692099?v=4", "events_url": "https://api.github.com/users/bilelomrani1/events{/privacy}", "followers_url": "https://api.github.com/users/bilelomrani1/followers", "following_url": "https://api.github.com/users/bilelomrani1/following{/other_user}", "gists_url": "https://api.github.com/users/bilelomrani1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bilelomrani1", "id": 16692099, "login": "bilelomrani1", "node_id": "MDQ6VXNlcjE2NjkyMDk5", "organizations_url": "https://api.github.com/users/bilelomrani1/orgs", "received_events_url": "https://api.github.com/users/bilelomrani1/received_events", "repos_url": "https://api.github.com/users/bilelomrani1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bilelomrani1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bilelomrani1/subscriptions", "type": "User", "url": "https://api.github.com/users/bilelomrani1" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
[ "Thanks for reporting, @bilelomrani1.\r\n\r\nWe are fixing it. ", "We have fixed it. We are planning to do a patch release today." ]
2023-07-30T23:03:00Z
2023-08-03T10:00:48Z
2023-08-01T10:38:55Z
NONE
null
null
null
### Describe the bug Loading a dataset from a GCP bucket raises a type error. This bug was introduced recently (either in 2.14 or 2.14.1), and appeared during a migration from 2.13.1. ### Steps to reproduce the bug Load any file from a GCP bucket: ```python import datasets datasets.load_dataset("json", data_files=["gs://..."]) ``` The following exception is raised: ```python Traceback (most recent call last): ... packages/datasets/data_files.py", line 335, in resolve_pattern protocol_prefix = fs.protocol + "://" if fs.protocol != "file" else "" TypeError: can only concatenate tuple (not "str") to tuple ``` With a `GoogleFileSystem`, the attribute `fs.protocol` is a tuple `('gs', 'gcs')` and hence cannot be concatenated with a string. ### Expected behavior The file should be loaded without exception. ### Environment info - `datasets` version: 2.14.1 - Platform: macOS-13.2.1-x86_64-i386-64bit - Python version: 3.10.12 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6100/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6100/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/4371
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4371/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4371/comments
https://api.github.com/repos/huggingface/datasets/issues/4371/events
https://github.com/huggingface/datasets/pull/4371
1,241,500,906
PR_kwDODunzps44GzSZ
4,371
Add missing language tags for udhr dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-05-19T09:34:10Z
2022-06-08T12:03:24Z
2022-05-20T09:43:10Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4371.diff", "html_url": "https://github.com/huggingface/datasets/pull/4371", "merged_at": "2022-05-20T09:43:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/4371.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4371" }
Related to #4362.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4371/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4371/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4343
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4343/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4343/comments
https://api.github.com/repos/huggingface/datasets/issues/4343/events
https://github.com/huggingface/datasets/issues/4343
1,234,864,168
I_kwDODunzps5Jmogo
4,343
Metrics documentation is not accessible in the datasets doc UI
{ "avatar_url": "https://avatars.githubusercontent.com/u/9808326?v=4", "events_url": "https://api.github.com/users/fxmarty/events{/privacy}", "followers_url": "https://api.github.com/users/fxmarty/followers", "following_url": "https://api.github.com/users/fxmarty/following{/other_user}", "gists_url": "https://api.github.com/users/fxmarty/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/fxmarty", "id": 9808326, "login": "fxmarty", "node_id": "MDQ6VXNlcjk4MDgzMjY=", "organizations_url": "https://api.github.com/users/fxmarty/orgs", "received_events_url": "https://api.github.com/users/fxmarty/received_events", "repos_url": "https://api.github.com/users/fxmarty/repos", "site_admin": false, "starred_url": "https://api.github.com/users/fxmarty/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fxmarty/subscriptions", "type": "User", "url": "https://api.github.com/users/fxmarty" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "d722e8", "default": fals...
closed
false
null
[]
null
[ "Hey @fxmarty :) Yes we are working on showing the docs of all the metrics on the Hugging face website. If you want to follow the advancements you can check the [evaluate](https://github.com/huggingface/evaluate) repository cc @lvwerra @sashavor " ]
2022-05-13T07:46:30Z
2022-06-03T08:50:25Z
2022-06-03T08:50:25Z
CONTRIBUTOR
null
null
null
**Is your feature request related to a problem? Please describe.** Search for a metric name like "seqeval" yields no results on https://huggingface.co/docs/datasets/master/en/index . One needs to go look in `datasets/metrics/README.md` to find the doc. Even in the `README.md`, it can be hard to understand what the metric expects as an input, for example for `squad` there is a [key `id`](https://github.com/huggingface/datasets/blob/1a4c185663a6958f48ec69624473fdc154a36a9d/metrics/squad/squad.py#L42) documented only in the function doc but not in the `README.md`, and one needs to go look into the code to understand what the metric expects. **Describe the solution you'd like** Have the documentation for metrics appear as well in the doc UI, e.g. this https://github.com/huggingface/datasets/blob/1a4c185663a6958f48ec69624473fdc154a36a9d/metrics/squad/squad.py#L21-L63 I know there are plans to migrate metrics to the evaluate library, but just pointing this out.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4343/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4343/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/2737
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2737/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2737/comments
https://api.github.com/repos/huggingface/datasets/issues/2737/events
https://github.com/huggingface/datasets/issues/2737
957,124,881
MDU6SXNzdWU5NTcxMjQ4ODE=
2,737
SacreBLEU update
{ "avatar_url": "https://avatars.githubusercontent.com/u/46989091?v=4", "events_url": "https://api.github.com/users/devrimcavusoglu/events{/privacy}", "followers_url": "https://api.github.com/users/devrimcavusoglu/followers", "following_url": "https://api.github.com/users/devrimcavusoglu/following{/other_user}", "gists_url": "https://api.github.com/users/devrimcavusoglu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/devrimcavusoglu", "id": 46989091, "login": "devrimcavusoglu", "node_id": "MDQ6VXNlcjQ2OTg5MDkx", "organizations_url": "https://api.github.com/users/devrimcavusoglu/orgs", "received_events_url": "https://api.github.com/users/devrimcavusoglu/received_events", "repos_url": "https://api.github.com/users/devrimcavusoglu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/devrimcavusoglu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/devrimcavusoglu/subscriptions", "type": "User", "url": "https://api.github.com/users/devrimcavusoglu" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[ "Hi @devrimcavusoglu, \r\nI tried your code with latest version of `datasets`and `sacrebleu==1.5.1` and it's running fine after changing one small thing:\r\n```\r\nsacrebleu = datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of t...
2021-07-30T23:53:08Z
2021-09-22T10:47:41Z
2021-08-03T04:23:37Z
NONE
null
null
null
With the latest release of [sacrebleu](https://github.com/mjpost/sacrebleu), `datasets.metrics.sacrebleu` is broken, and getting error. AttributeError: module 'sacrebleu' has no attribute 'DEFAULT_TOKENIZER' this happens since in new version of sacrebleu there is no `DEFAULT_TOKENIZER`, but sacrebleu.py tries to import it anyways. This can be fixed currently with fixing `sacrebleu==1.5.0` ## Steps to reproduce the bug ```python sacrebleu= datasets.load_metric('sacrebleu') predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] references = ["It is a guide to action that ensures that the military will forever heed Party commands"] results = sacrebleu.compute(predictions=predictions, references=references) print(results) ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.11.0 - Platform: Windows-10-10.0.19041-SP0 - Python version: Python 3.8.0 - PyArrow version: 5.0.0
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2737/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2737/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/1900
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1900/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1900/comments
https://api.github.com/repos/huggingface/datasets/issues/1900/events
https://github.com/huggingface/datasets/pull/1900
810,512,488
MDExOlB1bGxSZXF1ZXN0NTc1MTkxNTc3
1,900
Issue #1895: Bugfix for string_to_arrow timestamp[ns] support
{ "avatar_url": "https://avatars.githubusercontent.com/u/7731709?v=4", "events_url": "https://api.github.com/users/justin-yan/events{/privacy}", "followers_url": "https://api.github.com/users/justin-yan/followers", "following_url": "https://api.github.com/users/justin-yan/following{/other_user}", "gists_url": "https://api.github.com/users/justin-yan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/justin-yan", "id": 7731709, "login": "justin-yan", "node_id": "MDQ6VXNlcjc3MzE3MDk=", "organizations_url": "https://api.github.com/users/justin-yan/orgs", "received_events_url": "https://api.github.com/users/justin-yan/received_events", "repos_url": "https://api.github.com/users/justin-yan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/justin-yan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/justin-yan/subscriptions", "type": "User", "url": "https://api.github.com/users/justin-yan" }
[]
closed
false
null
[]
null
[ "OK! Thank you for the review - I will follow up with a separate PR for the comments here (https://github.com/huggingface/datasets/pull/1900#discussion_r578319725)!" ]
2021-02-17T20:26:04Z
2021-02-19T18:27:11Z
2021-02-19T18:27:11Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1900.diff", "html_url": "https://github.com/huggingface/datasets/pull/1900", "merged_at": "2021-02-19T18:27:11Z", "patch_url": "https://github.com/huggingface/datasets/pull/1900.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1900" }
Should resolve https://github.com/huggingface/datasets/issues/1895 The main part of this PR adds additional parsing in `string_to_arrow` to convert the timestamp dtypes that result from `str(pa_type)` back into the pa.DataType TimestampType. While adding unit-testing, I noticed that support for the double/float types also don't invert correctly, so I added them, which I believe would hypothetically make this section of `Value` redundant: ``` def __post_init__(self): if self.dtype == "double": # fix inferred type self.dtype = "float64" if self.dtype == "float": # fix inferred type self.dtype = "float32" ``` However, since I think Value.dtype is part of the public interface, removing that would result in a backward-incompatible change, so I didn't muck with that. The rest of the PR consists of docstrings that I added while developing locally so I could keep track of which functions were supposed to be inverses of each other, and thought I'd include them initially in case you want to keep them around, but I'm happy to delete or remove any of them at your request!
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1900/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1900/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/1534
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1534/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1534/comments
https://api.github.com/repos/huggingface/datasets/issues/1534/events
https://github.com/huggingface/datasets/pull/1534
764,934,681
MDExOlB1bGxSZXF1ZXN0NTM4Nzc1Njky
1,534
adding dataset for diplomacy detection
{ "avatar_url": "https://avatars.githubusercontent.com/u/15351802?v=4", "events_url": "https://api.github.com/users/MisbahKhan789/events{/privacy}", "followers_url": "https://api.github.com/users/MisbahKhan789/followers", "following_url": "https://api.github.com/users/MisbahKhan789/following{/other_user}", "gists_url": "https://api.github.com/users/MisbahKhan789/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/MisbahKhan789", "id": 15351802, "login": "MisbahKhan789", "node_id": "MDQ6VXNlcjE1MzUxODAy", "organizations_url": "https://api.github.com/users/MisbahKhan789/orgs", "received_events_url": "https://api.github.com/users/MisbahKhan789/received_events", "repos_url": "https://api.github.com/users/MisbahKhan789/repos", "site_admin": false, "starred_url": "https://api.github.com/users/MisbahKhan789/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MisbahKhan789/subscriptions", "type": "User", "url": "https://api.github.com/users/MisbahKhan789" }
[]
closed
false
null
[]
null
[ "Requested changes made and new PR submitted here: https://github.com/huggingface/datasets/pull/1580 " ]
2020-12-13T04:38:43Z
2020-12-15T19:52:52Z
2020-12-15T19:52:25Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1534.diff", "html_url": "https://github.com/huggingface/datasets/pull/1534", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1534.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1534" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1534/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1534/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2390
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2390/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2390/comments
https://api.github.com/repos/huggingface/datasets/issues/2390/events
https://github.com/huggingface/datasets/pull/2390
897,903,642
MDExOlB1bGxSZXF1ZXN0NjQ5ODQ0NjQ2
2,390
Add check for task templates on dataset load
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
[]
closed
false
null
[]
null
[ "LGTM now, thank you =)" ]
2021-05-21T10:16:57Z
2021-05-21T15:49:09Z
2021-05-21T15:49:06Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/2390.diff", "html_url": "https://github.com/huggingface/datasets/pull/2390", "merged_at": "2021-05-21T15:49:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/2390.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2390" }
This PR adds a check that the features of a dataset match the schema of each compatible task template.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2390/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2390/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/5409
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5409/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5409/comments
https://api.github.com/repos/huggingface/datasets/issues/5409/events
https://github.com/huggingface/datasets/pull/5409
1,520,374,219
PR_kwDODunzps5Gs3nL
5,409
Fix deprecation warning when use_auth_token passed to download_and_prepare
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-01-05T09:10:58Z
2023-01-06T11:06:16Z
2023-01-06T10:59:13Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/5409.diff", "html_url": "https://github.com/huggingface/datasets/pull/5409", "merged_at": "2023-01-06T10:59:13Z", "patch_url": "https://github.com/huggingface/datasets/pull/5409.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5409" }
The `DatasetBuilder.download_and_prepare` argument `use_auth_token` was deprecated in: - #5302 However, `use_auth_token` is still passed to `download_and_prepare` in our built-in `io` readers (csv, json, parquet,...). This PR fixes it, so that no deprecation warning is raised. Fix #5407.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5409/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5409/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4649
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4649/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4649/comments
https://api.github.com/repos/huggingface/datasets/issues/4649/events
https://github.com/huggingface/datasets/issues/4649
1,296,673,712
I_kwDODunzps5NSauw
4,649
Add PAQ dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/4755430?v=4", "events_url": "https://api.github.com/users/omarespejel/events{/privacy}", "followers_url": "https://api.github.com/users/omarespejel/followers", "following_url": "https://api.github.com/users/omarespejel/following{/other_user}", "gists_url": "https://api.github.com/users/omarespejel/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/omarespejel", "id": 4755430, "login": "omarespejel", "node_id": "MDQ6VXNlcjQ3NTU0MzA=", "organizations_url": "https://api.github.com/users/omarespejel/orgs", "received_events_url": "https://api.github.com/users/omarespejel/received_events", "repos_url": "https://api.github.com/users/omarespejel/repos", "site_admin": false, "starred_url": "https://api.github.com/users/omarespejel/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/omarespejel/subscriptions", "type": "User", "url": "https://api.github.com/users/omarespejel" }
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
null
[]
null
[ "uploaded dataset [here](https://huggingface.co/datasets/embedding-data/PAQ_pairs)" ]
2022-07-07T01:29:42Z
2022-07-14T02:06:27Z
2022-07-14T02:06:27Z
NONE
null
null
null
## Adding a Dataset - **Name:** *PAQ* - **Description:** *This repository contains code and models to support the research paper PAQ: 65 Million Probably-Asked Questions and What You Can Do With Them* - **Paper:** *https://arxiv.org/abs/2102.07033* - **Data:** *https://huggingface.co/datasets/sentence-transformers/embedding-training-data/resolve/main/PAQ_pairs.jsonl.gz* - **Motivation:** *Dataset for training and evaluating models of conversational response*
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4649/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4649/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/1330
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1330/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1330/comments
https://api.github.com/repos/huggingface/datasets/issues/1330/events
https://github.com/huggingface/datasets/pull/1330
759,657,324
MDExOlB1bGxSZXF1ZXN0NTM0NjI0MzMx
1,330
added un_ga dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/26374564?v=4", "events_url": "https://api.github.com/users/param087/events{/privacy}", "followers_url": "https://api.github.com/users/param087/followers", "following_url": "https://api.github.com/users/param087/following{/other_user}", "gists_url": "https://api.github.com/users/param087/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/param087", "id": 26374564, "login": "param087", "node_id": "MDQ6VXNlcjI2Mzc0NTY0", "organizations_url": "https://api.github.com/users/param087/orgs", "received_events_url": "https://api.github.com/users/param087/received_events", "repos_url": "https://api.github.com/users/param087/repos", "site_admin": false, "starred_url": "https://api.github.com/users/param087/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/param087/subscriptions", "type": "User", "url": "https://api.github.com/users/param087" }
[]
closed
false
null
[]
null
[ "Looks like this PR includes changes about many other files than the ones for un_ga\r\n\r\nCan you create another branch an another PR please ?", "@lhoestq, Thank you for suggestions. I have made the changes and raised the new PR https://github.com/huggingface/datasets/pull/1569. " ]
2020-12-08T17:58:38Z
2020-12-14T17:52:34Z
2020-12-14T17:52:34Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1330.diff", "html_url": "https://github.com/huggingface/datasets/pull/1330", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1330.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1330" }
Hi :hugs:, This is a PR for [United nations general assembly resolutions: A six-language parallel corpus](http://opus.nlpl.eu/UN.php) dataset
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1330/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1330/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3202
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3202/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3202/comments
https://api.github.com/repos/huggingface/datasets/issues/3202/events
https://github.com/huggingface/datasets/issues/3202
1,043,213,660
I_kwDODunzps4-Li1c
3,202
Add mIoU metric
{ "avatar_url": "https://avatars.githubusercontent.com/u/48327001?v=4", "events_url": "https://api.github.com/users/NielsRogge/events{/privacy}", "followers_url": "https://api.github.com/users/NielsRogge/followers", "following_url": "https://api.github.com/users/NielsRogge/following{/other_user}", "gists_url": "https://api.github.com/users/NielsRogge/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NielsRogge", "id": 48327001, "login": "NielsRogge", "node_id": "MDQ6VXNlcjQ4MzI3MDAx", "organizations_url": "https://api.github.com/users/NielsRogge/orgs", "received_events_url": "https://api.github.com/users/NielsRogge/received_events", "repos_url": "https://api.github.com/users/NielsRogge/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NielsRogge/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NielsRogge/subscriptions", "type": "User", "url": "https://api.github.com/users/NielsRogge" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[ "Resolved via https://github.com/huggingface/datasets/pull/3745." ]
2021-11-03T08:42:32Z
2022-06-01T17:39:05Z
2022-06-01T17:39:04Z
CONTRIBUTOR
null
null
null
**Is your feature request related to a problem? Please describe.** Recently, some semantic segmentation models were added to HuggingFace Transformers, including [SegFormer](https://huggingface.co/transformers/model_doc/segformer.html) and [BEiT](https://huggingface.co/transformers/model_doc/beit.html). Semantic segmentation (which is the task of labeling every pixel of an image with a corresponding class) is typically evaluated using the Mean Intersection and Union (mIoU). Together with the upcoming Image Feature, adding this metric could be very handy when creating example scripts to fine-tune any Transformer-based model on a semantic segmentation dataset. An implementation can be found [here](https://github.com/open-mmlab/mmsegmentation/blob/504965184c3e6bc9ec43af54237129ef21981a5f/mmseg/core/evaluation/metrics.py#L132) for instance.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3202/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3202/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/1781
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1781/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1781/comments
https://api.github.com/repos/huggingface/datasets/issues/1781/events
https://github.com/huggingface/datasets/issues/1781
793,914,556
MDU6SXNzdWU3OTM5MTQ1NTY=
1,781
AttributeError: module 'pyarrow' has no attribute 'PyExtensionType' during import
{ "avatar_url": "https://avatars.githubusercontent.com/u/45964869?v=4", "events_url": "https://api.github.com/users/PalaashAgrawal/events{/privacy}", "followers_url": "https://api.github.com/users/PalaashAgrawal/followers", "following_url": "https://api.github.com/users/PalaashAgrawal/following{/other_user}", "gists_url": "https://api.github.com/users/PalaashAgrawal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/PalaashAgrawal", "id": 45964869, "login": "PalaashAgrawal", "node_id": "MDQ6VXNlcjQ1OTY0ODY5", "organizations_url": "https://api.github.com/users/PalaashAgrawal/orgs", "received_events_url": "https://api.github.com/users/PalaashAgrawal/received_events", "repos_url": "https://api.github.com/users/PalaashAgrawal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/PalaashAgrawal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/PalaashAgrawal/subscriptions", "type": "User", "url": "https://api.github.com/users/PalaashAgrawal" }
[]
closed
false
null
[]
null
[ "Hi ! I'm not able to reproduce the issue. Can you try restarting your runtime ?\r\n\r\nThe PyExtensionType is available in pyarrow starting 0.17.1 iirc. If restarting your runtime doesn't fix this, can you try updating pyarrow ?\r\n```\r\npip install pyarrow --upgrade\r\n```", "We should bump up the version test...
2021-01-26T04:18:35Z
2022-10-05T12:37:06Z
2022-10-05T12:37:06Z
NONE
null
null
null
I'm using Colab. And suddenly this morning, there is this error. Have a look below! ![screenshot-colab research google com-2021 01 26-08-15-36](https://user-images.githubusercontent.com/45964869/105799890-fdaf3b80-5fae-11eb-8f06-11b65cdccc30.png)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1781/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1781/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/5478
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5478/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5478/comments
https://api.github.com/repos/huggingface/datasets/issues/5478/events
https://github.com/huggingface/datasets/pull/5478
1,560,357,583
PR_kwDODunzps5ItXQG
5,478
Tip for recomputing metadata
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-01-27T20:01:22Z
2023-01-30T19:22:21Z
2023-01-30T19:15:26Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/5478.diff", "html_url": "https://github.com/huggingface/datasets/pull/5478", "merged_at": "2023-01-30T19:15:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/5478.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5478" }
From this [feedback](https://discuss.huggingface.co/t/nonmatchingsplitssizeserror/30033) on the forum, thought I'd include a tip for recomputing the metadata numbers if it is your own dataset.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5478/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5478/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/5742
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5742/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5742/comments
https://api.github.com/repos/huggingface/datasets/issues/5742/events
https://github.com/huggingface/datasets/pull/5742
1,666,209,738
PR_kwDODunzps5OOH-W
5,742
Warning specifying future change in to_tf_dataset behaviour
{ "avatar_url": "https://avatars.githubusercontent.com/u/22614925?v=4", "events_url": "https://api.github.com/users/amyeroberts/events{/privacy}", "followers_url": "https://api.github.com/users/amyeroberts/followers", "following_url": "https://api.github.com/users/amyeroberts/following{/other_user}", "gists_url": "https://api.github.com/users/amyeroberts/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/amyeroberts", "id": 22614925, "login": "amyeroberts", "node_id": "MDQ6VXNlcjIyNjE0OTI1", "organizations_url": "https://api.github.com/users/amyeroberts/orgs", "received_events_url": "https://api.github.com/users/amyeroberts/received_events", "repos_url": "https://api.github.com/users/amyeroberts/repos", "site_admin": false, "starred_url": "https://api.github.com/users/amyeroberts/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amyeroberts/subscriptions", "type": "User", "url": "https://api.github.com/users/amyeroberts" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-04-13T11:10:00Z
2023-04-21T13:18:14Z
2023-04-21T13:11:09Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/5742.diff", "html_url": "https://github.com/huggingface/datasets/pull/5742", "merged_at": "2023-04-21T13:11:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/5742.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5742" }
Warning specifying future changes happening to `to_tf_dataset` behaviour when #5602 is merged in
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5742/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5742/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2905
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2905/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2905/comments
https://api.github.com/repos/huggingface/datasets/issues/2905/events
https://github.com/huggingface/datasets/pull/2905
995,843,964
PR_kwDODunzps4ruL5X
2,905
Update BibTeX entry
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[]
2021-09-14T10:16:17Z
2021-09-14T12:25:37Z
2021-09-14T12:25:37Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/2905.diff", "html_url": "https://github.com/huggingface/datasets/pull/2905", "merged_at": "2021-09-14T12:25:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/2905.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2905" }
Update BibTeX entry.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2905/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2905/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/5964
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5964/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5964/comments
https://api.github.com/repos/huggingface/datasets/issues/5964/events
https://github.com/huggingface/datasets/pull/5964
1,763,513,574
PR_kwDODunzps5TVweZ
5,964
Always return list in `list_datasets`
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-19T13:07:08Z
2023-06-19T17:29:37Z
2023-06-19T17:22:41Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/5964.diff", "html_url": "https://github.com/huggingface/datasets/pull/5964", "merged_at": "2023-06-19T17:22:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/5964.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5964" }
Fix #5925 Plus, deprecate `list_datasets`/`inspect_dataset` in favor of `huggingface_hub.list_datasets`/"git clone workflow" (downloads data files)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5964/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5964/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/6207
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6207/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6207/comments
https://api.github.com/repos/huggingface/datasets/issues/6207/events
https://github.com/huggingface/datasets/issues/6207
1,879,555,234
I_kwDODunzps5wB7yi
6,207
No-script datasets with ZIP files do not load
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
[]
2023-09-04T05:50:27Z
2023-09-04T09:13:33Z
2023-09-04T09:13:33Z
MEMBER
null
null
null
While investigating an issue on a Hub dataset, I have discovered the no-script datasets containing ZIP files do not load. For example, that no-script dataset containing ZIP files, raises NonMatchingSplitsSizesError: ```python In [2]: ds = load_dataset("sidovic/LearningQ-qg") NonMatchingSplitsSizesError: [ { 'expected': SplitInfo(name='train', num_bytes=0, num_examples=188660, shard_lengths=None, dataset_name=None), 'recorded': SplitInfo(name='train', num_bytes=0, num_examples=0, shard_lengths=None, dataset_name='learning_q-qg') }, { 'expected': SplitInfo(name='validation', num_bytes=0, num_examples=20630, shard_lengths=None, dataset_name=None), 'recorded': SplitInfo(name='validation', num_bytes=0, num_examples=0, shard_lengths=None, dataset_name='learning_q-qg') }, { 'expected': SplitInfo(name='test', num_bytes=0, num_examples=18227, shard_lengths=None, dataset_name=None), 'recorded': SplitInfo(name='test', num_bytes=0, num_examples=0, shard_lengths=None, dataset_name='learning_q-qg') } ] ``` As another example, a no-script dataset containing just a (CSV)-ZIP file, raises a DatasetGenerationError: ``` > num_examples, num_bytes = writer.finalize() src/datasets/builder.py:1949: > raise SchemaInferenceError("Please pass `features` or at least one example when writing data") E datasets.arrow_writer.SchemaInferenceError: Please pass `features` or at least one example when writing data src/datasets/arrow_writer.py:598: SchemaInferenceError The above exception was the direct cause of the following exception: src/datasets/load.py:2143: in load_dataset builder_instance.download_and_prepare( src/datasets/builder.py:954: in download_and_prepare self._download_and_prepare( src/datasets/builder.py:1049: in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) src/datasets/builder.py:1813: in _prepare_split for job_id, done, content in self._prepare_split_single( > raise DatasetGenerationError("An error occurred while generating the dataset") from e E datasets.builder.DatasetGenerationError: An error occurred while generating the dataset src/datasets/builder.py:1958: DatasetGenerationError ``` After investigating, I think this bug was introduced in this PR: - #5972 Related to: - https://huggingface.co/datasets/sidovic/LearningQ-qg/discussions/1 CC: @lhoestq
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6207/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6207/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/5788
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5788/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5788/comments
https://api.github.com/repos/huggingface/datasets/issues/5788/events
https://github.com/huggingface/datasets/pull/5788
1,681,136,256
PR_kwDODunzps5O_v4B
5,788
Prepare tests for hfh 0.14
{ "avatar_url": "https://avatars.githubusercontent.com/u/11801849?v=4", "events_url": "https://api.github.com/users/Wauplin/events{/privacy}", "followers_url": "https://api.github.com/users/Wauplin/followers", "following_url": "https://api.github.com/users/Wauplin/following{/other_user}", "gists_url": "https://api.github.com/users/Wauplin/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Wauplin", "id": 11801849, "login": "Wauplin", "node_id": "MDQ6VXNlcjExODAxODQ5", "organizations_url": "https://api.github.com/users/Wauplin/orgs", "received_events_url": "https://api.github.com/users/Wauplin/received_events", "repos_url": "https://api.github.com/users/Wauplin/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Wauplin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wauplin/subscriptions", "type": "User", "url": "https://api.github.com/users/Wauplin" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-04-24T12:13:03Z
2023-04-25T14:32:56Z
2023-04-25T14:25:30Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/5788.diff", "html_url": "https://github.com/huggingface/datasets/pull/5788", "merged_at": "2023-04-25T14:25:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/5788.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5788" }
Related to the coming release of `huggingface_hub==0.14.0`. It will break some internal tests. The PR fixes these tests. Let's double-check the CI but I expect the fixed tests to be running fine with both `hfh<=0.13.4` and `hfh==0.14`. Worth case scenario, existing PRs will have to be rebased once this fix is merged. See related [discussion](https://huggingface.slack.com/archives/C02V5EA0A95/p1682337463368609?thread_ts=1681994202.635609&cid=C02V5EA0A95) (private slack). cc @lhoestq
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5788/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5788/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4249
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4249/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4249/comments
https://api.github.com/repos/huggingface/datasets/issues/4249/events
https://github.com/huggingface/datasets/pull/4249
1,218,524,424
PR_kwDODunzps42742y
4,249
Support streaming XGLUE dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-04-28T10:27:23Z
2022-05-06T08:38:21Z
2022-04-28T16:08:03Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4249.diff", "html_url": "https://github.com/huggingface/datasets/pull/4249", "merged_at": "2022-04-28T16:08:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/4249.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4249" }
Support streaming XGLUE dataset. Fix #4247. CC: @severo
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4249/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4249/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/5873
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5873/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5873/comments
https://api.github.com/repos/huggingface/datasets/issues/5873/events
https://github.com/huggingface/datasets/issues/5873
1,713,269,724
I_kwDODunzps5mHmvc
5,873
Allow setting the environment variable for the lock file path
{ "avatar_url": "https://avatars.githubusercontent.com/u/83260933?v=4", "events_url": "https://api.github.com/users/xin3he/events{/privacy}", "followers_url": "https://api.github.com/users/xin3he/followers", "following_url": "https://api.github.com/users/xin3he/following{/other_user}", "gists_url": "https://api.github.com/users/xin3he/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/xin3he", "id": 83260933, "login": "xin3he", "node_id": "MDQ6VXNlcjgzMjYwOTMz", "organizations_url": "https://api.github.com/users/xin3he/orgs", "received_events_url": "https://api.github.com/users/xin3he/received_events", "repos_url": "https://api.github.com/users/xin3he/repos", "site_admin": false, "starred_url": "https://api.github.com/users/xin3he/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xin3he/subscriptions", "type": "User", "url": "https://api.github.com/users/xin3he" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[]
2023-05-17T07:10:02Z
2023-05-17T07:11:05Z
null
NONE
null
null
null
### Feature request Add an environment variable to replace the default lock file path. ### Motivation Usually, dataset path is a read-only path while the lock file needs to be modified each time. It would be convenient if the path can be reset individually. ### Your contribution ```/src/datasets/utils/filelock.py class UnixFileLock(BaseFileLock): def __init__(self, lock_file, timeout=-1, max_filename_length=None): #------------------- if os.getenv('DS_TMP_PATH'): file_name = str(lock_file).split('/')[-1] dataset_tmp_path = os.getenv('DS_TMP_PATH') lock_file = os.path.join(dataset_tmp_path, file_name) #------------------- max_filename_length = os.statvfs(os.path.dirname(lock_file)).f_namemax super().__init__(lock_file, timeout=timeout, max_filename_length=max_filename_length) ``` A simple demo is as upper. Thanks.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5873/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5873/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/5753
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5753/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5753/comments
https://api.github.com/repos/huggingface/datasets/issues/5753/events
https://github.com/huggingface/datasets/issues/5753
1,668,659,536
I_kwDODunzps5jdblQ
5,753
[IterableDatasets] Add column followed by interleave datasets gives bogus outputs
{ "avatar_url": "https://avatars.githubusercontent.com/u/93869735?v=4", "events_url": "https://api.github.com/users/sanchit-gandhi/events{/privacy}", "followers_url": "https://api.github.com/users/sanchit-gandhi/followers", "following_url": "https://api.github.com/users/sanchit-gandhi/following{/other_user}", "gists_url": "https://api.github.com/users/sanchit-gandhi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sanchit-gandhi", "id": 93869735, "login": "sanchit-gandhi", "node_id": "U_kgDOBZhWpw", "organizations_url": "https://api.github.com/users/sanchit-gandhi/orgs", "received_events_url": "https://api.github.com/users/sanchit-gandhi/received_events", "repos_url": "https://api.github.com/users/sanchit-gandhi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sanchit-gandhi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sanchit-gandhi/subscriptions", "type": "User", "url": "https://api.github.com/users/sanchit-gandhi" }
[]
closed
false
null
[]
null
[ "Problem with the code snippet! Using global vars and functions was not a good idea with iterable datasets!\r\n\r\nIf we update to:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\noriginal_dataset = load_dataset(\"librispeech_asr\", \"clean\", split=\"validation\", streaming=True)\r\n\r\n# now add a new co...
2023-04-14T17:32:31Z
2023-04-14T17:45:52Z
2023-04-14T17:36:37Z
CONTRIBUTOR
null
null
null
### Describe the bug If we add a new column to our iterable dataset using the hack described in #5752, when we then interleave datasets the new column is pinned to one value. ### Steps to reproduce the bug What we're going to do here is: 1. Load an iterable dataset in streaming mode (`original_dataset`) 2. Add a new column to this dataset using the hack in #5752 (`modified_dataset_1`) 3. Create another new dataset by adding a column with the same key but different values (`modified_dataset_2`) 4. Interleave our new datasets (`modified_dataset_1` + `modified_dataset_2`) 5. Check the value of our newly added column (`new_column`) ```python from datasets import load_dataset # load an iterable dataset original_dataset = load_dataset("librispeech_asr", "clean", split="validation", streaming=True) # now add a new column to our streaming dataset using our hack from 5752 name = "new_column" column = [f"new dataset 1, row {i}" for i in range(50)] new_features = original_dataset.features.copy() new_features[name] = new_features["file"] # I know that "file" has the right column type to match our new feature def add_column_fn(example, idx): if name in example: raise ValueError(f"Error when adding {name}: column {name} is already in the dataset.") return {name: column[idx]} modified_dataset_1 = original_dataset.map(add_column_fn, with_indices=True, features=new_features) # now create a second modified dataset using the same trick column = [f"new dataset 2, row {i}" for i in range(50)] def add_column_fn(example, idx): if name in example: raise ValueError(f"Error when adding {name}: column {name} is already in the dataset.") return {name: column[idx]} modified_dataset_2 = original_dataset.map(add_column_fn, with_indices=True, features=new_features) # interleave these datasets interleaved_dataset = interleave_datasets([modified_dataset_1, modified_dataset_2]) # now check what the value of the added column is for i, sample in enumerate(interleaved_dataset): print(sample["new_column"]) if i == 10: break ``` **Print Output:** ``` new dataset 2, row 0 new dataset 2, row 0 new dataset 2, row 1 new dataset 2, row 1 new dataset 2, row 2 new dataset 2, row 2 new dataset 2, row 3 new dataset 2, row 3 new dataset 2, row 4 new dataset 2, row 4 new dataset 2, row 5 ``` We see that we only get outputs from our second dataset. ### Expected behavior We should interleave between dataset 1 and 2 and increase in row value: ``` new dataset 1, row 0 new dataset 2, row 0 new dataset 1, row 1 new dataset 2, row 1 new dataset 1, row 2 new dataset 2, row 2 ... ``` ### Environment info - datasets version: 2.10.2.dev0 - Platform: Linux-4.19.0-23-cloud-amd64-x86_64-with-glibc2.28 - Python version: 3.9.16 - Huggingface_hub version: 0.13.3 - PyArrow version: 10.0.1 - Pandas version: 1.5.2
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5753/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5753/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/4991
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4991/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4991/comments
https://api.github.com/repos/huggingface/datasets/issues/4991/events
https://github.com/huggingface/datasets/pull/4991
1,378,898,752
PR_kwDODunzps4_P5hI
4,991
Fix missing tags in dataset cards
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-09-20T06:42:07Z
2022-09-22T12:25:32Z
2022-09-20T07:37:30Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4991.diff", "html_url": "https://github.com/huggingface/datasets/pull/4991", "merged_at": "2022-09-20T07:37:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/4991.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4991" }
Fix missing tags in dataset cards: - aeslc - empathetic_dialogues - event2Mind - gap - iwslt2017 - newsgroup - qa4mre - scicite This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833 - #4891 - #4896 - #4908 - #4921 - #4931 - #4979
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4991/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4991/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/5514
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5514/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5514/comments
https://api.github.com/repos/huggingface/datasets/issues/5514/events
https://github.com/huggingface/datasets/issues/5514
1,576,453,837
I_kwDODunzps5d9sbN
5,514
Improve inconsistency of `Dataset.map` interface for `load_from_cache_file`
{ "avatar_url": "https://avatars.githubusercontent.com/u/22773355?v=4", "events_url": "https://api.github.com/users/HallerPatrick/events{/privacy}", "followers_url": "https://api.github.com/users/HallerPatrick/followers", "following_url": "https://api.github.com/users/HallerPatrick/following{/other_user}", "gists_url": "https://api.github.com/users/HallerPatrick/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/HallerPatrick", "id": 22773355, "login": "HallerPatrick", "node_id": "MDQ6VXNlcjIyNzczMzU1", "organizations_url": "https://api.github.com/users/HallerPatrick/orgs", "received_events_url": "https://api.github.com/users/HallerPatrick/received_events", "repos_url": "https://api.github.com/users/HallerPatrick/repos", "site_admin": false, "starred_url": "https://api.github.com/users/HallerPatrick/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/HallerPatrick/subscriptions", "type": "User", "url": "https://api.github.com/users/HallerPatrick" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[ "Hi, thanks for noticing this! We can't just remove the cache control as this allows us to control where the arrow files generated by the ops are written (cached on disk if enabled or a temporary directory if disabled). The right way to address this inconsistency would be by having `load_from_cache_file=None` by de...
2023-02-08T16:40:44Z
2023-02-14T14:26:44Z
2023-02-14T14:26:44Z
CONTRIBUTOR
null
null
null
### Feature request 1. Replace the `load_from_cache_file` default value to `True`. 2. Remove or alter checks from `is_caching_enabled` logic. ### Motivation I stumbled over an inconsistency in the `Dataset.map` interface. The documentation (and source) states for the parameter `load_from_cache_file`: ``` load_from_cache_file (`bool`, defaults to `True` if caching is enabled): If a cache file storing the current computation from `function` can be identified, use it instead of recomputing. ``` 1. `load_from_cache_file` default value is `None`, while being annotated as `bool` 2. It is inconsistent with other method signatures like `filter`, that have the default value `True` 3. The logic is inconsistent, as the `map` method checks if caching is enabled through `is_caching_enabled`. This logic is not used for other similar methods. ### Your contribution I am not fully aware of the logic behind caching checks. If this is just a inconsistency that historically grew, I would suggest to remove the `is_caching_enabled` logic as the "default" logic. Maybe someone can give insights, if environment variables have a higher priority than local variables or vice versa. If this is clarified, I could adjust the source according to the "Feature request" section of this issue.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5514/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5514/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/2483
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2483/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2483/comments
https://api.github.com/repos/huggingface/datasets/issues/2483/events
https://github.com/huggingface/datasets/pull/2483
918,871,712
MDExOlB1bGxSZXF1ZXN0NjY4MjU1Mjg1
2,483
Use gc.collect only when needed to avoid slow downs
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
closed
false
null
[]
null
[ "I continue thinking that the origin of the issue has to do with tqdm (and not with Arrow): this issue only arises for version 4.50.0 (and later) of tqdm, not for previous versions of tqdm.\r\n\r\nMy guess is that tqdm made a change from version 4.50.0 that does not properly release the iterable. ", "FR" ]
2021-06-11T15:09:30Z
2021-06-18T19:25:06Z
2021-06-11T15:31:36Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/2483.diff", "html_url": "https://github.com/huggingface/datasets/pull/2483", "merged_at": "2021-06-11T15:31:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/2483.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2483" }
In https://github.com/huggingface/datasets/commit/42320a110d9d072703814e1f630a0d90d626a1e6 we added a call to gc.collect to resolve some issues on windows (see https://github.com/huggingface/datasets/pull/2482) However calling gc.collect too often causes significant slow downs (the CI run time doubled). So I just moved the gc.collect call to the exact place where it's actually needed: when post-processing a dataset
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2483/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2483/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3744
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3744/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3744/comments
https://api.github.com/repos/huggingface/datasets/issues/3744/events
https://github.com/huggingface/datasets/issues/3744
1,141,461,165
I_kwDODunzps5ECVCt
3,744
Better shards shuffling in streaming mode
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "fef2c0", "default": fals...
closed
false
null
[]
null
[]
2022-02-17T15:07:21Z
2022-02-23T15:00:58Z
2022-02-23T15:00:58Z
MEMBER
null
null
null
Sometimes a dataset script has a `_split_generators` that returns several files as well as the corresponding metadata of each file. It often happens that they end up in two separate lists in the `gen_kwargs`: ```python gen_kwargs = { "files": [os.path.join(data_dir, filename) for filename in all_files], "metadata_files": [all_metadata[filename] for filename in all_files], } ``` It happened for Multilingual Spoken Words for example in #3666 However currently **the two lists are shuffled independently** when shuffling the shards in streaming mode. This leads to `_generate_examples` not having the right metadata for each file. To prevent this issue I suggest that we always shuffle lists of the same length the exact same way to avoid such a big but silent issue. cc @polinaeterna
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3744/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3744/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/3974
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3974/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3974/comments
https://api.github.com/repos/huggingface/datasets/issues/3974/events
https://github.com/huggingface/datasets/pull/3974
1,174,485,044
PR_kwDODunzps40ssrA
3,974
Add XFUN dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/24835382?v=4", "events_url": "https://api.github.com/users/qqaatw/events{/privacy}", "followers_url": "https://api.github.com/users/qqaatw/followers", "following_url": "https://api.github.com/users/qqaatw/following{/other_user}", "gists_url": "https://api.github.com/users/qqaatw/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/qqaatw", "id": 24835382, "login": "qqaatw", "node_id": "MDQ6VXNlcjI0ODM1Mzgy", "organizations_url": "https://api.github.com/users/qqaatw/orgs", "received_events_url": "https://api.github.com/users/qqaatw/received_events", "repos_url": "https://api.github.com/users/qqaatw/repos", "site_admin": false, "starred_url": "https://api.github.com/users/qqaatw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/qqaatw/subscriptions", "type": "User", "url": "https://api.github.com/users/qqaatw" }
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "Not sure how to generate dummy data.\r\n\r\nThe downloaded file structure is \r\n\r\n- document file paths\r\n - (a json file containing all documents info, document images folder)\r\n - (a json file containing all documents in...
2022-03-20T09:24:54Z
2022-10-03T09:38:16Z
2022-10-03T09:36:22Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3974.diff", "html_url": "https://github.com/huggingface/datasets/pull/3974", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/3974.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3974" }
This PR adds XFUN dataset. Home page and repository: https://github.com/doc-analysis/XFUND Source code: https://github.com/microsoft/unilm/blob/master/layoutlmft/layoutlmft/data/datasets/xfun.py
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3974/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3974/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/5517
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5517/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5517/comments
https://api.github.com/repos/huggingface/datasets/issues/5517/events
https://github.com/huggingface/datasets/issues/5517
1,577,976,608
I_kwDODunzps5eDgMg
5,517
`with_format("numpy")` silently downcasts float64 to float32 features
{ "avatar_url": "https://avatars.githubusercontent.com/u/1250234?v=4", "events_url": "https://api.github.com/users/ernestum/events{/privacy}", "followers_url": "https://api.github.com/users/ernestum/followers", "following_url": "https://api.github.com/users/ernestum/following{/other_user}", "gists_url": "https://api.github.com/users/ernestum/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ernestum", "id": 1250234, "login": "ernestum", "node_id": "MDQ6VXNlcjEyNTAyMzQ=", "organizations_url": "https://api.github.com/users/ernestum/orgs", "received_events_url": "https://api.github.com/users/ernestum/received_events", "repos_url": "https://api.github.com/users/ernestum/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ernestum/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ernestum/subscriptions", "type": "User", "url": "https://api.github.com/users/ernestum" }
[]
open
false
null
[]
{ "closed_at": null, "closed_issues": 0, "created_at": "2023-02-13T16:22:42Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }, "description": "Next major release", "due_on": null, "html_url": "https://github.com/huggingface/datasets/milestone/10", "id": 9038583, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/10/labels", "node_id": "MI_kwDODunzps4Aier3", "number": 10, "open_issues": 4, "state": "open", "title": "3.0", "updated_at": "2023-09-22T14:07:52Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/10" }
[ "Hi! This behavior stems from these lines:\r\n\r\nhttps://github.com/huggingface/datasets/blob/b065547654efa0ec633cf373ac1512884c68b2e1/src/datasets/formatting/np_formatter.py#L45-L46\r\n\r\nI agree we should preserve the original type whenever possible and downcast explicitly with a warning.\r\n\r\n@lhoestq Do you...
2023-02-09T14:18:00Z
2023-02-14T15:38:54Z
null
NONE
null
null
null
### Describe the bug When I create a dataset with a `float64` feature, then apply numpy formatting the returned numpy arrays are silently downcasted to `float32`. ### Steps to reproduce the bug ```python import datasets dataset = datasets.Dataset.from_dict({'a': [1.0, 2.0, 3.0]}).with_format("numpy") print("feature dtype:", dataset.features['a'].dtype) print("array dtype:", dataset['a'].dtype) ``` output: ``` feature dtype: float64 array dtype: float32 ``` ### Expected behavior ``` feature dtype: float64 array dtype: float64 ``` ### Environment info - `datasets` version: 2.8.0 - Platform: Linux-5.4.0-135-generic-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 10.0.1 - Pandas version: 1.4.4 ### Suggested Fix Changing [the `_tensorize` function of the numpy formatter](https://github.com/huggingface/datasets/blob/b065547654efa0ec633cf373ac1512884c68b2e1/src/datasets/formatting/np_formatter.py#L32) to ```python def _tensorize(self, value): if isinstance(value, (str, bytes, type(None))): return value elif isinstance(value, (np.character, np.ndarray)) and np.issubdtype(value.dtype, np.character): return value elif isinstance(value, np.number): return value return np.asarray(value, **self.np_array_kwargs) ``` fixes this particular issue for me. Not sure if this would break other tests. This should also avoid unnecessary copying of the array.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5517/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5517/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2623
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2623/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2623/comments
https://api.github.com/repos/huggingface/datasets/issues/2623/events
https://github.com/huggingface/datasets/pull/2623
941,265,342
MDExOlB1bGxSZXF1ZXN0Njg3MTk0MjM3
2,623
[Metrics] added wiki_split metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/26653468?v=4", "events_url": "https://api.github.com/users/bhadreshpsavani/events{/privacy}", "followers_url": "https://api.github.com/users/bhadreshpsavani/followers", "following_url": "https://api.github.com/users/bhadreshpsavani/following{/other_user}", "gists_url": "https://api.github.com/users/bhadreshpsavani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bhadreshpsavani", "id": 26653468, "login": "bhadreshpsavani", "node_id": "MDQ6VXNlcjI2NjUzNDY4", "organizations_url": "https://api.github.com/users/bhadreshpsavani/orgs", "received_events_url": "https://api.github.com/users/bhadreshpsavani/received_events", "repos_url": "https://api.github.com/users/bhadreshpsavani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bhadreshpsavani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhadreshpsavani/subscriptions", "type": "User", "url": "https://api.github.com/users/bhadreshpsavani" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/followin...
null
[ "Looks all good to me thanks :)\r\nJust did some minor corrections in the docstring" ]
2021-07-10T14:51:50Z
2021-07-14T14:28:13Z
2021-07-12T22:34:31Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/2623.diff", "html_url": "https://github.com/huggingface/datasets/pull/2623", "merged_at": "2021-07-12T22:34:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/2623.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2623" }
Fixes: #2606 This pull request adds combine metrics for the wikisplit or English sentence split task Reviewer: @patrickvonplaten
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2623/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2623/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3570
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3570/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3570/comments
https://api.github.com/repos/huggingface/datasets/issues/3570/events
https://github.com/huggingface/datasets/pull/3570
1,100,480,791
PR_kwDODunzps4w3Xez
3,570
Add the KMWP dataset (extension of #3564)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42150335?v=4", "events_url": "https://api.github.com/users/sooftware/events{/privacy}", "followers_url": "https://api.github.com/users/sooftware/followers", "following_url": "https://api.github.com/users/sooftware/following{/other_user}", "gists_url": "https://api.github.com/users/sooftware/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sooftware", "id": 42150335, "login": "sooftware", "node_id": "MDQ6VXNlcjQyMTUwMzM1", "organizations_url": "https://api.github.com/users/sooftware/orgs", "received_events_url": "https://api.github.com/users/sooftware/received_events", "repos_url": "https://api.github.com/users/sooftware/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sooftware/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sooftware/subscriptions", "type": "User", "url": "https://api.github.com/users/sooftware" }
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
closed
false
null
[]
null
[ "Sorry, I'm late to check! I'll send it to you soon!", "Thanks for your contribution, @sooftware. Are you still interested in adding this dataset?\r\n\r\nWe are removing the dataset scripts from this GitHub repo and moving them to the Hugging Face Hub: https://huggingface.co/datasets\r\n\r\nWe would suggest you c...
2022-01-12T15:33:08Z
2022-10-01T06:43:16Z
2022-10-01T06:43:16Z
NONE
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3570.diff", "html_url": "https://github.com/huggingface/datasets/pull/3570", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/3570.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3570" }
New pull request of #3564 (Add the KMWP dataset)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3570/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3570/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4440
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4440/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4440/comments
https://api.github.com/repos/huggingface/datasets/issues/4440/events
https://github.com/huggingface/datasets/pull/4440
1,258,494,469
PR_kwDODunzps44_io_
4,440
Update docs around audio and vision
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[ { "color": "0075ca", "default": true, "description": "Improvements or additions to documentation", "id": 1935892861, "name": "documentation", "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation" } ]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "> Let me know what you think, especially if we should include some code samples for training a model in the audio/vision sections. I left this out since we already showed it in the NLP section. I want to keep the focus on using Datas...
2022-06-02T17:42:03Z
2022-06-23T16:33:19Z
2022-06-23T16:23:02Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4440.diff", "html_url": "https://github.com/huggingface/datasets/pull/4440", "merged_at": "2022-06-23T16:23:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/4440.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4440" }
As part of the strategy to center the docs around the different modalities, this PR updates the quickstart to include audio and vision examples. This improves the developer experience by making audio and vision content more discoverable, enabling users working in these modalities to also quickly get started without digging too deeply into the docs. Other changes include: - Moved the installation guide to the Get Started section because it should be part of a user's onboarding to the library before exploring tutorials or how-to's. - Updated the native TF code at creating a `tf.data.Dataset` because it was throwing an error. The `to_tensor()` bit was redundant and removing it fixed the error (please double-check me here!). - Added some UI components to the quickstart so it's easier for users to navigate directly to the relevant section with context about what to expect. - Reverted to the code tabs for content that don't have any framework-specific text. I think this saves space compared to the code blocks. We'll still use the code blocks if the `torch` text is different from the `tf` text. Let me know what you think, especially if we should include some code samples for training a model in the audio/vision sections. I left this out since we already showed it in the NLP section. I want to keep the focus on using Datasets to load and process a dataset, and not so much the training part. Maybe we can add links to the Transformers docs instead?
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4440/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4440/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/1722
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1722/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1722/comments
https://api.github.com/repos/huggingface/datasets/issues/1722/events
https://github.com/huggingface/datasets/pull/1722
783,921,679
MDExOlB1bGxSZXF1ZXN0NTUzMTk3MTg4
1,722
Added unfiltered versions of the Wiki-Auto training data for the GEM simplification task.
{ "avatar_url": "https://avatars.githubusercontent.com/u/11708999?v=4", "events_url": "https://api.github.com/users/mounicam/events{/privacy}", "followers_url": "https://api.github.com/users/mounicam/followers", "following_url": "https://api.github.com/users/mounicam/following{/other_user}", "gists_url": "https://api.github.com/users/mounicam/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mounicam", "id": 11708999, "login": "mounicam", "node_id": "MDQ6VXNlcjExNzA4OTk5", "organizations_url": "https://api.github.com/users/mounicam/orgs", "received_events_url": "https://api.github.com/users/mounicam/received_events", "repos_url": "https://api.github.com/users/mounicam/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mounicam/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mounicam/subscriptions", "type": "User", "url": "https://api.github.com/users/mounicam" }
[]
closed
false
null
[]
null
[ "The current version of Wiki-Auto dataset contains a filtered version of the aligned dataset. The commit adds unfiltered versions of the data that can be useful the GEM task participants." ]
2021-01-12T05:26:04Z
2021-01-12T18:14:53Z
2021-01-12T17:35:57Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/1722.diff", "html_url": "https://github.com/huggingface/datasets/pull/1722", "merged_at": "2021-01-12T17:35:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/1722.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1722" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1722/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1722/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3654
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3654/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3654/comments
https://api.github.com/repos/huggingface/datasets/issues/3654/events
https://github.com/huggingface/datasets/pull/3654
1,119,717,475
PR_kwDODunzps4x2kiX
3,654
Better TQDM output
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[ "@lhoestq I've created a notebook for you to see the difference: https://colab.research.google.com/drive/1by3EqnoKvC2p-yKW4lPDGOFOZHyGVyeQ?usp=sharing.\r\n\r\nFeel free to suggest better descriptions for the progress bars. \r\n\r\nIf everything looks good, think we can merge." ]
2022-01-31T17:22:43Z
2022-02-03T15:55:34Z
2022-02-03T15:55:33Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3654.diff", "html_url": "https://github.com/huggingface/datasets/pull/3654", "merged_at": "2022-02-03T15:55:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/3654.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3654" }
This PR does the following: * if `dataset_infos.json` exists for a dataset, uses `num_examples` to print the total number of examples that needs to be generated (in `builder.py`) * fixes `tqdm` + multiprocessing in Jupyter Notebook/Colab (the issue stems from this commit in the `tqdm` repo: https://github.com/tqdm/tqdm/commit/f7722edecc3010cb35cc1c923ac4850a76336f82) * adds the missing `drop_last_batch` and `with_ranks` params to `DatasetDict.map` * correctly computes the number of iterations in `map` and the CSV/JSON loader when `batched=True` to fix `tqdm` progress bars * removes the `bool(logging.get_verbosity() == logging.NOTSET)` (or simplifies `bool(logging.get_verbosity() == logging.NOTSET) or not utils.is_progress_bar_enabled()` to `not utils.is_progress_bar_enabled()`) condition and uses `utils.is_progress_bar_enabled` to check if `tqdm` output is enabled (this comment from @stas00 explains why the `bool(logging.get_verbosity() == logging.NOTSET)` check is problematic: https://github.com/huggingface/transformers/issues/14889#issue-1087318463) Fix #2630
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3654/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3654/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/5242
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5242/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5242/comments
https://api.github.com/repos/huggingface/datasets/issues/5242/events
https://github.com/huggingface/datasets/issues/5242
1,449,069,382
I_kwDODunzps5WXwtG
5,242
Failed Data Processing upon upload with zip file full of images
{ "avatar_url": "https://avatars.githubusercontent.com/u/82735473?v=4", "events_url": "https://api.github.com/users/scrambled2/events{/privacy}", "followers_url": "https://api.github.com/users/scrambled2/followers", "following_url": "https://api.github.com/users/scrambled2/following{/other_user}", "gists_url": "https://api.github.com/users/scrambled2/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/scrambled2", "id": 82735473, "login": "scrambled2", "node_id": "MDQ6VXNlcjgyNzM1NDcz", "organizations_url": "https://api.github.com/users/scrambled2/orgs", "received_events_url": "https://api.github.com/users/scrambled2/received_events", "repos_url": "https://api.github.com/users/scrambled2/repos", "site_admin": false, "starred_url": "https://api.github.com/users/scrambled2/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/scrambled2/subscriptions", "type": "User", "url": "https://api.github.com/users/scrambled2" }
[]
open
false
null
[]
null
[ "cc @abhishekkrthakur @SBrandeis " ]
2022-11-15T02:47:52Z
2022-11-15T17:59:23Z
null
NONE
null
null
null
I went to autotrain and under image classification arrived where it was time to prepare my dataset. Screenshot below ![image](https://user-images.githubusercontent.com/82735473/201814099-3cc5ff8a-88dc-4f5f-8140-f19560641d83.png) I chose the method 2 option. I have a csv file with two columns. ~23,000 files. I uploaded this and chose the image_relpath, and target columns. The image uploader said that I could only upload 10,000 singular images at a time so the 2nd option was to zip the images up and upload a zip archive which I did. That all uploaded. Now I have the message below. It appears the zip archive does just uncompress on the Hugging Face end? What am I missing here? ![image](https://user-images.githubusercontent.com/82735473/201813838-b50dbbbc-34e8-4d73-9c07-12f9e41c62eb.png)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5242/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5242/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/6230
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6230/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6230/comments
https://api.github.com/repos/huggingface/datasets/issues/6230/events
https://github.com/huggingface/datasets/pull/6230
1,890,521,006
PR_kwDODunzps5aBh6L
6,230
Don't skip hidden files in `dl_manager.iter_files` when they are given as input
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
2023-09-11T13:29:19Z
2023-09-13T18:21:28Z
2023-09-13T18:12:09Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6230.diff", "html_url": "https://github.com/huggingface/datasets/pull/6230", "merged_at": "2023-09-13T18:12:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/6230.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6230" }
Required for `load_dataset(<format>, data_files=["path/to/.hidden_file"])` to work as expected
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6230/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6230/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3719
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3719/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3719/comments
https://api.github.com/repos/huggingface/datasets/issues/3719/events
https://github.com/huggingface/datasets/pull/3719
1,137,237,622
PR_kwDODunzps4yyFv7
3,719
Check if indices values in `Dataset.select` are within bounds
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[]
2022-02-14T12:31:41Z
2022-02-14T19:19:22Z
2022-02-14T19:19:22Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3719.diff", "html_url": "https://github.com/huggingface/datasets/pull/3719", "merged_at": "2022-02-14T19:19:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/3719.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3719" }
Fix #3707 Instead of reusing `_check_valid_index_key` from `datasets.formatting`, I defined a new function to provide a more meaningful error message.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/3719/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3719/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3318
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3318/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3318/comments
https://api.github.com/repos/huggingface/datasets/issues/3318/events
https://github.com/huggingface/datasets/pull/3318
1,062,369,717
PR_kwDODunzps4u9m-k
3,318
Finish transition to PyArrow 3.0.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
closed
false
null
[]
null
[]
2021-11-24T12:30:14Z
2021-11-24T15:35:05Z
2021-11-24T15:35:04Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3318.diff", "html_url": "https://github.com/huggingface/datasets/pull/3318", "merged_at": "2021-11-24T15:35:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/3318.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3318" }
Finish transition to PyArrow 3.0.0 that was started in #3098.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3318/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3318/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3526
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3526/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3526/comments
https://api.github.com/repos/huggingface/datasets/issues/3526/events
https://github.com/huggingface/datasets/pull/3526
1,093,833,446
PR_kwDODunzps4wiMaQ
3,526
Update license to bookcorpus dataset card
{ "avatar_url": "https://avatars.githubusercontent.com/u/90473723?v=4", "events_url": "https://api.github.com/users/meg-huggingface/events{/privacy}", "followers_url": "https://api.github.com/users/meg-huggingface/followers", "following_url": "https://api.github.com/users/meg-huggingface/following{/other_user}", "gists_url": "https://api.github.com/users/meg-huggingface/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/meg-huggingface", "id": 90473723, "login": "meg-huggingface", "node_id": "MDQ6VXNlcjkwNDczNzIz", "organizations_url": "https://api.github.com/users/meg-huggingface/orgs", "received_events_url": "https://api.github.com/users/meg-huggingface/received_events", "repos_url": "https://api.github.com/users/meg-huggingface/repos", "site_admin": false, "starred_url": "https://api.github.com/users/meg-huggingface/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/meg-huggingface/subscriptions", "type": "User", "url": "https://api.github.com/users/meg-huggingface" }
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
closed
false
null
[]
null
[ "The smashwords ToS apply for this dataset, we did the same for https://github.com/huggingface/datasets/pull/3525", "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-01-04T23:25:23Z
2022-09-30T10:23:38Z
2022-09-30T10:21:20Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3526.diff", "html_url": "https://github.com/huggingface/datasets/pull/3526", "merged_at": "2022-09-30T10:21:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/3526.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3526" }
Not entirely sure, following the links here, but it seems the relevant license is at https://github.com/soskek/bookcorpus/blob/master/LICENSE
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3526/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3526/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/1627
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1627/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1627/comments
https://api.github.com/repos/huggingface/datasets/issues/1627/events
https://github.com/huggingface/datasets/issues/1627
773,960,255
MDU6SXNzdWU3NzM5NjAyNTU=
1,627
`Dataset.map` disable progress bar
{ "avatar_url": "https://avatars.githubusercontent.com/u/8767964?v=4", "events_url": "https://api.github.com/users/Nickil21/events{/privacy}", "followers_url": "https://api.github.com/users/Nickil21/followers", "following_url": "https://api.github.com/users/Nickil21/following{/other_user}", "gists_url": "https://api.github.com/users/Nickil21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Nickil21", "id": 8767964, "login": "Nickil21", "node_id": "MDQ6VXNlcjg3Njc5NjQ=", "organizations_url": "https://api.github.com/users/Nickil21/orgs", "received_events_url": "https://api.github.com/users/Nickil21/received_events", "repos_url": "https://api.github.com/users/Nickil21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Nickil21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Nickil21/subscriptions", "type": "User", "url": "https://api.github.com/users/Nickil21" }
[]
closed
false
null
[]
null
[ "Progress bar can be disabled like this:\r\n```python\r\nfrom datasets.utils.logging import set_verbosity_error\r\nset_verbosity_error()\r\n```\r\n\r\nThere is this line in `Dataset.map`:\r\n```python\r\nnot_verbose = bool(logger.getEffectiveLevel() > WARNING)\r\n```\r\n\r\nSo any logging level higher than `WARNING...
2020-12-23T17:53:42Z
2023-02-08T02:37:47Z
2020-12-26T19:57:17Z
NONE
null
null
null
I can't find anything to turn off the `tqdm` progress bars while running a preprocessing function using `Dataset.map`. I want to do akin to `disable_tqdm=True` in the case of `transformers`. Is there something like that?
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/1627/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1627/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/4735
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4735/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4735/comments
https://api.github.com/repos/huggingface/datasets/issues/4735/events
https://github.com/huggingface/datasets/pull/4735
1,314,501,641
PR_kwDODunzps477CuP
4,735
Pin rouge_score test dependency
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-07-22T07:18:21Z
2022-07-22T07:58:14Z
2022-07-22T07:45:18Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4735.diff", "html_url": "https://github.com/huggingface/datasets/pull/4735", "merged_at": "2022-07-22T07:45:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/4735.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4735" }
Temporarily pin `rouge_score` (to avoid latest version 0.7.0) until the issue is fixed. Fix #4734
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4735/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4735/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/5651
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5651/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5651/comments
https://api.github.com/repos/huggingface/datasets/issues/5651/events
https://github.com/huggingface/datasets/issues/5651
1,631,967,509
I_kwDODunzps5hRdkV
5,651
expanduser in save_to_disk
{ "avatar_url": "https://avatars.githubusercontent.com/u/42400165?v=4", "events_url": "https://api.github.com/users/RmZeta2718/events{/privacy}", "followers_url": "https://api.github.com/users/RmZeta2718/followers", "following_url": "https://api.github.com/users/RmZeta2718/following{/other_user}", "gists_url": "https://api.github.com/users/RmZeta2718/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/RmZeta2718", "id": 42400165, "login": "RmZeta2718", "node_id": "MDQ6VXNlcjQyNDAwMTY1", "organizations_url": "https://api.github.com/users/RmZeta2718/orgs", "received_events_url": "https://api.github.com/users/RmZeta2718/received_events", "repos_url": "https://api.github.com/users/RmZeta2718/repos", "site_admin": false, "starred_url": "https://api.github.com/users/RmZeta2718/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/RmZeta2718/subscriptions", "type": "User", "url": "https://api.github.com/users/RmZeta2718" }
[ { "color": "7057ff", "default": true, "description": "Good for newcomers", "id": 1935892877, "name": "good first issue", "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue" } ]
open
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/35114142?v=4", "events_url": "https://api.github.com/users/benjaminbrown038/events{/privacy}", "followers_url": "https://api.github.com/users/benjaminbrown038/followers", "following_url": "https://api.github.com/users/benjaminbrown038/following{/other_user}", "gists_url": "https://api.github.com/users/benjaminbrown038/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/benjaminbrown038", "id": 35114142, "login": "benjaminbrown038", "node_id": "MDQ6VXNlcjM1MTE0MTQy", "organizations_url": "https://api.github.com/users/benjaminbrown038/orgs", "received_events_url": "https://api.github.com/users/benjaminbrown038/received_events", "repos_url": "https://api.github.com/users/benjaminbrown038/repos", "site_admin": false, "starred_url": "https://api.github.com/users/benjaminbrown038/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/benjaminbrown038/subscriptions", "type": "User", "url": "https://api.github.com/users/benjaminbrown038" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/35114142?v=4", "events_url": "https://api.github.com/users/benjaminbrown038/events{/privacy}", "followers_url": "https://api.github.com/users/benjaminbrown038/followers", "following_url": "https://api.github.com/users/benjaminbrown038/followin...
null
[ "`save_to_disk` should indeed expand `~`. Marking it as a \"good first issue\".", "#self-assign\r\n\r\nFile path to code: \r\n\r\nhttps://github.com/huggingface/datasets/blob/2.13.0/src/datasets/arrow_dataset.py#L1364\r\n\r\n@RmZeta2718 I created a pull request for this issue. ", "Hello, \r\nIt says `save_to_di...
2023-03-20T12:02:18Z
2023-07-26T16:18:06Z
null
NONE
null
null
null
### Describe the bug save_to_disk() does not expand `~` 1. `dataset = load_datasets("any dataset")` 2. `dataset.save_to_disk("~/data")` 3. a folder named "~" created in current folder 4. FileNotFoundError is raised, because the expanded path does not exist (`/home/<user>/data`) related issue https://github.com/huggingface/transformers/issues/10628 ### Steps to reproduce the bug As described above. ### Expected behavior expanduser correctly ### Environment info - datasets 2.10.1 - python 3.10
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5651/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5651/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/4919
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4919/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4919/comments
https://api.github.com/repos/huggingface/datasets/issues/4919/events
https://github.com/huggingface/datasets/pull/4919
1,357,441,599
PR_kwDODunzps4-IxDZ
4,919
feat: improve error message on Keys mismatch. closes #4917
{ "avatar_url": "https://avatars.githubusercontent.com/u/25532159?v=4", "events_url": "https://api.github.com/users/PaulLerner/events{/privacy}", "followers_url": "https://api.github.com/users/PaulLerner/followers", "following_url": "https://api.github.com/users/PaulLerner/following{/other_user}", "gists_url": "https://api.github.com/users/PaulLerner/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/PaulLerner", "id": 25532159, "login": "PaulLerner", "node_id": "MDQ6VXNlcjI1NTMyMTU5", "organizations_url": "https://api.github.com/users/PaulLerner/orgs", "received_events_url": "https://api.github.com/users/PaulLerner/received_events", "repos_url": "https://api.github.com/users/PaulLerner/repos", "site_admin": false, "starred_url": "https://api.github.com/users/PaulLerner/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/PaulLerner/subscriptions", "type": "User", "url": "https://api.github.com/users/PaulLerner" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._", "We are having an unrelated issue that makes several tests fail. We are working on that. Once fixed, you will be able to merge the main branch into this, so that you get the fix and the tests pass..." ]
2022-08-31T14:41:36Z
2022-09-05T08:46:01Z
2022-09-05T08:43:33Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4919.diff", "html_url": "https://github.com/huggingface/datasets/pull/4919", "merged_at": "2022-09-05T08:43:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/4919.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4919" }
Hi @lhoestq what do you think? Let me give you a code sample: ```py >>> import datasets >>> foo = datasets.Dataset.from_dict({'foo':[0,1], 'bar':[2,3]}) >>> foo.save_to_disk('foo') # edit foo/dataset_info.json e.g. rename the 'foo' feature to 'baz' >>> datasets.load_from_disk('foo') --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-2-4863e606b330> in <module> ----> 1 datasets.load_from_disk('foo') ~/code/datasets/src/datasets/load.py in load_from_disk(dataset_path, fs, keep_in_memory) 1851 raise FileNotFoundError(f"Directory {dataset_path} not found") 1852 if fs.isfile(Path(dest_dataset_path, config.DATASET_INFO_FILENAME).as_posix()): -> 1853 return Dataset.load_from_disk(dataset_path, fs, keep_in_memory=keep_in_memory) 1854 elif fs.isfile(Path(dest_dataset_path, config.DATASETDICT_JSON_FILENAME).as_posix()): 1855 return DatasetDict.load_from_disk(dataset_path, fs, keep_in_memory=keep_in_memory) ~/code/datasets/src/datasets/arrow_dataset.py in load_from_disk(dataset_path, fs, keep_in_memory) 1230 info=dataset_info, 1231 split=split, -> 1232 fingerprint=state["_fingerprint"], 1233 ) 1234 ~/code/datasets/src/datasets/arrow_dataset.py in __init__(self, arrow_table, info, split, indices_table, fingerprint) 687 self.info.features = inferred_features 688 else: # make sure the nested columns are in the right order --> 689 self.info.features = self.info.features.reorder_fields_as(inferred_features) 690 691 # Infer fingerprint if None ~/code/datasets/src/datasets/features/features.py in reorder_fields_as(self, other) 1771 return source 1772 -> 1773 return Features(recursive_reorder(self, other)) 1774 1775 def flatten(self, max_depth=16) -> "Features": ~/code/datasets/src/datasets/features/features.py in recursive_reorder(source, target, stack) 1760 f"{source.keys()-target.keys()} are missing from dataset.arrow " 1761 f"and {target.keys()-source.keys()} are missing from dataset_info.json"+stack_position) -> 1762 raise ValueError(message) 1763 return {key: recursive_reorder(source[key], target[key], stack + f".{key}") for key in target} 1764 elif isinstance(source, list): ValueError: Keys mismatch: between {'baz': Value(dtype='int64', id=None), 'bar': Value(dtype='int64', id=None)} (dataset_info.json) and {'foo': Value(dtype='int64', id=None), 'bar': Value(dtype='int64', id=None)} (inferred from dataset.arrow). {'baz'} are missing from dataset.arrow and {'foo'} are missing from dataset_info.json ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4919/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4919/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4833
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4833/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4833/comments
https://api.github.com/repos/huggingface/datasets/issues/4833/events
https://github.com/huggingface/datasets/pull/4833
1,336,946,965
PR_kwDODunzps49E_Nk
4,833
Fix missing tags in dataset cards
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-08-12T09:04:52Z
2022-09-22T14:41:23Z
2022-08-12T09:45:55Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4833.diff", "html_url": "https://github.com/huggingface/datasets/pull/4833", "merged_at": "2022-08-12T09:45:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/4833.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4833" }
Fix missing tags in dataset cards: - boolq - break_data - definite_pronoun_resolution - emo - kor_nli - pg19 - quartz - sciq - squad_es - wmt14 - wmt15 - wmt16 - wmt17 - wmt18 - wmt19 - wmt_t2t This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4833/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4833/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/5516
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5516/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5516/comments
https://api.github.com/repos/huggingface/datasets/issues/5516/events
https://github.com/huggingface/datasets/pull/5516
1,577,661,640
PR_kwDODunzps5JmzPQ
5,516
Reload features from Parquet metadata
{ "avatar_url": "https://avatars.githubusercontent.com/u/6368040?v=4", "events_url": "https://api.github.com/users/MFreidank/events{/privacy}", "followers_url": "https://api.github.com/users/MFreidank/followers", "following_url": "https://api.github.com/users/MFreidank/following{/other_user}", "gists_url": "https://api.github.com/users/MFreidank/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/MFreidank", "id": 6368040, "login": "MFreidank", "node_id": "MDQ6VXNlcjYzNjgwNDA=", "organizations_url": "https://api.github.com/users/MFreidank/orgs", "received_events_url": "https://api.github.com/users/MFreidank/received_events", "repos_url": "https://api.github.com/users/MFreidank/repos", "site_admin": false, "starred_url": "https://api.github.com/users/MFreidank/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MFreidank/subscriptions", "type": "User", "url": "https://api.github.com/users/MFreidank" }
[]
closed
false
null
[]
null
[ "Thanks a lot for your help @lhoestq. I've simplified what turned out to be a simple fix and added the unit test.\r\n\r\nDoes this look ready to be merged or is there anything I'm still missing?", "Cool ! I think you just need to remove the unused import in `io/parquet.py`\r\n```\r\nsrc/datasets/io/parquet.py:4:1...
2023-02-09T10:52:15Z
2023-02-12T16:00:00Z
2023-02-12T15:57:01Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/5516.diff", "html_url": "https://github.com/huggingface/datasets/pull/5516", "merged_at": "2023-02-12T15:57:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/5516.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5516" }
Resolves #5482. Attaches feature metadata to parquet files serialised using `Dataset.to_parquet`. This allows retrieving data with "rich" feature types (e.g., `datasets.features.image.Image` or `datasets.features.audio.Audio`) from parquet files without cumbersome casting (for an example, see #5482). @lhoestq It seems that it is sufficient to attach metadata to the schema prior to serialising and features are loaded back with correct types afterwards automatically. I used the following script to test the implementation: ```python from pathlib import Path import datasets dataset_name = "Maysee/tiny-imagenet" ds = datasets.load_dataset(dataset_name, split=datasets.Split.TRAIN) output_directory_path = Path(__file__).parent.joinpath("example_test_outputs", dataset_name.replace("/", "_")) output_directory_path.mkdir(exist_ok=True, parents=True) output_filepath = output_directory_path.joinpath("ds.parquet") ds.to_parquet(str(output_filepath)) reloaded_ds = datasets.load_dataset(str(output_directory_path), split=datasets.Split.TRAIN) assert ds.features == reloaded_ds.features ``` Prior to the change in this PR this script raises an `AssertionError` and the `Image` features lose their type after serialisation. After the change in this PR, the assertion does not raise an error and manual inspection of the features shows type `Image` for the respective columns of `reloaded_ds `. Some open questions: * How/where can I best add new unit tests for this implementation? * What dataset would I best use in the tests? I chose `Maysee/tiny-imagenet` mainly because it is small and contains an ?Image` feature that can be used to test, but I'd be happy for suggestions on a suitable data source to use. * Currently I'm calling `datasets.arrow_writer.ArrowWriter._build_metadata` as I need the same logic. However, I'm not happy with the coupling between `datasets.io.parquet` and `datasets.arrow_writer` it leaves me with. Suggest to factor this common logic out into a helper function and reuse it from both of these. Do you agree and if yes, could you please guide me where I would best place this function? Many thanks in advance and kind regards, MFreidank
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5516/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5516/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2554
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2554/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2554/comments
https://api.github.com/repos/huggingface/datasets/issues/2554/events
https://github.com/huggingface/datasets/issues/2554
931,453,855
MDU6SXNzdWU5MzE0NTM4NTU=
2,554
Multilabel metrics not supported
{ "avatar_url": "https://avatars.githubusercontent.com/u/37592763?v=4", "events_url": "https://api.github.com/users/GuillemGSubies/events{/privacy}", "followers_url": "https://api.github.com/users/GuillemGSubies/followers", "following_url": "https://api.github.com/users/GuillemGSubies/following{/other_user}", "gists_url": "https://api.github.com/users/GuillemGSubies/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/GuillemGSubies", "id": 37592763, "login": "GuillemGSubies", "node_id": "MDQ6VXNlcjM3NTkyNzYz", "organizations_url": "https://api.github.com/users/GuillemGSubies/orgs", "received_events_url": "https://api.github.com/users/GuillemGSubies/received_events", "repos_url": "https://api.github.com/users/GuillemGSubies/repos", "site_admin": false, "starred_url": "https://api.github.com/users/GuillemGSubies/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/GuillemGSubies/subscriptions", "type": "User", "url": "https://api.github.com/users/GuillemGSubies" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[ "Hi @GuillemGSubies, thanks for reporting.\r\n\r\nI have made a PR to fix this issue and allow metrics to be computed also for multilabel classification problems.", "Looks nice, thank you very much! 🚀 ", "Sorry for reopening but I just noticed that the `_compute` method for the F1 metric is still not good enou...
2021-06-28T11:09:46Z
2021-10-13T12:29:13Z
2021-07-08T08:40:15Z
NONE
null
null
null
When I try to use a metric like F1 macro I get the following error: ``` TypeError: int() argument must be a string, a bytes-like object or a number, not 'list' ``` There is an explicit casting here: https://github.com/huggingface/datasets/blob/fc79f61cbbcfa0e8c68b28c0a8257f17e768a075/src/datasets/features.py#L274 And looks like this is because here https://github.com/huggingface/datasets/blob/fc79f61cbbcfa0e8c68b28c0a8257f17e768a075/metrics/f1/f1.py#L88 the features can only be integers, so we cannot use that F1 for multilabel. Instead, if I create the following F1 (ints replaced with sequence of ints), it will work: ```python class F1(datasets.Metric): def _info(self): return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("int32")), "references": datasets.Sequence(datasets.Value("int32")), } ), reference_urls=["https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html"], ) def _compute(self, predictions, references, labels=None, pos_label=1, average="binary", sample_weight=None): return { "f1": f1_score( references, predictions, labels=labels, pos_label=pos_label, average=average, sample_weight=sample_weight, ), } ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2554/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2554/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/4767
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4767/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4767/comments
https://api.github.com/repos/huggingface/datasets/issues/4767/events
https://github.com/huggingface/datasets/pull/4767
1,321,843,538
PR_kwDODunzps48TCpI
4,767
Add 2.4.0 version added to docstrings
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
closed
false
null
[]
null
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
2022-07-29T07:01:56Z
2022-07-29T11:16:49Z
2022-07-29T11:03:58Z
MEMBER
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/4767.diff", "html_url": "https://github.com/huggingface/datasets/pull/4767", "merged_at": "2022-07-29T11:03:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/4767.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4767" }
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4767/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4767/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/3520
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3520/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3520/comments
https://api.github.com/repos/huggingface/datasets/issues/3520/events
https://github.com/huggingface/datasets/pull/3520
1,093,747,753
PR_kwDODunzps4wh6oD
3,520
Audio datacard update - first pass
{ "avatar_url": "https://avatars.githubusercontent.com/u/90473723?v=4", "events_url": "https://api.github.com/users/meg-huggingface/events{/privacy}", "followers_url": "https://api.github.com/users/meg-huggingface/followers", "following_url": "https://api.github.com/users/meg-huggingface/following{/other_user}", "gists_url": "https://api.github.com/users/meg-huggingface/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/meg-huggingface", "id": 90473723, "login": "meg-huggingface", "node_id": "MDQ6VXNlcjkwNDczNzIz", "organizations_url": "https://api.github.com/users/meg-huggingface/orgs", "received_events_url": "https://api.github.com/users/meg-huggingface/received_events", "repos_url": "https://api.github.com/users/meg-huggingface/repos", "site_admin": false, "starred_url": "https://api.github.com/users/meg-huggingface/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/meg-huggingface/subscriptions", "type": "User", "url": "https://api.github.com/users/meg-huggingface" }
[]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/90473723?v=4", "events_url": "https://api.github.com/users/meg-huggingface/events{/privacy}", "followers_url": "https://api.github.com/users/meg-huggingface/followers", "following_url": "https://api.github.com/users/meg-huggingface/following{/other_user}", "gists_url": "https://api.github.com/users/meg-huggingface/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/meg-huggingface", "id": 90473723, "login": "meg-huggingface", "node_id": "MDQ6VXNlcjkwNDczNzIz", "organizations_url": "https://api.github.com/users/meg-huggingface/orgs", "received_events_url": "https://api.github.com/users/meg-huggingface/received_events", "repos_url": "https://api.github.com/users/meg-huggingface/repos", "site_admin": false, "starred_url": "https://api.github.com/users/meg-huggingface/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/meg-huggingface/subscriptions", "type": "User", "url": "https://api.github.com/users/meg-huggingface" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/90473723?v=4", "events_url": "https://api.github.com/users/meg-huggingface/events{/privacy}", "followers_url": "https://api.github.com/users/meg-huggingface/followers", "following_url": "https://api.github.com/users/meg-huggingface/following{/...
null
[ "I'm not sure that we want to change the tags at the top of the cards by hand. Those are used to create the tags in the hub. Although looking at all the tags now, we might want to normalize the current tags again (hyphens or no, \".0\" or no). Maybe we could add a binary tag for public domain or not?", "> \r\n\r\...
2022-01-04T20:58:25Z
2022-01-05T12:30:21Z
2022-01-05T12:30:20Z
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/3520.diff", "html_url": "https://github.com/huggingface/datasets/pull/3520", "merged_at": "2022-01-05T12:30:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/3520.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3520" }
Filling out data card "Personal and Sensitive Information" for speech datasets to note PII concerns
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3520/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3520/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2688
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2688/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2688/comments
https://api.github.com/repos/huggingface/datasets/issues/2688/events
https://github.com/huggingface/datasets/issues/2688
949,182,074
MDU6SXNzdWU5NDkxODIwNzQ=
2,688
hebrew language codes he and iw should be treated as aliases
{ "avatar_url": "https://avatars.githubusercontent.com/u/4436747?v=4", "events_url": "https://api.github.com/users/eyaler/events{/privacy}", "followers_url": "https://api.github.com/users/eyaler/followers", "following_url": "https://api.github.com/users/eyaler/following{/other_user}", "gists_url": "https://api.github.com/users/eyaler/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/eyaler", "id": 4436747, "login": "eyaler", "node_id": "MDQ6VXNlcjQ0MzY3NDc=", "organizations_url": "https://api.github.com/users/eyaler/orgs", "received_events_url": "https://api.github.com/users/eyaler/received_events", "repos_url": "https://api.github.com/users/eyaler/repos", "site_admin": false, "starred_url": "https://api.github.com/users/eyaler/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eyaler/subscriptions", "type": "User", "url": "https://api.github.com/users/eyaler" }
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
[]
null
[ "Hi @eyaler, thanks for reporting.\r\n\r\nWhile you are true with respect the Hebrew language tag (\"iw\" is deprecated and \"he\" is the preferred value), in the \"mc4\" dataset (which is a derived dataset) we have kept the language tags present in the original dataset: [Google C4](https://www.tensorflow.org/datas...
2021-07-20T23:13:52Z
2021-07-21T16:34:53Z
2021-07-21T16:34:53Z
NONE
null
null
null
https://huggingface.co/datasets/mc4 not listed when searching for hebrew datasets (he) as it uses the older language code iw, preventing discoverability.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2688/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2688/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/6231
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6231/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6231/comments
https://api.github.com/repos/huggingface/datasets/issues/6231/events
https://github.com/huggingface/datasets/pull/6231
1,890,863,249
PR_kwDODunzps5aCr8_
6,231
Overwrite legacy default config name in `dataset_infos.json` in packaged datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6231). All of your documentation changes will be reflected on that endpoint.", "realized that this pr is still not merged, @lhoestq maybe you can take a look at it? ", "I think https://github.com/huggingface/datasets/pull/621...
2023-09-11T16:27:09Z
2023-09-26T11:19:36Z
null
CONTRIBUTOR
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/6231.diff", "html_url": "https://github.com/huggingface/datasets/pull/6231", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/6231.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6231" }
Currently if we push data as default config with `.push_to_hub` to a repo that has a legacy `dataset_infos.json` file containing a legacy default config name like `{username}--{dataset_name}`, new key `"default"` is added to `dataset_infos.json` along with the legacy one. I think the legacy one should be dropped in this case. Also, in `load.py` I suggest to check if a legacy config name is indeed a legacy config name because after this fix it might not be the case (this check was first introduced in https://github.com/huggingface/datasets/pull/6218)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6231/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6231/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/4605
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4605/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4605/comments
https://api.github.com/repos/huggingface/datasets/issues/4605/events
https://github.com/huggingface/datasets/issues/4605
1,290,058,970
I_kwDODunzps5M5Lza
4,605
Dataset Viewer issue for boris/gis_filtered
{ "avatar_url": "https://avatars.githubusercontent.com/u/41203448?v=4", "events_url": "https://api.github.com/users/WaterKnight1998/events{/privacy}", "followers_url": "https://api.github.com/users/WaterKnight1998/followers", "following_url": "https://api.github.com/users/WaterKnight1998/following{/other_user}", "gists_url": "https://api.github.com/users/WaterKnight1998/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/WaterKnight1998", "id": 41203448, "login": "WaterKnight1998", "node_id": "MDQ6VXNlcjQxMjAzNDQ4", "organizations_url": "https://api.github.com/users/WaterKnight1998/orgs", "received_events_url": "https://api.github.com/users/WaterKnight1998/received_events", "repos_url": "https://api.github.com/users/WaterKnight1998/repos", "site_admin": false, "starred_url": "https://api.github.com/users/WaterKnight1998/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/WaterKnight1998/subscriptions", "type": "User", "url": "https://api.github.com/users/WaterKnight1998" }
[ { "color": "fef2c0", "default": false, "description": "", "id": 3287858981, "name": "streaming", "node_id": "MDU6TGFiZWwzMjg3ODU4OTgx", "url": "https://api.github.com/repos/huggingface/datasets/labels/streaming" } ]
closed
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
[ "Yes, this dataset is \"gated\": you first have to go to https://huggingface.co/datasets/boris/gis_filtered and click \"Access repository\" (if you accept to share your contact information with the repository authors).", "I already did that, it returns error when using streaming", "Oh, sorry, I misread. Looking...
2022-06-30T12:23:34Z
2022-07-06T12:34:19Z
2022-07-06T12:34:19Z
NONE
null
null
null
### Link https://huggingface.co/datasets/boris/gis_filtered/viewer/boris--gis_filtered/train ### Description When I try to access this from the website I get this error: Status code: 400 Exception: ClientResponseError Message: 401, message='Unauthorized', url=URL('https://huggingface.co/datasets/boris/gis_filtered/resolve/80b805053ce61d4eb487b6b8d9095d775c2c466e/data/train/0000.parquet') If I try to load with code I also get the same issue: ```python dataset2_train=load_dataset("boris/gis_filtered", use_auth_token=os.environ["HF_TOKEN"],split="train",streaming=True) dataset2_validation=load_dataset("boris/gis_filtered", use_auth_token=os.environ["HF_TOKEN"], split="validation",streaming=True) ``` ### Owner No
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4605/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4605/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/5243
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5243/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5243/comments
https://api.github.com/repos/huggingface/datasets/issues/5243/events
https://github.com/huggingface/datasets/issues/5243
1,449,523,962
I_kwDODunzps5WZfr6
5,243
Download only split data
{ "avatar_url": "https://avatars.githubusercontent.com/u/48530104?v=4", "events_url": "https://api.github.com/users/capsabogdan/events{/privacy}", "followers_url": "https://api.github.com/users/capsabogdan/followers", "following_url": "https://api.github.com/users/capsabogdan/following{/other_user}", "gists_url": "https://api.github.com/users/capsabogdan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/capsabogdan", "id": 48530104, "login": "capsabogdan", "node_id": "MDQ6VXNlcjQ4NTMwMTA0", "organizations_url": "https://api.github.com/users/capsabogdan/orgs", "received_events_url": "https://api.github.com/users/capsabogdan/received_events", "repos_url": "https://api.github.com/users/capsabogdan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/capsabogdan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/capsabogdan/subscriptions", "type": "User", "url": "https://api.github.com/users/capsabogdan" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[ "Hi @capsabogdan! Unfortunately, it's hard to implement because quite often datasets data is being hosted in a single archive for all splits :( So we have to download the whole archive to split it into splits. This is the case for CommonVoice too. \r\n\r\nHowever, for cases when data is distributed in separate arch...
2022-11-15T10:15:54Z
2023-05-02T09:27:51Z
null
NONE
null
null
null
### Feature request Is it possible to download only the data that I am requesting and not the entire dataset? I run out of disk spaceas it seems to download the entire dataset, instead of only the part needed. common_voice["test"] = load_dataset("mozilla-foundation/common_voice_11_0", "en", split="test", cache_dir="cache/path...", use_auth_token=True, download_config=DownloadConfig(delete_extracted='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') ) ### Motivation efficiency improvement ### Your contribution n/a
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5243/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5243/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/5327
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5327/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5327/comments
https://api.github.com/repos/huggingface/datasets/issues/5327/events
https://github.com/huggingface/datasets/pull/5327
1,471,657,247
PR_kwDODunzps5EE_3Q
5,327
Avoid unwanted behaviour when splits from script and metadata are not matching because of outdated metadata
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
open
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_use...
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5327). All of your documentation changes will be reflected on that endpoint." ]
2022-12-01T17:05:23Z
2023-01-23T12:48:29Z
null
CONTRIBUTOR
null
1
{ "diff_url": "https://github.com/huggingface/datasets/pull/5327.diff", "html_url": "https://github.com/huggingface/datasets/pull/5327", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5327.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5327" }
will fix #5315
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5327/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5327/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/2826
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2826/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2826/comments
https://api.github.com/repos/huggingface/datasets/issues/2826/events
https://github.com/huggingface/datasets/issues/2826
976,974,254
MDU6SXNzdWU5NzY5NzQyNTQ=
2,826
Add a Text Classification dataset: KanHope
{ "avatar_url": "https://avatars.githubusercontent.com/u/46108405?v=4", "events_url": "https://api.github.com/users/adeepH/events{/privacy}", "followers_url": "https://api.github.com/users/adeepH/followers", "following_url": "https://api.github.com/users/adeepH/following{/other_user}", "gists_url": "https://api.github.com/users/adeepH/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/adeepH", "id": 46108405, "login": "adeepH", "node_id": "MDQ6VXNlcjQ2MTA4NDA1", "organizations_url": "https://api.github.com/users/adeepH/orgs", "received_events_url": "https://api.github.com/users/adeepH/received_events", "repos_url": "https://api.github.com/users/adeepH/repos", "site_admin": false, "starred_url": "https://api.github.com/users/adeepH/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/adeepH/subscriptions", "type": "User", "url": "https://api.github.com/users/adeepH" }
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
null
[]
null
[ "Hi ! In your script it looks like you're trying to load the dataset `bn_hate_speech,`, not KanHope.\r\n\r\nMoreover the error `KeyError: ' '` means that you have a feature of type ClassLabel, but for a certain example of the dataset, it looks like the label is empty (it's just a string with a space). Can you make ...
2021-08-23T12:21:58Z
2021-10-01T18:06:59Z
2021-10-01T18:06:59Z
CONTRIBUTOR
null
null
null
## Adding a Dataset - **Name:** *KanHope* - **Description:** *A code-mixed English-Kannada dataset for Hope speech detection* - **Paper:** *https://arxiv.org/abs/2108.04616* (I am the author of the paper} - **Author:** *[AdeepH](https://github.com/adeepH)* - **Data:** *https://github.com/adeepH/KanHope/tree/main/dataset* - **Motivation:** *The dataset is amongst the very few resources available for code-mixed Dravidian languages* - I tried following the steps as per the instructions. However, could not resolve an error. Any help would be appreciated. - The dataset card and the scripts for the dataset *https://github.com/adeepH/datasets/tree/multilingual-hope-speech/datasets/mhs_eval* ``` Using custom data configuration default Downloading and preparing dataset bn_hate_speech/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to /root/.cache/huggingface/datasets/bn_hate_speech/default/0.0.0/5f417ddc89777278abd29988f909f39495f0ec802090f7d8fa63b5bffb121762... --------------------------------------------------------------------------- KeyError Traceback (most recent call last) <ipython-input-114-4a9cdb519e4c> in <module>() 1 from datasets import load_dataset 2 ----> 3 data = load_dataset('/content/bn') 9 frames /usr/local/lib/python3.7/dist-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, task, streaming, **config_kwargs) 850 ignore_verifications=ignore_verifications, 851 try_from_hf_gcs=try_from_hf_gcs, --> 852 use_auth_token=use_auth_token, 853 ) 854 /usr/local/lib/python3.7/dist-packages/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs) 614 if not downloaded_from_gcs: 615 self._download_and_prepare( --> 616 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs 617 ) 618 # Sync info /usr/local/lib/python3.7/dist-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 691 try: 692 # Prepare split will record examples associated to the split --> 693 self._prepare_split(split_generator, **prepare_split_kwargs) 694 except OSError as e: 695 raise OSError( /usr/local/lib/python3.7/dist-packages/datasets/builder.py in _prepare_split(self, split_generator) 1107 disable=bool(logging.get_verbosity() == logging.NOTSET), 1108 ): -> 1109 example = self.info.features.encode_example(record) 1110 writer.write(example, key) 1111 finally: /usr/local/lib/python3.7/dist-packages/datasets/features.py in encode_example(self, example) 1015 """ 1016 example = cast_to_python_objects(example) -> 1017 return encode_nested_example(self, example) 1018 1019 def encode_batch(self, batch): /usr/local/lib/python3.7/dist-packages/datasets/features.py in encode_nested_example(schema, obj) 863 if isinstance(schema, dict): 864 return { --> 865 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj) 866 } 867 elif isinstance(schema, (list, tuple)): /usr/local/lib/python3.7/dist-packages/datasets/features.py in <dictcomp>(.0) 863 if isinstance(schema, dict): 864 return { --> 865 k: encode_nested_example(sub_schema, sub_obj) for k, (sub_schema, sub_obj) in utils.zip_dict(schema, obj) 866 } 867 elif isinstance(schema, (list, tuple)): /usr/local/lib/python3.7/dist-packages/datasets/features.py in encode_nested_example(schema, obj) 890 # ClassLabel will convert from string to int, TranslationVariableLanguages does some checks 891 elif isinstance(schema, (ClassLabel, TranslationVariableLanguages, Value, _ArrayXD)): --> 892 return schema.encode_example(obj) 893 # Other object should be directly convertible to a native Arrow type (like Translation and Translation) 894 return obj /usr/local/lib/python3.7/dist-packages/datasets/features.py in encode_example(self, example_data) 665 # If a string is given, convert to associated integer 666 if isinstance(example_data, str): --> 667 example_data = self.str2int(example_data) 668 669 # Allowing -1 to mean no label. /usr/local/lib/python3.7/dist-packages/datasets/features.py in str2int(self, values) 623 if value not in self._str2int: 624 value = str(value).strip() --> 625 output.append(self._str2int[str(value)]) 626 else: 627 # No names provided, try to integerize KeyError: ' ' ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2826/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2826/timeline
null
completed
false