url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
48
51
id
int64
600M
2.19B
node_id
stringlengths
18
24
number
int64
2
6.73k
title
stringlengths
1
290
user
dict
labels
listlengths
0
4
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
listlengths
0
30
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
3 values
active_lock_reason
null
draft
null
pull_request
null
body
stringlengths
0
228k
βŒ€
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
state_reason
stringclasses
3 values
https://api.github.com/repos/huggingface/datasets/issues/5874
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5874/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5874/comments
https://api.github.com/repos/huggingface/datasets/issues/5874/events
https://github.com/huggingface/datasets/issues/5874
1,715,708,930
I_kwDODunzps5mQ6QC
5,874
Using as_dataset on a "parquet" builder
{ "login": "rems75", "id": 9039058, "node_id": "MDQ6VXNlcjkwMzkwNTg=", "avatar_url": "https://avatars.githubusercontent.com/u/9039058?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rems75", "html_url": "https://github.com/rems75", "followers_url": "https://api.github.com/users/rems75/foll...
[]
closed
false
null
[]
null
[ "Hi! You can refer to [this doc](https://huggingface.co/docs/datasets/filesystems#load-and-save-your-datasets-using-your-cloud-storage-filesystem) to see the intended usage (basically, it skips the Arrow -> Parquet conversion step in `ds = load_dataset(...); ds.to_parquet(\"path/to/parquet\")`) and allows writing P...
2023-05-18T14:09:03
2023-05-31T13:23:55
2023-05-31T13:23:55
NONE
null
null
null
### Describe the bug I used a custom builder to ``download_and_prepare`` a dataset. The first (very minor) issue is that the doc seems to suggest ``download_and_prepare`` will return the dataset, while it does not ([builder.py](https://github.com/huggingface/datasets/blob/main/src/datasets/builder.py#L718-L738)). ```...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5874/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5874/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5873
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5873/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5873/comments
https://api.github.com/repos/huggingface/datasets/issues/5873/events
https://github.com/huggingface/datasets/issues/5873
1,713,269,724
I_kwDODunzps5mHmvc
5,873
Allow setting the environment variable for the lock file path
{ "login": "xin3he", "id": 83260933, "node_id": "MDQ6VXNlcjgzMjYwOTMz", "avatar_url": "https://avatars.githubusercontent.com/u/83260933?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xin3he", "html_url": "https://github.com/xin3he", "followers_url": "https://api.github.com/users/xin3he/fo...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[]
2023-05-17T07:10:02
2023-05-17T07:11:05
null
NONE
null
null
null
### Feature request Add an environment variable to replace the default lock file path. ### Motivation Usually, dataset path is a read-only path while the lock file needs to be modified each time. It would be convenient if the path can be reset individually. ### Your contribution ```/src/datasets/utils/fi...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5873/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5873/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5871
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5871/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5871/comments
https://api.github.com/repos/huggingface/datasets/issues/5871/events
https://github.com/huggingface/datasets/issues/5871
1,712,573,073
I_kwDODunzps5mE8qR
5,871
data configuration hash suffix depends on uncanonicalized data_dir
{ "login": "kylrth", "id": 5044802, "node_id": "MDQ6VXNlcjUwNDQ4MDI=", "avatar_url": "https://avatars.githubusercontent.com/u/5044802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kylrth", "html_url": "https://github.com/kylrth", "followers_url": "https://api.github.com/users/kylrth/foll...
[ { "id": 1935892877, "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue", "name": "good first issue", "color": "7057ff", "default": true, "description": "Good for newcomers" } ]
closed
false
{ "login": "kylrth", "id": 5044802, "node_id": "MDQ6VXNlcjUwNDQ4MDI=", "avatar_url": "https://avatars.githubusercontent.com/u/5044802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kylrth", "html_url": "https://github.com/kylrth", "followers_url": "https://api.github.com/users/kylrth/foll...
[ { "login": "kylrth", "id": 5044802, "node_id": "MDQ6VXNlcjUwNDQ4MDI=", "avatar_url": "https://avatars.githubusercontent.com/u/5044802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kylrth", "html_url": "https://github.com/kylrth", "followers_url": "https://api.github.c...
null
[ "It could even use `os.path.realpath` to resolve symlinks.", "Indeed, it makes sense to normalize `data_dir`. Feel free to submit a PR (this can be \"fixed\" [here](https://github.com/huggingface/datasets/blob/89f775226321ba94e5bf4670a323c0fb44f5f65c/src/datasets/builder.py#L173))", "#self-assign" ]
2023-05-16T18:56:04
2023-06-02T15:52:05
2023-06-02T15:52:05
CONTRIBUTOR
null
null
null
### Describe the bug I am working with the `recipe_nlg` dataset, which requires manual download. Once it's downloaded, I've noticed that the hash in the custom data configuration is different if I add a trailing `/` to my `data_dir`. It took me a while to notice that the hashes were different, and to understand that...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5871/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5871/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5870
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5870/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5870/comments
https://api.github.com/repos/huggingface/datasets/issues/5870/events
https://github.com/huggingface/datasets/issues/5870
1,712,156,282
I_kwDODunzps5mDW56
5,870
Behaviour difference between datasets.map and IterableDatasets.map
{ "login": "llStringll", "id": 30209072, "node_id": "MDQ6VXNlcjMwMjA5MDcy", "avatar_url": "https://avatars.githubusercontent.com/u/30209072?v=4", "gravatar_id": "", "url": "https://api.github.com/users/llStringll", "html_url": "https://github.com/llStringll", "followers_url": "https://api.github.com/use...
[]
open
false
null
[]
null
[ "PS - some work is definitely needed for 'special cases' docs, not explanations, just usages of 'functions' under mixture of special cases, like a combination of custom databuilder + iterable dataset for large size + dynamic .map() application." ]
2023-05-16T14:32:57
2023-05-16T14:36:05
null
NONE
null
null
null
### Describe the bug All the examples in all the docs mentioned throughout huggingface datasets correspond to datasets object, and not IterableDatasets object. At one point of time, they might have been in sync, but the code for datasets version >=2.9.0 is very different as compared to the docs. I basically need to ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5870/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5870/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5869
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5869/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5869/comments
https://api.github.com/repos/huggingface/datasets/issues/5869/events
https://github.com/huggingface/datasets/issues/5869
1,711,990,003
I_kwDODunzps5mCuTz
5,869
Image Encoding Issue when submitting a Parquet Dataset
{ "login": "PhilippeMoussalli", "id": 47530815, "node_id": "MDQ6VXNlcjQ3NTMwODE1", "avatar_url": "https://avatars.githubusercontent.com/u/47530815?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PhilippeMoussalli", "html_url": "https://github.com/PhilippeMoussalli", "followers_url": "https...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi @PhilippeMoussalli thanks for opening a detailed issue. It seems the issue is more related to the `datasets` library so I'll ping @lhoestq @mariosasko on this one :) \n\n(edit: also can one of you move the issue to the datasets repo? Thanks in advance πŸ™)", "Hi ! The `Image()` info is stored in the **schema m...
2023-05-16T09:42:58
2023-06-16T12:48:38
2023-06-16T09:30:48
NONE
null
null
null
### Describe the bug Hello, I'd like to report an issue related to pushing a dataset represented as a Parquet file to a dataset repository using Dask. Here are the details: We attempted to load an example dataset in Parquet format from the Hugging Face (HF) filesystem using Dask with the following code snippet...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5869/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5869/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5868
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5868/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5868/comments
https://api.github.com/repos/huggingface/datasets/issues/5868/events
https://github.com/huggingface/datasets/issues/5868
1,711,173,098
I_kwDODunzps5l_m3q
5,868
Is it possible to change a cached file and 're-cache' it instead of re-generating?
{ "login": "zyh3826", "id": 31238754, "node_id": "MDQ6VXNlcjMxMjM4NzU0", "avatar_url": "https://avatars.githubusercontent.com/u/31238754?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zyh3826", "html_url": "https://github.com/zyh3826", "followers_url": "https://api.github.com/users/zyh382...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
[ "Arrow files/primitives (tables and arrays) are immutable, so re-generating them is the only option, I'm afraid.", "> \r\n\r\nGot it, thanks for your reply" ]
2023-05-16T03:45:42
2023-05-17T11:21:36
2023-05-17T11:21:36
NONE
null
null
null
### Feature request Hi, I have a huge cached file using `map`(over 500GB), and I want to change an attribution of each element, is there possible to do it using some method instead of re-generating, because `map` takes over 24 hours ### Motivation For large datasets, I think it is very important because we always f...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5868/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5868/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5866
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5866/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5866/comments
https://api.github.com/repos/huggingface/datasets/issues/5866/events
https://github.com/huggingface/datasets/issues/5866
1,710,496,993
I_kwDODunzps5l9Bzh
5,866
Issue with Sequence features
{ "login": "alialamiidrissi", "id": 14365168, "node_id": "MDQ6VXNlcjE0MzY1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/14365168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/alialamiidrissi", "html_url": "https://github.com/alialamiidrissi", "followers_url": "https://api...
[]
closed
false
null
[]
null
[ "Thanks for reporting! I've opened a PR with a fix." ]
2023-05-15T17:13:29
2023-05-26T11:57:17
2023-05-26T11:57:17
NONE
null
null
null
### Describe the bug Sequences features sometimes causes errors when the specified length is not -1 ### Steps to reproduce the bug ```python import numpy as np from datasets import Features, ClassLabel, Sequence, Value, Dataset feats = Features(**{'target': ClassLabel(names=[0, 1]),'x': Sequence(feature=Va...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5866/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5866/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5864
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5864/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5864/comments
https://api.github.com/repos/huggingface/datasets/issues/5864/events
https://github.com/huggingface/datasets/issues/5864
1,710,450,047
I_kwDODunzps5l82V_
5,864
Slow iteration over Torch tensors
{ "login": "crisostomi", "id": 51738205, "node_id": "MDQ6VXNlcjUxNzM4MjA1", "avatar_url": "https://avatars.githubusercontent.com/u/51738205?v=4", "gravatar_id": "", "url": "https://api.github.com/users/crisostomi", "html_url": "https://github.com/crisostomi", "followers_url": "https://api.github.com/use...
[]
open
false
null
[]
null
[ "I am highly interested performance of dataset so I ran your example as a curious user.\r\n```python\r\ntrain_dataset.cast_column(\"x\", Array3D(shape=img_shape, dtype=\"float32\"))\r\n```\r\nhave return values and \"x\" is a new column, it shoulde be\r\n```python\r\nds=train_dataset.cast_column(\"img\", Array3D(sh...
2023-05-15T16:43:58
2023-05-16T03:27:38
null
NONE
null
null
null
### Describe the bug I have a problem related to this [issue](https://github.com/huggingface/datasets/issues/5841): I get a way slower iteration when using a Torch dataloader if I use vanilla Numpy tensors or if I first apply a ToTensor transform to the input. In particular, it takes 5 seconds to iterate over the vani...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5864/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5864/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5862
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5862/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5862/comments
https://api.github.com/repos/huggingface/datasets/issues/5862/events
https://github.com/huggingface/datasets/issues/5862
1,710,140,646
I_kwDODunzps5l7qzm
5,862
IndexError: list index out of range with data hosted on Zenodo
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "This error is also raised when data is hosted on Google Drive:\r\n- https://huggingface.co/datasets/docred/discussions/5\r\n- https://huggingface.co/datasets/linnaeus/discussions/3\r\n- https://huggingface.co/datasets/poleval2019_mt/discussions/3\r\n- https://huggingface.co/datasets/reddit_tifu/discussions/2\r\n- ...
2023-05-15T13:47:19
2023-09-25T12:09:51
null
MEMBER
null
null
null
The dataset viewer sometimes raises an `IndexError`: ``` IndexError: list index out of range ``` See: - huggingface/datasets-server#1151 - https://huggingface.co/datasets/reddit/discussions/5 - huggingface/datasets-server#1118 - https://huggingface.co/datasets/krr-oxford/OntoLAMA/discussions/1 - https://hu...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5862/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5862/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5858
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5858/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5858/comments
https://api.github.com/repos/huggingface/datasets/issues/5858/events
https://github.com/huggingface/datasets/issues/5858
1,709,332,632
I_kwDODunzps5l4liY
5,858
Throw an error when dataset improperly indexed
{ "login": "sarahwie", "id": 8027676, "node_id": "MDQ6VXNlcjgwMjc2NzY=", "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sarahwie", "html_url": "https://github.com/sarahwie", "followers_url": "https://api.github.com/users/sarah...
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Thanks for reporting, @sarahwie.\r\n\r\nPlease note that in `datasets` we do not have vectorized operation like `pandas`. Therefore, your equality comparisons above are `False`:\r\n- For example: `squad['question']` returns a `list`, and this list is not equal to `\"Who was the Norse leader?\"`\r\n\r\nThe `False` ...
2023-05-15T05:15:53
2023-05-25T16:23:19
2023-05-25T16:23:19
NONE
null
null
null
### Describe the bug Pandas-style subset indexing on dataset does not throw an error, when maybe it should. Instead returns the first instance of the dataset regardless of index condition. ### Steps to reproduce the bug Steps to reproduce the behavior: 1. `squad = datasets.load_dataset("squad_v2", split="validati...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5858/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5858/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5857
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5857/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5857/comments
https://api.github.com/repos/huggingface/datasets/issues/5857/events
https://github.com/huggingface/datasets/issues/5857
1,709,326,622
I_kwDODunzps5l4kEe
5,857
Adding chemistry dataset/models in huggingface
{ "login": "knc6", "id": 16902896, "node_id": "MDQ6VXNlcjE2OTAyODk2", "avatar_url": "https://avatars.githubusercontent.com/u/16902896?v=4", "gravatar_id": "", "url": "https://api.github.com/users/knc6", "html_url": "https://github.com/knc6", "followers_url": "https://api.github.com/users/knc6/followers"...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
[ "Hi! \r\n\r\nThis would be a nice addition to the Hub! You can find the existing chemistry datasets/models on the Hub (using the `chemistry` tag) [here](https://huggingface.co/search/full-text?q=chemistry&type=model&type=dataset).\r\n\r\nFeel free to ping us here on the Hub if you need help adding the datasets.\r\n...
2023-05-15T05:09:49
2023-07-21T13:45:40
2023-07-21T13:45:40
NONE
null
null
null
### Feature request Huggingface is really amazing platform for open science. In addition to computer vision, video and NLP, would it be of interest to add chemistry/materials science dataset/models in Huggingface? Or, if its already done, can you provide some pointers. We have been working on a comprehensive ben...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5857/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5857/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5856
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5856/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5856/comments
https://api.github.com/repos/huggingface/datasets/issues/5856/events
https://github.com/huggingface/datasets/issues/5856
1,709,218,242
I_kwDODunzps5l4JnC
5,856
Error loading natural_questions
{ "login": "Crownor", "id": 19185508, "node_id": "MDQ6VXNlcjE5MTg1NTA4", "avatar_url": "https://avatars.githubusercontent.com/u/19185508?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Crownor", "html_url": "https://github.com/Crownor", "followers_url": "https://api.github.com/users/Crowno...
[]
closed
false
null
[]
null
[ "Hi! You can avoid this error by using the preprocessed version:\r\n```python\r\nimport datasets\r\nds = datasets.load_dataset('natural_questions')\r\n```\r\n\r\nPS: Once we finish https://github.com/huggingface/datasets/pull/5364, this error will no longer be a problem.", "> Hi! You can avoid this error by using...
2023-05-15T02:46:04
2023-06-05T09:11:19
2023-06-05T09:11:18
NONE
null
null
null
### Describe the bug When try to load natural_questions through datasets == 2.12.0 with python == 3.8.9: ```python import datasets datasets.load_dataset('natural_questions',beam_runner='DirectRunner') ``` It failed with following info: `pyarrow.lib.ArrowNotImplementedError: Nested data conversions not impl...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5856/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5856/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5855
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5855/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5855/comments
https://api.github.com/repos/huggingface/datasets/issues/5855/events
https://github.com/huggingface/datasets/issues/5855
1,708,784,943
I_kwDODunzps5l2f0v
5,855
`to_tf_dataset` consumes too much memory
{ "login": "massquantity", "id": 28751760, "node_id": "MDQ6VXNlcjI4NzUxNzYw", "avatar_url": "https://avatars.githubusercontent.com/u/28751760?v=4", "gravatar_id": "", "url": "https://api.github.com/users/massquantity", "html_url": "https://github.com/massquantity", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Cc @amyeroberts @Rocketknight1 \r\n\r\nIndded I think it's because it does something like this under the hood when there's no multiprocessing:\r\n\r\n```python\r\ntf_dataset = tf_dataset.shuffle(len(dataset))\r\n```\r\n\r\nPS: with multiprocessing it appears to be different:\r\n\r\n```python\r\nindices = np.arange...
2023-05-14T01:22:29
2023-06-08T16:32:52
2023-06-08T16:32:52
NONE
null
null
null
### Describe the bug Hi, I'm using `to_tf_dataset` to convert a _large_ dataset to `tf.data.Dataset`. I observed that the data loading *before* training took a lot of time and memory, even with `batch_size=1`. After some digging, i believe the reason lies in the shuffle behavior. The [source code](https://github....
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5855/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5855/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5854
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5854/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5854/comments
https://api.github.com/repos/huggingface/datasets/issues/5854/events
https://github.com/huggingface/datasets/issues/5854
1,708,779,300
I_kwDODunzps5l2eck
5,854
Can not load audiofolder dataset on kaggle
{ "login": "ILG2021", "id": 93691919, "node_id": "U_kgDOBZWgDw", "avatar_url": "https://avatars.githubusercontent.com/u/93691919?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ILG2021", "html_url": "https://github.com/ILG2021", "followers_url": "https://api.github.com/users/ILG2021/follow...
[]
closed
false
null
[]
null
[ "Hi! `audiofolder` requires `datasets>=2.5.0`, so please update the `datasets`' installation (`pip install -U datasets`) in the environment (and restart the env for the update to take effect) to resolve the issue.", "> Hi! `audiofolder` requires `datasets>=2.5.0`, so please update the `datasets`' installation (`p...
2023-05-14T00:50:47
2023-08-16T13:35:36
2023-07-21T13:53:45
NONE
null
null
null
### Describe the bug It's crash log: FileNotFoundError: Couldn't find a dataset script at /kaggle/working/audiofolder/audiofolder.py or any data file in the same directory. Couldn't find 'audiofolder' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingfac...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5854/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5854/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5849
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5849/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5849/comments
https://api.github.com/repos/huggingface/datasets/issues/5849/events
https://github.com/huggingface/datasets/issues/5849
1,707,551,511
I_kwDODunzps5lxysX
5,849
CSV datasets should only read the CSV data files in the repo
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[]
2023-05-12T12:29:53
2023-06-22T14:16:27
2023-06-22T14:16:27
MEMBER
null
null
null
When a no-script dataset has many CSV files and a JPG file, the library infers to use the Csv builder, but tries to read as CSV all files in the repo, also the JPG file. I think the Csv builder should filter out non-CSV files when reading. An analogue solution should be implemented for other packaged builders. ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5849/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5849/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5847
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5847/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5847/comments
https://api.github.com/repos/huggingface/datasets/issues/5847/events
https://github.com/huggingface/datasets/issues/5847
1,706,616,634
I_kwDODunzps5luOc6
5,847
Streaming IterableDataset not working with translation pipeline
{ "login": "jlquinn", "id": 826841, "node_id": "MDQ6VXNlcjgyNjg0MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/826841?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jlquinn", "html_url": "https://github.com/jlquinn", "followers_url": "https://api.github.com/users/jlquinn/fo...
[]
open
false
null
[]
null
[ "I wasn't sure to file this against transformers or datasets.", "[`KeyDataset`](https://github.com/huggingface/transformers/blob/7f8b909189547944617741d8d3c6c84504701693/src/transformers/pipelines/pt_utils.py#L296) doesn't support iterable datasets, so you either need to implement a version that does (and also in...
2023-05-11T21:52:38
2023-05-16T15:59:55
null
NONE
null
null
null
### Describe the bug I'm trying to use a streaming dataset for translation inference to avoid downloading the training data. I'm using a pipeline and a dataset, and following the guidance in the tutorial. Instead I get an exception that IterableDataset has no len(). ### Steps to reproduce the bug CODE: ``` ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5847/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5847/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5851
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5851/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5851/comments
https://api.github.com/repos/huggingface/datasets/issues/5851/events
https://github.com/huggingface/datasets/issues/5851
1,707,907,048
I_kwDODunzps5lzJfo
5,851
Error message not clear in interleaving datasets
{ "login": "surya-narayanan", "id": 17240858, "node_id": "MDQ6VXNlcjE3MjQwODU4", "avatar_url": "https://avatars.githubusercontent.com/u/17240858?v=4", "gravatar_id": "", "url": "https://api.github.com/users/surya-narayanan", "html_url": "https://github.com/surya-narayanan", "followers_url": "https://api...
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[]
2023-05-11T20:52:13
2023-05-23T10:32:59
2023-05-23T10:32:59
NONE
null
null
null
### System Info standard env ### Who can help? _No response_ ### Information - [ ] The official example scripts - [X] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [X] My own task or dataset (give details below) ### Reproduction I'm tr...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5851/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5851/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5846
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5846/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5846/comments
https://api.github.com/repos/huggingface/datasets/issues/5846/events
https://github.com/huggingface/datasets/issues/5846
1,706,289,290
I_kwDODunzps5ls-iK
5,846
load_dataset('bigcode/the-stack-dedup', streaming=True) very slow!
{ "login": "tbenthompson", "id": 4241811, "node_id": "MDQ6VXNlcjQyNDE4MTE=", "avatar_url": "https://avatars.githubusercontent.com/u/4241811?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tbenthompson", "html_url": "https://github.com/tbenthompson", "followers_url": "https://api.github.com...
[]
open
false
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[ { "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https:...
null
[ "This is due to the slow resolution of the data files: https://github.com/huggingface/datasets/issues/5537.\r\n\r\nWe plan to switch to `huggingface_hub`'s `HfFileSystem` soon to make the resolution faster (will be up to 20x faster once we merge https://github.com/huggingface/huggingface_hub/pull/1443)\r\n\r\n", ...
2023-05-11T17:58:57
2023-05-16T03:23:46
null
NONE
null
null
null
### Describe the bug Running ``` import datasets ds = datasets.load_dataset('bigcode/the-stack-dedup', streaming=True) ``` takes about 2.5 minutes! I would expect this to be near instantaneous. With other datasets, the runtime is one or two seconds. ### Environment info - `datasets` version: 2.1...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5846/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5846/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5844
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5844/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5844/comments
https://api.github.com/repos/huggingface/datasets/issues/5844/events
https://github.com/huggingface/datasets/issues/5844
1,705,907,812
I_kwDODunzps5lrhZk
5,844
TypeError: Couldn't cast array of type struct<answer: struct<unanswerable: bool, answerType: string, free_form_answer: string, evidence: list<item: string>, evidenceAnnotate: list<item: string>, highlighted_evidence: list<item: string>>> to ...
{ "login": "chen-coding", "id": 54010030, "node_id": "MDQ6VXNlcjU0MDEwMDMw", "avatar_url": "https://avatars.githubusercontent.com/u/54010030?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chen-coding", "html_url": "https://github.com/chen-coding", "followers_url": "https://api.github.com/...
[]
open
false
null
[]
null
[]
2023-05-11T14:15:01
2023-05-11T14:15:01
null
NONE
null
null
null
### Describe the bug TypeError: Couldn't cast array of type struct<answer: struct<unanswerable: bool, answerType: string, free_form_answer: string, evidence: list<item: string>, evidenceAnnotate: list<item: string>, highlighted_evidence: list<item: string>>> to {'answer': {'unanswerable': Value(dtype='bool', id=None),...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5844/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5844/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5841
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5841/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5841/comments
https://api.github.com/repos/huggingface/datasets/issues/5841/events
https://github.com/huggingface/datasets/issues/5841
1,705,286,639
I_kwDODunzps5lpJvv
5,841
Abusurdly slow on iteration
{ "login": "fecet", "id": 41792945, "node_id": "MDQ6VXNlcjQxNzkyOTQ1", "avatar_url": "https://avatars.githubusercontent.com/u/41792945?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fecet", "html_url": "https://github.com/fecet", "followers_url": "https://api.github.com/users/fecet/follow...
[]
closed
false
null
[]
null
[ "Hi ! You can try to use the [Image](https://huggingface.co/docs/datasets/v2.12.0/en/package_reference/main_classes#datasets.Image) type which [decodes images on-the-fly](https://huggingface.co/docs/datasets/v2.12.0/en/about_dataset_features#image-feature) into pytorch tensors :)\r\n\r\n```python\r\nds = Dataset.fr...
2023-05-11T08:04:09
2023-05-15T15:38:13
2023-05-15T15:38:13
NONE
null
null
null
### Describe the bug I am attempting to iterate through an image dataset, but I am encountering a significant slowdown in the iteration speed. In order to investigate this issue, I conducted the following experiment: ```python a=torch.randn(100,224) a=torch.stack([a] * 10000) a.shape # %% ds=Dataset.from_d...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5841/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5841/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5840
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5840/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5840/comments
https://api.github.com/repos/huggingface/datasets/issues/5840/events
https://github.com/huggingface/datasets/issues/5840
1,705,212,085
I_kwDODunzps5lo3i1
5,840
load model error.
{ "login": "LanShanPi", "id": 58167546, "node_id": "MDQ6VXNlcjU4MTY3NTQ2", "avatar_url": "https://avatars.githubusercontent.com/u/58167546?v=4", "gravatar_id": "", "url": "https://api.github.com/users/LanShanPi", "html_url": "https://github.com/LanShanPi", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Please report this in the `transformers` repo, as it's not related to `datasets`" ]
2023-05-11T07:12:38
2023-05-12T13:44:07
2023-05-12T13:44:06
NONE
null
null
null
### Describe the bug I had trained one model use deepspeed, when I load the final load I get the follow error: OSError: Can't load tokenizer for '/XXX/DeepSpeedExamples/applications/DeepSpeed-Chat/output/step3-models/1.3b/actor'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5840/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5840/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5842
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5842/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5842/comments
https://api.github.com/repos/huggingface/datasets/issues/5842/events
https://github.com/huggingface/datasets/issues/5842
1,705,510,602
I_kwDODunzps5lqAbK
5,842
Remove columns in interable dataset
{ "login": "surya-narayanan", "id": 17240858, "node_id": "MDQ6VXNlcjE3MjQwODU4", "avatar_url": "https://avatars.githubusercontent.com/u/17240858?v=4", "gravatar_id": "", "url": "https://api.github.com/users/surya-narayanan", "html_url": "https://github.com/surya-narayanan", "followers_url": "https://api...
[]
closed
false
null
[]
null
[ "Transferring this issue as it's related to the πŸ€— Datasets library ", "Hi @surya-narayanan! Could you provide some code snippet?", "This method has been recently added to the `IterableDataset`, so you need to update the `datasets`' installation (`pip install -U datasets`) to use it." ]
2023-05-11T03:48:46
2023-06-21T16:36:42
2023-06-21T16:36:41
NONE
null
null
null
### Feature request Right now, remove_columns() produces a NotImplementedError for iterable style datasets ### Motivation It would be great to have the same functionality irrespective of whether one is using an iterable or a map-style dataset ### Your contribution hope and courage.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5842/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5842/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5843
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5843/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5843/comments
https://api.github.com/repos/huggingface/datasets/issues/5843/events
https://github.com/huggingface/datasets/issues/5843
1,705,514,551
I_kwDODunzps5lqBY3
5,843
Can't add iterable datasets to a Dataset Dict.
{ "login": "surya-narayanan", "id": 17240858, "node_id": "MDQ6VXNlcjE3MjQwODU4", "avatar_url": "https://avatars.githubusercontent.com/u/17240858?v=4", "gravatar_id": "", "url": "https://api.github.com/users/surya-narayanan", "html_url": "https://github.com/surya-narayanan", "followers_url": "https://api...
[]
closed
false
null
[]
null
[ "Transferring as this is relating to the πŸ€— Datasets library", "You need to use `IterableDatasetDict` instead of `DatasetDict` for iterable datasets." ]
2023-05-11T02:09:29
2023-05-25T04:51:59
2023-05-25T04:51:59
NONE
null
null
null
### System Info standard env ### Who can help? _No response_ ### Information - [ ] The official example scripts - [X] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [ ] My own task or dataset (give details below) ### Reproduction Get th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5843/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5843/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5839
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5839/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5839/comments
https://api.github.com/repos/huggingface/datasets/issues/5839/events
https://github.com/huggingface/datasets/issues/5839
1,704,554,718
I_kwDODunzps5lmXDe
5,839
Make models/functions optimized with `torch.compile` hashable
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[ { "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https:...
null
[]
2023-05-10T20:02:08
2023-11-28T16:29:33
2023-11-28T16:29:33
CONTRIBUTOR
null
null
null
As reported in https://github.com/huggingface/datasets/issues/5819, hashing functions/transforms that reference a model, or a function, optimized with `torch.compile` currently fails due to them not being picklable (the concrete error can be found in the linked issue). The solutions to consider: 1. hashing/pickling...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5839/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5839/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5838
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5838/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5838/comments
https://api.github.com/repos/huggingface/datasets/issues/5838/events
https://github.com/huggingface/datasets/issues/5838
1,703,210,848
I_kwDODunzps5lhO9g
5,838
Streaming support for `load_from_disk`
{ "login": "Nilabhra", "id": 5437792, "node_id": "MDQ6VXNlcjU0Mzc3OTI=", "avatar_url": "https://avatars.githubusercontent.com/u/5437792?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Nilabhra", "html_url": "https://github.com/Nilabhra", "followers_url": "https://api.github.com/users/Nilab...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
[ "As the name says, `load_from_disk` load the data from your disk. If the data is hosted on S3, it is first downloaded locally and then loaded from your disk.\r\n\r\nThere is a discussion on streaming data from S3 here though: #5281 ", "@lhoestq \r\nThanks for your comment. I have checked out the discussion before...
2023-05-10T06:25:22
2023-05-12T09:37:45
2023-05-12T09:37:45
NONE
null
null
null
### Feature request Support for streaming datasets stored in object stores in `load_from_disk`. ### Motivation The `load_from_disk` function supports fetching datasets stored in object stores such as `s3`. In many cases, the datasets that are stored in object stores are very large and being able to stream the data ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5838/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5838/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5837
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5837/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5837/comments
https://api.github.com/repos/huggingface/datasets/issues/5837/events
https://github.com/huggingface/datasets/issues/5837
1,703,019,816
I_kwDODunzps5lggUo
5,837
Use DeepSpeed load myself " .csv " dataset.
{ "login": "LanShanPi", "id": 58167546, "node_id": "MDQ6VXNlcjU4MTY3NTQ2", "avatar_url": "https://avatars.githubusercontent.com/u/58167546?v=4", "gravatar_id": "", "url": "https://api.github.com/users/LanShanPi", "html_url": "https://github.com/LanShanPi", "followers_url": "https://api.github.com/users/...
[]
open
false
null
[]
null
[ "Hi ! Doing `load_dataset(\"path/to/data.csv\")` is not supported yet, but you can do\r\n\r\n```python\r\nds = load_dataset(\"csv\", data_files=[\"path/to/data.csv\"])\r\n```", "@lhoestq thank you.", "The other question: \r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n Fil...
2023-05-10T02:39:28
2023-05-15T03:51:36
null
NONE
null
null
null
### Describe the bug When I use DeepSpeed train a model with my own " XXX.csv" dataset I got the follow question: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/fm001/.conda/envs/hzl/lib/python3.8/site-packages/datasets/load.py", line 1767, in load_dataset builder_instan...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5837/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5837/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5834
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5834/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5834/comments
https://api.github.com/repos/huggingface/datasets/issues/5834/events
https://github.com/huggingface/datasets/issues/5834
1,702,448,892
I_kwDODunzps5leU78
5,834
Is uint8 supported?
{ "login": "ryokan0123", "id": 17979572, "node_id": "MDQ6VXNlcjE3OTc5NTcy", "avatar_url": "https://avatars.githubusercontent.com/u/17979572?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ryokan0123", "html_url": "https://github.com/ryokan0123", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "Hi ! The numpy formatting detaults to int64 and float32 - but you can use uint8 using\r\n```python\r\nds = ds.with_format(\"numpy\", dtype=np.uint8)\r\n```", "Related to https://github.com/huggingface/datasets/issues/5517.", "Thank you!\r\nBy setting `ds.with_format(\"numpy\", dtype=np.uint8)`, the dataset ret...
2023-05-09T17:31:13
2023-05-13T05:04:21
2023-05-13T05:04:21
NONE
null
null
null
### Describe the bug I expect the dataset to store the data in the `uint8` data type, but it's returning `int64` instead. While I've found that `datasets` doesn't yet support float16 (https://github.com/huggingface/datasets/issues/4981), I'm wondering if this is the case for other data types as well. Is there a way ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5834/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5834/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5833
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5833/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5833/comments
https://api.github.com/repos/huggingface/datasets/issues/5833/events
https://github.com/huggingface/datasets/issues/5833
1,702,280,682
I_kwDODunzps5ldr3q
5,833
Unable to push dataset - `create_pr` problem
{ "login": "agombert", "id": 17645711, "node_id": "MDQ6VXNlcjE3NjQ1NzEx", "avatar_url": "https://avatars.githubusercontent.com/u/17645711?v=4", "gravatar_id": "", "url": "https://api.github.com/users/agombert", "html_url": "https://github.com/agombert", "followers_url": "https://api.github.com/users/ago...
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Thanks for reporting, @agombert.\r\n\r\nIn this case, I think the root issue is authentication: before pushing to Hub, you should authenticate. See our docs: https://huggingface.co/docs/datasets/upload_dataset#upload-with-python\r\n> 2. To upload a dataset on the Hub in Python, you need to log in to your Hugging F...
2023-05-09T15:32:55
2023-10-24T18:22:29
2023-10-24T18:22:29
NONE
null
null
null
### Describe the bug I can't upload to the hub the dataset I manually created locally (Image dataset). I have a problem when using the method `.push_to_hub` which asks for a `create_pr` attribute which is not compatible. ### Steps to reproduce the bug here what I have: ```python dataset.push_to_hub("agomberto/Fr...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5833/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5833/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5832
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5832/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5832/comments
https://api.github.com/repos/huggingface/datasets/issues/5832/events
https://github.com/huggingface/datasets/issues/5832
1,702,135,336
I_kwDODunzps5ldIYo
5,832
404 Client Error: Not Found for url: https://huggingface.co/api/models/bert-large-cased
{ "login": "varungupta31", "id": 51288316, "node_id": "MDQ6VXNlcjUxMjg4MzE2", "avatar_url": "https://avatars.githubusercontent.com/u/51288316?v=4", "gravatar_id": "", "url": "https://api.github.com/users/varungupta31", "html_url": "https://github.com/varungupta31", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "moved to https://github.com/huggingface/transformers/issues/23233" ]
2023-05-09T14:14:59
2023-05-09T14:25:59
2023-05-09T14:25:59
NONE
null
null
null
### Describe the bug Running [Bert-Large-Cased](https://huggingface.co/bert-large-cased) model causes `HTTPError`, with the following traceback- ``` HTTPError Traceback (most recent call last) <ipython-input-6-5c580443a1ad> in <module> ----> 1 tokenizer = BertTokenizer.from_pretra...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5832/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5832/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5831
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5831/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5831/comments
https://api.github.com/repos/huggingface/datasets/issues/5831/events
https://github.com/huggingface/datasets/issues/5831
1,701,813,835
I_kwDODunzps5lb55L
5,831
[Bug]504 Server Error when loading dataset which was already cached
{ "login": "SingL3", "id": 20473466, "node_id": "MDQ6VXNlcjIwNDczNDY2", "avatar_url": "https://avatars.githubusercontent.com/u/20473466?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SingL3", "html_url": "https://github.com/SingL3", "followers_url": "https://api.github.com/users/SingL3/fo...
[]
open
false
null
[]
null
[ "I am experiencing the same problem with the following environment:\r\n\r\n* `datasets` version: 2.11.0\r\n* Platform: `Linux 5.19.0-41-generic x86_64 GNU/Linux`\r\n* Python version: `3.8.5`\r\n* Huggingface_hub version: 0.13.3\r\n* PyArrow version: `11.0.0`\r\n* Pandas version: `1.5.3`\r\n\r\nTrying to get some di...
2023-05-09T10:31:07
2023-05-10T01:48:20
null
NONE
null
null
null
### Describe the bug I have already cached the dataset using: ``` dataset = load_dataset("databricks/databricks-dolly-15k", cache_dir="/mnt/data/llm/datasets/databricks-dolly-15k") ``` After that, I tried to load it again using the same machine, I got this error: ``` Traceback (most rece...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5831/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5831/timeline
null
reopened
https://api.github.com/repos/huggingface/datasets/issues/5829
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5829/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5829/comments
https://api.github.com/repos/huggingface/datasets/issues/5829/events
https://github.com/huggingface/datasets/issues/5829
1,699,958,189
I_kwDODunzps5lU02t
5,829
(mach-o file, but is an incompatible architecture (have 'arm64', need 'x86_64'))
{ "login": "elcolie", "id": 18206728, "node_id": "MDQ6VXNlcjE4MjA2NzI4", "avatar_url": "https://avatars.githubusercontent.com/u/18206728?v=4", "gravatar_id": "", "url": "https://api.github.com/users/elcolie", "html_url": "https://github.com/elcolie", "followers_url": "https://api.github.com/users/elcoli...
[]
closed
false
null
[]
null
[ "Can you paste the error stack trace?", "That is weird. I can't reproduce it again after reboot.\r\n```python\r\nIn [2]: import platform\r\n\r\nIn [3]: platform.platform()\r\nOut[3]: 'macOS-13.2-arm64-arm-64bit'\r\n\r\nIn [4]: from datasets import load_dataset\r\n ...:\r\n ...: jazzy = load_dataset(\"nomic-ai...
2023-05-08T10:07:14
2023-06-30T11:39:14
2023-05-09T00:46:42
NONE
null
null
null
### Describe the bug M2 MBP can't run ```python from datasets import load_dataset jazzy = load_dataset("nomic-ai/gpt4all-j-prompt-generations", revision='v1.2-jazzy') ``` ### Steps to reproduce the bug 1. Use M2 MBP 2. Python 3.10.10 from pyenv 3. Run ``` from datasets import load_dataset jazzy = load_...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5829/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5829/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5828
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5828/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5828/comments
https://api.github.com/repos/huggingface/datasets/issues/5828/events
https://github.com/huggingface/datasets/issues/5828
1,699,235,739
I_kwDODunzps5lSEeb
5,828
Stream data concatenation issue
{ "login": "krishnapriya-18", "id": 48817796, "node_id": "MDQ6VXNlcjQ4ODE3Nzk2", "avatar_url": "https://avatars.githubusercontent.com/u/48817796?v=4", "gravatar_id": "", "url": "https://api.github.com/users/krishnapriya-18", "html_url": "https://github.com/krishnapriya-18", "followers_url": "https://api...
[]
closed
false
null
[]
null
[ "Hi! \r\n\r\nYou can call `map` as follows to avoid the error:\r\n```python\r\naugmented_dataset_cln = dataset_cln['train'].map(augment_dataset, features=dataset_cln['train'].features)\r\n```", "Thanks it is solved", "Hi! \r\nI have run into the same problem with you. Could you please let me know how you solve ...
2023-05-07T21:02:54
2023-06-29T20:07:56
2023-05-10T05:05:47
NONE
null
null
null
### Describe the bug I am not able to concatenate the augmentation of the stream data. I am using the latest version of dataset. ValueError: The features can't be aligned because the key audio of features {'audio_id': Value(dtype='string', id=None), 'audio': {'array': Sequence(feature=Value(dtype='float32', id=...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5828/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5828/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5827
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5827/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5827/comments
https://api.github.com/repos/huggingface/datasets/issues/5827/events
https://github.com/huggingface/datasets/issues/5827
1,698,891,246
I_kwDODunzps5lQwXu
5,827
load json dataset interrupt when dtype cast problem occured
{ "login": "1014661165", "id": 46060451, "node_id": "MDQ6VXNlcjQ2MDYwNDUx", "avatar_url": "https://avatars.githubusercontent.com/u/46060451?v=4", "gravatar_id": "", "url": "https://api.github.com/users/1014661165", "html_url": "https://github.com/1014661165", "followers_url": "https://api.github.com/use...
[]
open
false
null
[]
null
[ "Indeed the JSON dataset builder raises an error when it encounters an unexpected type.\r\n\r\nThere's an old PR open to add away to ignore such elements though, if it can help: https://github.com/huggingface/datasets/pull/2838" ]
2023-05-07T04:52:09
2023-05-10T12:32:28
null
NONE
null
null
null
### Describe the bug i have a json like this: [ {"id": 1, "name": 1}, {"id": 2, "name": "Nan"}, {"id": 3, "name": 3}, .... ] ,which have several problematic rows data like row 2, then i load it with datasets.load_dataset('json', data_files=['xx.json'], split='train'), it will report like this: ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5827/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5827/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5825
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5825/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5825/comments
https://api.github.com/repos/huggingface/datasets/issues/5825/events
https://github.com/huggingface/datasets/issues/5825
1,697,327,483
I_kwDODunzps5lKyl7
5,825
FileNotFound even though exists
{ "login": "Muennighoff", "id": 62820084, "node_id": "MDQ6VXNlcjYyODIwMDg0", "avatar_url": "https://avatars.githubusercontent.com/u/62820084?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Muennighoff", "html_url": "https://github.com/Muennighoff", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[ "Hi! \r\n\r\nThis would only work if `bigscience/xP3` was a no-code dataset, but it isn't (it has a Python builder script).\r\n\r\nBut this should work: \r\n```python\r\nload_dataset(\"json\", data_files=\"https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_a...
2023-05-05T09:49:55
2023-08-16T10:02:01
2023-08-16T10:02:01
CONTRIBUTOR
null
null
null
### Describe the bug I'm trying to download https://huggingface.co/datasets/bigscience/xP3/resolve/main/ur/xp3_facebook_flores_spa_Latn-urd_Arab_devtest_ab-spa_Latn-urd_Arab.jsonl which works fine in my webbrowser, but somehow not with datasets. Am I doing sth wrong? ``` Downloading builder script: 100% 2.82k/2.8...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5825/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5825/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5823
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5823/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5823/comments
https://api.github.com/repos/huggingface/datasets/issues/5823/events
https://github.com/huggingface/datasets/issues/5823
1,697,024,789
I_kwDODunzps5lJosV
5,823
[2.12.0] DatasetDict.save_to_disk not saving to S3
{ "login": "thejamesmarq", "id": 5233185, "node_id": "MDQ6VXNlcjUyMzMxODU=", "avatar_url": "https://avatars.githubusercontent.com/u/5233185?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thejamesmarq", "html_url": "https://github.com/thejamesmarq", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "Hi ! Can you try adding the `s3://` prefix ?\r\n```python\r\nf\"s3://{s3_bucket}/{s3_dir}/{dataset_name}\"\r\n```", "Ugh, yeah that was it. Thank you!" ]
2023-05-05T05:22:59
2023-05-05T15:01:18
2023-05-05T15:01:17
NONE
null
null
null
### Describe the bug When trying to save a `DatasetDict` to a private S3 bucket using `save_to_disk`, the artifacts are instead saved locally, and not in the S3 bucket. I have tried using the deprecated `fs` as well as the `storage_options` arguments and I get the same results. ### Steps to reproduce the bug 1. C...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5823/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5823/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5822
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5822/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5822/comments
https://api.github.com/repos/huggingface/datasets/issues/5822/events
https://github.com/huggingface/datasets/issues/5822
1,696,627,308
I_kwDODunzps5lIHps
5,822
Audio Dataset with_format torch problem
{ "login": "paulbauriegel", "id": 20282916, "node_id": "MDQ6VXNlcjIwMjgyOTE2", "avatar_url": "https://avatars.githubusercontent.com/u/20282916?v=4", "gravatar_id": "", "url": "https://api.github.com/users/paulbauriegel", "html_url": "https://github.com/paulbauriegel", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "Hi ! Can you try with a more recent version of `datasets` ?", "Ok, yes it worked with the most recent version. Thanks" ]
2023-05-04T20:07:51
2023-05-11T20:45:53
2023-05-11T20:45:53
NONE
null
null
null
### Describe the bug Common Voice v10 Delta (German) Dataset from here https://commonvoice.mozilla.org/de/datasets ``` audio_dataset = \ (Dataset .from_dict({"audio": ('/tmp/cv-corpus-10.0-delta-2022-07-04/de/clips/' + df.path).to_list()}) .cast_column("audio", Audio(sampling_rate=16_000)) .with...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5822/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5822/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5820
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5820/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5820/comments
https://api.github.com/repos/huggingface/datasets/issues/5820/events
https://github.com/huggingface/datasets/issues/5820
1,695,892,811
I_kwDODunzps5lFUVL
5,820
Incomplete docstring for `BuilderConfig`
{ "login": "Laurent2916", "id": 21087104, "node_id": "MDQ6VXNlcjIxMDg3MTA0", "avatar_url": "https://avatars.githubusercontent.com/u/21087104?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Laurent2916", "html_url": "https://github.com/Laurent2916", "followers_url": "https://api.github.com/...
[ { "id": 1935892877, "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue", "name": "good first issue", "color": "7057ff", "default": true, "description": "Good for newcomers" } ]
closed
false
null
[]
null
[ "Thanks for reporting! You are more than welcome to improve `BuilderConfig`'s docstring.\r\n\r\nThis class serves an identical purpose as `tensorflow_datasets`'s `BuilderConfig`, and its docstring is [here](https://github.com/tensorflow/datasets/blob/a95e38b5bb018312c3d3720619c2a8ef83ebf57f/tensorflow_datasets/core...
2023-05-04T12:14:34
2023-05-05T12:31:56
2023-05-05T12:31:56
CONTRIBUTOR
null
null
null
Hi guys ! I stumbled upon this docstring while working on a project. Some of the attributes have missing descriptions. https://github.com/huggingface/datasets/blob/bc5fef5b6d91f009e4101684adcb374df2c170f6/src/datasets/builder.py#L104-L117
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5820/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5820/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5819
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5819/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5819/comments
https://api.github.com/repos/huggingface/datasets/issues/5819/events
https://github.com/huggingface/datasets/issues/5819
1,695,536,738
I_kwDODunzps5lD9Zi
5,819
Cannot pickle error in Dataset.from_generator()
{ "login": "xinghaow99", "id": 50691954, "node_id": "MDQ6VXNlcjUwNjkxOTU0", "avatar_url": "https://avatars.githubusercontent.com/u/50691954?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xinghaow99", "html_url": "https://github.com/xinghaow99", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "Hi! It should work if you put `model = torch.compile(model)` inside the `generate_data` function. If a referenced object is outside, it needs to be pickable, and that's not the case for the compiled models (or functions). ", "> Hi! It should work if you put `model = torch.compile(model)` inside the `generate_da...
2023-05-04T08:39:09
2023-05-05T19:20:59
2023-05-05T19:20:58
NONE
null
null
null
### Describe the bug I'm trying to use Dataset.from_generator() to generate a large dataset. ### Steps to reproduce the bug Code to reproduce: ``` from transformers import T5Tokenizer, T5ForConditionalGeneration, GenerationConfig import torch from tqdm import tqdm from datasets import load_dataset tokenizer...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5819/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5819/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5818
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5818/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5818/comments
https://api.github.com/repos/huggingface/datasets/issues/5818/events
https://github.com/huggingface/datasets/issues/5818
1,695,052,555
I_kwDODunzps5lCHML
5,818
Ability to update a dataset
{ "login": "davidgilbertson", "id": 4443482, "node_id": "MDQ6VXNlcjQ0NDM0ODI=", "avatar_url": "https://avatars.githubusercontent.com/u/4443482?v=4", "gravatar_id": "", "url": "https://api.github.com/users/davidgilbertson", "html_url": "https://github.com/davidgilbertson", "followers_url": "https://api.g...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "This [reply](https://discuss.huggingface.co/t/how-do-i-add-things-rows-to-an-already-saved-dataset/27423) from @mariosasko on the forums may be useful :)", "In this case, I think we can avoid the `PermissionError` by unpacking the underlying `ConcatenationTable` and saving only the newly added data blocks (in ne...
2023-05-04T01:08:13
2023-05-04T20:43:39
null
NONE
null
null
null
### Feature request The ability to load a dataset, add or change something, and save it back to disk. Maybe it's possible, but I can't work out how to do it, e.g. this fails: ```py import datasets dataset = datasets.load_from_disk("data/test1") dataset = dataset.add_item({"text": "A new item"}) dataset.sav...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5818/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5818/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5817
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5817/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5817/comments
https://api.github.com/repos/huggingface/datasets/issues/5817/events
https://github.com/huggingface/datasets/issues/5817
1,694,891,866
I_kwDODunzps5lBf9a
5,817
Setting `num_proc` errors when `.map` returns additional items.
{ "login": "davidgilbertson", "id": 4443482, "node_id": "MDQ6VXNlcjQ0NDM0ODI=", "avatar_url": "https://avatars.githubusercontent.com/u/4443482?v=4", "gravatar_id": "", "url": "https://api.github.com/users/davidgilbertson", "html_url": "https://github.com/davidgilbertson", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "Hi ! Unfortunately I couldn't reproduce on my side locally and with datasets 2.11 and python 3.10.11 on colab.\r\nWhat version of `multiprocess` are you using ?", "I've got `multiprocess` version `0.70.14`.\r\n\r\nI've done some more testing and the error only occurs in PyCharm's Python Console. It seems to be [...
2023-05-03T21:46:53
2023-05-04T21:14:21
2023-05-04T20:22:25
NONE
null
null
null
### Describe the bug I'm using a map function that returns more rows than are passed in. If I try to use `num_proc` I get: ``` File "/home/davidg/.virtualenvs/learning/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 563, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kw...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5817/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5817/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5815
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5815/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5815/comments
https://api.github.com/repos/huggingface/datasets/issues/5815/events
https://github.com/huggingface/datasets/issues/5815
1,693,701,743
I_kwDODunzps5k89Zv
5,815
Easy way to create a Kaggle dataset from a Huggingface dataset?
{ "login": "hrbigelow", "id": 5355286, "node_id": "MDQ6VXNlcjUzNTUyODY=", "avatar_url": "https://avatars.githubusercontent.com/u/5355286?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hrbigelow", "html_url": "https://github.com/hrbigelow", "followers_url": "https://api.github.com/users/hr...
[]
open
false
null
[]
null
[ "Hi @hrbigelow , I'm no expert for such a question so I'll ping @lhoestq from the `datasets` library (also this issue could be moved there if someone with permission can do it :) )", "Hi ! Many datasets are made of several files, and how they are parsed often requires a python script. Because of that, datasets li...
2023-05-02T21:43:33
2023-07-26T16:13:31
null
NONE
null
null
null
I'm not sure whether this is more appropriately addressed with HuggingFace or Kaggle. I would like to somehow directly create a Kaggle dataset from a HuggingFace Dataset. While Kaggle does provide the option to create a dataset from a URI, that URI must point to a single file. For example: ![image](https://user...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5815/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5815/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5812
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5812/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5812/comments
https://api.github.com/repos/huggingface/datasets/issues/5812/events
https://github.com/huggingface/datasets/issues/5812
1,691,798,169
I_kwDODunzps5k1sqZ
5,812
Cannot shuffle interleaved IterableDataset with "all_exhausted" stopping strategy
{ "login": "offchan42", "id": 15215732, "node_id": "MDQ6VXNlcjE1MjE1NzMy", "avatar_url": "https://avatars.githubusercontent.com/u/15215732?v=4", "gravatar_id": "", "url": "https://api.github.com/users/offchan42", "html_url": "https://github.com/offchan42", "followers_url": "https://api.github.com/users/...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 3287858981, "node_id": "MDU6TGFiZWwzMjg3ODU...
closed
false
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[ { "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https:...
null
[]
2023-05-02T05:26:17
2023-05-04T14:24:51
2023-05-04T14:24:51
NONE
null
null
null
### Describe the bug Shuffling interleaved `IterableDataset` with "all_exhausted" strategy yields non-exhaustive sampling. ### Steps to reproduce the bug ```py from datasets import IterableDataset, interleave_datasets def gen(bias, length): for i in range(length): yield dict(a=bias+i) seed = 42 ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5812/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5812/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5811
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5811/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5811/comments
https://api.github.com/repos/huggingface/datasets/issues/5811/events
https://github.com/huggingface/datasets/issues/5811
1,689,919,046
I_kwDODunzps5kuh5G
5,811
load_dataset: TypeError: 'NoneType' object is not callable, on local dataset filename changes
{ "login": "durapensa", "id": 50685483, "node_id": "MDQ6VXNlcjUwNjg1NDgz", "avatar_url": "https://avatars.githubusercontent.com/u/50685483?v=4", "gravatar_id": "", "url": "https://api.github.com/users/durapensa", "html_url": "https://github.com/durapensa", "followers_url": "https://api.github.com/users/...
[]
open
false
null
[]
null
[ "This error means a `DatasetBuilder` subclass that generates the dataset could not be found inside the script, so make sure `dushowxa-characters/dushowxa-characters.py `is a valid dataset script (assuming `path_or_dataset` is `dushowxa-characters`)\r\n\r\nAlso, we should improve the error to make it more obvious wh...
2023-04-30T13:27:17
2023-05-05T17:44:03
null
NONE
null
null
null
### Describe the bug I've adapted Databrick's [train_dolly.py](/databrickslabs/dolly/blob/master/train_dolly.py) to train using a local dataset, which has been working. Upon changing the filenames of the `.json` & `.py` files in my local dataset directory, `dataset = load_dataset(path_or_dataset)["train"]` throws th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5811/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5811/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5809
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5809/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5809/comments
https://api.github.com/repos/huggingface/datasets/issues/5809/events
https://github.com/huggingface/datasets/issues/5809
1,689,797,293
I_kwDODunzps5kuEKt
5,809
wiki_dpr details for Open Domain Question Answering tasks
{ "login": "yulgok22", "id": 64122846, "node_id": "MDQ6VXNlcjY0MTIyODQ2", "avatar_url": "https://avatars.githubusercontent.com/u/64122846?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yulgok22", "html_url": "https://github.com/yulgok22", "followers_url": "https://api.github.com/users/yul...
[]
closed
false
null
[]
null
[ "Hi ! I don't remember exactly how it was done, but maybe you have to embed `f\"{title}<sep>{text}\"` ?\r\n\r\nUsing a HF tokenizer it corresponds to doing\r\n```python\r\ntokenized = tokenizer(titles, texts)\r\n```" ]
2023-04-30T06:12:04
2023-07-21T14:11:00
2023-07-21T14:11:00
NONE
null
null
null
Hey guys! Thanks for creating the wiki_dpr dataset! I am currently trying to combine wiki_dpr and my own datasets. but I don't know how to make the embedding value the same way as wiki_dpr. As an experiment, I embeds the text of id="7" of wiki_dpr, but this result was very different from wiki_dpr.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5809/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5809/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5806
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5806/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5806/comments
https://api.github.com/repos/huggingface/datasets/issues/5806/events
https://github.com/huggingface/datasets/issues/5806
1,688,598,095
I_kwDODunzps5kpfZP
5,806
Return the name of the currently loaded file in the load_dataset function.
{ "login": "s-JoL", "id": 16948304, "node_id": "MDQ6VXNlcjE2OTQ4MzA0", "avatar_url": "https://avatars.githubusercontent.com/u/16948304?v=4", "gravatar_id": "", "url": "https://api.github.com/users/s-JoL", "html_url": "https://github.com/s-JoL", "followers_url": "https://api.github.com/users/s-JoL/follow...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 1935892877, "node_id": "MDU6...
open
false
{ "login": "tsabbir96", "id": 49894149, "node_id": "MDQ6VXNlcjQ5ODk0MTQ5", "avatar_url": "https://avatars.githubusercontent.com/u/49894149?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tsabbir96", "html_url": "https://github.com/tsabbir96", "followers_url": "https://api.github.com/users/...
[ { "login": "tsabbir96", "id": 49894149, "node_id": "MDQ6VXNlcjQ5ODk0MTQ5", "avatar_url": "https://avatars.githubusercontent.com/u/49894149?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tsabbir96", "html_url": "https://github.com/tsabbir96", "followers_url": "https://a...
null
[ "Implementing this makes sense (e.g., `tensorflow_datasets`' imagefolder returns image filenames). Also, in Datasets 3.0, we plan only to store the bytes of an image/audio, not its path, so this feature would be useful when the path info is still needed.", "Hey @mariosasko, Can I work on this issue, this one seem...
2023-04-28T13:50:15
2024-01-21T16:38:29
null
NONE
null
null
null
### Feature request Add an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. ### Motivation When training large language models, machine problems may interrupt...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5806/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5806/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5805
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5805/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5805/comments
https://api.github.com/repos/huggingface/datasets/issues/5805/events
https://github.com/huggingface/datasets/issues/5805
1,688,558,577
I_kwDODunzps5kpVvx
5,805
Improve `Create a dataset` tutorial
{ "login": "polinaeterna", "id": 16348744, "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "gravatar_id": "", "url": "https://api.github.com/users/polinaeterna", "html_url": "https://github.com/polinaeterna", "followers_url": "https://api.github.c...
[ { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
open
false
null
[]
null
[ "I can work on this. The link to the tutorial seems to be broken though @polinaeterna. ", "@isunitha98selvan would be great, thank you! which link are you talking about? I think it should work: https://huggingface.co/docs/datasets/create_dataset" ]
2023-04-28T13:26:22
2023-06-23T14:58:44
null
CONTRIBUTOR
null
null
null
Our [tutorial on how to create a dataset](https://huggingface.co/docs/datasets/create_dataset) is a bit misleading. 1. In **Folder-based builders** section it says that we have two folder-based builders as standard builders, but we also have similar builders (that can be created from directory with data of required f...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5805/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/huggingface/datasets/issues/5805/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5799
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5799/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5799/comments
https://api.github.com/repos/huggingface/datasets/issues/5799/events
https://github.com/huggingface/datasets/issues/5799
1,686,334,572
I_kwDODunzps5kg2xs
5,799
Files downloaded to cache do not respect umask
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[]
2023-04-27T08:06:05
2023-04-27T09:30:17
2023-04-27T09:30:17
MEMBER
null
null
null
As reported by @stas00, files downloaded to the cache do not respect umask: ```bash $ ls -l /path/to/cache/datasets/downloads/ -rw------- 1 uername username 150M Apr 25 16:41 5e646c1d600f065adaeb134e536f6f2f296a6d804bd1f0e1fdcd20ee28c185c6 ``` Related to: - #2065
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5799/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5799/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5798
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5798/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5798/comments
https://api.github.com/repos/huggingface/datasets/issues/5798/events
https://github.com/huggingface/datasets/issues/5798
1,685,904,526
I_kwDODunzps5kfNyO
5,798
Support parallelized downloading and processing in load_dataset with Spark
{ "login": "es94129", "id": 12763339, "node_id": "MDQ6VXNlcjEyNzYzMzM5", "avatar_url": "https://avatars.githubusercontent.com/u/12763339?v=4", "gravatar_id": "", "url": "https://api.github.com/users/es94129", "html_url": "https://github.com/es94129", "followers_url": "https://api.github.com/users/es9412...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Hi ! We're using process pools for parallelism right now. I was wondering if there's a package that implements the same API as a process pool but runs with Spark under the hood ? That or something similar would be cool because users could use whatever distributed framework they want this way.\r\n\r\nFeel free to p...
2023-04-27T00:16:11
2023-05-25T14:11:41
null
CONTRIBUTOR
null
null
null
### Feature request When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes. ```python load_dataset(..., use_spark=True) ``` ### Motivation Further speed up ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5798/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5798/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5797
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5797/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5797/comments
https://api.github.com/repos/huggingface/datasets/issues/5797/events
https://github.com/huggingface/datasets/issues/5797
1,685,501,199
I_kwDODunzps5kdrUP
5,797
load_dataset is case sentitive?
{ "login": "haonan-li", "id": 34729065, "node_id": "MDQ6VXNlcjM0NzI5MDY1", "avatar_url": "https://avatars.githubusercontent.com/u/34729065?v=4", "gravatar_id": "", "url": "https://api.github.com/users/haonan-li", "html_url": "https://github.com/haonan-li", "followers_url": "https://api.github.com/users/...
[]
open
false
null
[]
null
[ "Hi @haonan-li , thank you for the report! It seems to be a bug on the [`huggingface_hub`](https://github.com/huggingface/huggingface_hub) site, there is even no such dataset as `mbzuai/bactrian-x` on the Hub. I opened and [issue](https://github.com/huggingface/huggingface_hub/issues/1453) there.", "I think `loa...
2023-04-26T18:19:04
2023-04-27T11:56:58
null
NONE
null
null
null
### Describe the bug load_dataset() function is case sensitive? ### Steps to reproduce the bug The following two code, get totally different behavior. 1. load_dataset('mbzuai/bactrian-x','en') 2. load_dataset('MBZUAI/Bactrian-X','en') ### Expected behavior Compare 1 and 2. 1 will download all 52 subsets, sh...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5797/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5797/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5794
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5794/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5794/comments
https://api.github.com/repos/huggingface/datasets/issues/5794/events
https://github.com/huggingface/datasets/issues/5794
1,685,196,061
I_kwDODunzps5kcg0d
5,794
CI ZeroDivisionError
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
[]
2023-04-26T14:55:23
2023-04-26T14:55:23
null
MEMBER
null
null
null
Sometimes when running our CI on Windows, we get a ZeroDivisionError: ``` FAILED tests/test_metric_common.py::LocalMetricTest::test_load_metric_frugalscore - ZeroDivisionError: float division by zero ``` See for example: - https://github.com/huggingface/datasets/actions/runs/4809358266/jobs/8560513110 - https:/...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5794/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5794/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5793
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5793/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5793/comments
https://api.github.com/repos/huggingface/datasets/issues/5793/events
https://github.com/huggingface/datasets/issues/5793
1,684,777,320
I_kwDODunzps5ka6lo
5,793
IterableDataset.with_format("torch") not working
{ "login": "jiangwangyi", "id": 39762734, "node_id": "MDQ6VXNlcjM5NzYyNzM0", "avatar_url": "https://avatars.githubusercontent.com/u/39762734?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jiangwangyi", "html_url": "https://github.com/jiangwangyi", "followers_url": "https://api.github.com/...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODk...
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "Hi ! Thanks for reporting, I'm working on it ;)" ]
2023-04-26T10:50:23
2023-06-13T15:57:06
2023-06-13T15:57:06
NONE
null
null
null
### Describe the bug After calling the with_format("torch") method on an IterableDataset instance, the data format is unchanged. ### Steps to reproduce the bug ```python from datasets import IterableDataset def gen(): for i in range(4): yield {"a": [i] * 4} dataset = IterableDataset.from_generator(g...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5793/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5793/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5791
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5791/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5791/comments
https://api.github.com/repos/huggingface/datasets/issues/5791/events
https://github.com/huggingface/datasets/issues/5791
1,683,473,943
I_kwDODunzps5kV8YX
5,791
TIFF/TIF support
{ "login": "sebasmos", "id": 31293221, "node_id": "MDQ6VXNlcjMxMjkzMjIx", "avatar_url": "https://avatars.githubusercontent.com/u/31293221?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sebasmos", "html_url": "https://github.com/sebasmos", "followers_url": "https://api.github.com/users/seb...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
[ "The issue with multichannel TIFF images has already been reported in Pillow (https://github.com/python-pillow/Pillow/issues/1888). We can't do much about it on our side.\r\n\r\nStill, to avoid the error, you can bypass the default Pillow decoding and define a custom one as follows:\r\n```python\r\nimport tifffile ...
2023-04-25T16:14:18
2024-01-15T16:40:33
2024-01-15T16:40:16
NONE
null
null
null
### Feature request I currently have a dataset (with tiff and json files) where I have to do this: `wget path_to_data/images.zip && unzip images.zip` `wget path_to_data/annotations.zip && unzip annotations.zip` Would it make sense a contribution that supports these type of files? ### Motivation instead o...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5791/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5791/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5789
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5789/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5789/comments
https://api.github.com/repos/huggingface/datasets/issues/5789/events
https://github.com/huggingface/datasets/issues/5789
1,682,611,179
I_kwDODunzps5kSpvr
5,789
Support streaming datasets that use jsonlines
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[]
2023-04-25T07:40:02
2023-04-25T07:40:03
null
MEMBER
null
null
null
Extend support for streaming datasets that use `jsonlines.open`. Currently, if `jsonlines` is installed, `datasets` raises a `FileNotFoundError`: ``` FileNotFoundError: [Errno 2] No such file or directory: 'https://...' ``` See: - https://huggingface.co/datasets/masakhane/afriqa/discussions/1
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5789/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5789/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5786
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5786/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5786/comments
https://api.github.com/repos/huggingface/datasets/issues/5786/events
https://github.com/huggingface/datasets/issues/5786
1,680,957,070
I_kwDODunzps5kMV6O
5,786
Multiprocessing in a `filter` or `map` function with a Pytorch model
{ "login": "HugoLaurencon", "id": 44556846, "node_id": "MDQ6VXNlcjQ0NTU2ODQ2", "avatar_url": "https://avatars.githubusercontent.com/u/44556846?v=4", "gravatar_id": "", "url": "https://api.github.com/users/HugoLaurencon", "html_url": "https://github.com/HugoLaurencon", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "Hi ! PyTorch may hang when calling `load_state_dict()` in a subprocess. To fix that, set the multiprocessing start method to \"spawn\". Since `datasets` uses `multiprocess`, you should do:\r\n\r\n```python\r\n# Required to avoid issues with pytorch (otherwise hangs during load_state_dict in multiprocessing)\r\nimp...
2023-04-24T10:38:07
2023-05-30T09:56:30
2023-04-24T10:43:58
MEMBER
null
null
null
### Describe the bug I am trying to use a Pytorch model loaded on CPUs with multiple processes with a `.map` or a `.filter` method. Usually, when dealing with models that are non-pickable, creating a class such that the `map` function is the method `__call__`, and adding `reduce` helps to solve the problem. Howe...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5786/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5786/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5785
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5785/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5785/comments
https://api.github.com/repos/huggingface/datasets/issues/5785/events
https://github.com/huggingface/datasets/issues/5785
1,680,956,964
I_kwDODunzps5kMV4k
5,785
Unsupported data files raise TypeError: 'NoneType' object is not iterable
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[]
2023-04-24T10:38:03
2023-04-27T12:57:30
2023-04-27T12:57:30
MEMBER
null
null
null
Currently, we raise a TypeError for unsupported data files: ``` TypeError: 'NoneType' object is not iterable ``` See: - https://github.com/huggingface/datasets-server/issues/1073 We should give a more informative error message.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5785/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5785/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5783
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5783/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5783/comments
https://api.github.com/repos/huggingface/datasets/issues/5783/events
https://github.com/huggingface/datasets/issues/5783
1,679,664,393
I_kwDODunzps5kHaUJ
5,783
Offset overflow while doing regex on a text column
{ "login": "nishanthcgit", "id": 5066268, "node_id": "MDQ6VXNlcjUwNjYyNjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5066268?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nishanthcgit", "html_url": "https://github.com/nishanthcgit", "followers_url": "https://api.github.com...
[]
open
false
null
[]
null
[ "Hi! This looks like an Arrow bug, but it can be avoided by reducing the `writer_batch_size`.\r\n\r\n(`ds = ds.map(get_text_caption, writer_batch_size=100)` in Colab runs without issues)\r\n", "@mariosasko I ran into this problem with load_dataset. What should I do", "@AisingioroHao0 You can also pass the `wri...
2023-04-22T19:12:03
2023-09-22T06:44:07
null
NONE
null
null
null
### Describe the bug `ArrowInvalid: offset overflow while concatenating arrays` Same error as [here](https://github.com/huggingface/datasets/issues/615) ### Steps to reproduce the bug Steps to reproduce: (dataset is a few GB big so try in colab maybe) ``` import datasets import re ds = datasets.lo...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5783/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5783/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5782
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5782/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5782/comments
https://api.github.com/repos/huggingface/datasets/issues/5782/events
https://github.com/huggingface/datasets/issues/5782
1,679,622,367
I_kwDODunzps5kHQDf
5,782
Support for various audio-loading backends instead of always relying on SoundFile
{ "login": "BoringDonut", "id": 129098876, "node_id": "U_kgDOB7HkfA", "avatar_url": "https://avatars.githubusercontent.com/u/129098876?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BoringDonut", "html_url": "https://github.com/BoringDonut", "followers_url": "https://api.github.com/users/...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
[ "Hi! \r\n\r\nYou can use `set_transform`/`with_transform` to define a custom decoding for audio formats not supported by `soundfile`:\r\n```python\r\naudio_dataset_amr = Dataset.from_dict({\"audio\": [\"audio_samples/audio.amr\"]})\r\n\r\ndef decode_audio(batch):\r\n batch[\"audio\"] = [read_ffmpeg(audio_path) f...
2023-04-22T17:09:25
2023-05-10T20:23:04
2023-05-10T20:23:04
NONE
null
null
null
### Feature request Introduce an option to select from a variety of audio-loading backends rather than solely relying on the SoundFile library. For instance, if the ffmpeg library is installed, it can serve as a fallback loading option. ### Motivation - The SoundFile library, used in [features/audio.py](https://gith...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5782/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5782/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5781
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5781/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5781/comments
https://api.github.com/repos/huggingface/datasets/issues/5781/events
https://github.com/huggingface/datasets/issues/5781
1,679,580,460
I_kwDODunzps5kHF0s
5,781
Error using `load_datasets`
{ "login": "gjyoungjr", "id": 61463108, "node_id": "MDQ6VXNlcjYxNDYzMTA4", "avatar_url": "https://avatars.githubusercontent.com/u/61463108?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gjyoungjr", "html_url": "https://github.com/gjyoungjr", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "It looks like an issue with your installation of scipy, can you try reinstalling it ?", "Sorry for the late reply, but that worked @lhoestq . Thanks for the assist." ]
2023-04-22T15:10:44
2023-05-02T23:41:25
2023-05-02T23:41:25
NONE
null
null
null
### Describe the bug I tried to load a dataset using the `datasets` library in a conda jupyter notebook and got the below error. ``` ImportError: dlopen(/Users/gilbertyoung/miniforge3/envs/review_sense/lib/python3.8/site-packages/scipy/sparse/linalg/_isolve/_iterative.cpython-38-darwin.so, 0x0002): Library not ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5781/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5781/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5780
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5780/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5780/comments
https://api.github.com/repos/huggingface/datasets/issues/5780/events
https://github.com/huggingface/datasets/issues/5780
1,679,367,149
I_kwDODunzps5kGRvt
5,780
TypeError: 'NoneType' object does not support item assignment
{ "login": "v-yunbin", "id": 38179632, "node_id": "MDQ6VXNlcjM4MTc5NjMy", "avatar_url": "https://avatars.githubusercontent.com/u/38179632?v=4", "gravatar_id": "", "url": "https://api.github.com/users/v-yunbin", "html_url": "https://github.com/v-yunbin", "followers_url": "https://api.github.com/users/v-y...
[]
closed
false
null
[]
null
[]
2023-04-22T06:22:43
2023-04-23T08:49:18
2023-04-23T08:49:18
NONE
null
null
null
command: ``` def load_datasets(formats, data_dir=datadir, data_files=datafileοΌ‰οΌš dataset = load_dataset(formats, data_dir=datadir, data_files=datafile, split=split, streaming=True, **kwargs) return dataset raw_datasets = DatasetDict() raw_datasets["train"] = load_datasets(β€œcsv”, args.datadir, "train.csv", s...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5780/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5780/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5778
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5778/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5778/comments
https://api.github.com/repos/huggingface/datasets/issues/5778/events
https://github.com/huggingface/datasets/issues/5778
1,678,125,951
I_kwDODunzps5kBit_
5,778
SchrΓΆdinger's dataset_dict
{ "login": "liujuncn", "id": 902005, "node_id": "MDQ6VXNlcjkwMjAwNQ==", "avatar_url": "https://avatars.githubusercontent.com/u/902005?v=4", "gravatar_id": "", "url": "https://api.github.com/users/liujuncn", "html_url": "https://github.com/liujuncn", "followers_url": "https://api.github.com/users/liujunc...
[]
closed
false
null
[]
null
[ "Hi ! Passing `data_files=\"path/test.json\"` is equivalent to `data_files={\"train\": [\"path/test.json\"]}`, that's why you end up with a train split. If you don't pass `data_files=`, then split names are inferred from the data files names" ]
2023-04-21T08:38:12
2023-07-24T15:15:14
2023-07-24T15:15:14
NONE
null
null
null
### Describe the bug If you use load_dataset('json', data_files="path/test.json"), it will return DatasetDict({train:...}). And if you use load_dataset("path"), it will return DatasetDict({test:...}). Why can't the output behavior be unified? ### Steps to reproduce the bug as description above. ### Expected b...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5778/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5778/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5777
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5777/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5777/comments
https://api.github.com/repos/huggingface/datasets/issues/5777/events
https://github.com/huggingface/datasets/issues/5777
1,677,655,969
I_kwDODunzps5j_v-h
5,777
datasets.load_dataset("code_search_net", "python") : NotADirectoryError: [Errno 20] Not a directory
{ "login": "jason-brian-anderson", "id": 34688597, "node_id": "MDQ6VXNlcjM0Njg4NTk3", "avatar_url": "https://avatars.githubusercontent.com/u/34688597?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jason-brian-anderson", "html_url": "https://github.com/jason-brian-anderson", "followers_url...
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Note:\r\nI listed the datasets and grepped around to find what appears to be an alternative source for this:\r\n\r\nraw_datasets = load_dataset(\"espejelomar/code_search_net_python_10000_examples\", \"python\")", "Thanks for reporting, @jason-brian-anderson.\r\n\r\nYes, this is a known issue: the [CodeSearchNet]...
2023-04-21T02:08:07
2023-06-05T05:49:52
2023-05-11T11:51:56
NONE
null
null
null
### Describe the bug While checking out the [tokenizer tutorial](https://huggingface.co/course/chapter6/2?fw=pt), i noticed getting an error while initially downloading the python dataset used in the examples. The [collab with the error is here](https://colab.research.google.com/github/huggingface/notebooks/blob/ma...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5777/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5777/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5776
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5776/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5776/comments
https://api.github.com/repos/huggingface/datasets/issues/5776/events
https://github.com/huggingface/datasets/issues/5776
1,677,116,100
I_kwDODunzps5j9sLE
5,776
Use Pandas' `read_json` in the JSON builder
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[ { "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https:...
null
[]
2023-04-20T17:15:49
2023-04-20T17:15:49
null
CONTRIBUTOR
null
null
null
Instead of PyArrow's `read_json`, we should use `pd.read_json` in the JSON builder for consistency with the CSV and SQL builders (e.g., to address https://github.com/huggingface/datasets/issues/5725). In Pandas2.0, to get the same performance, we can set the `engine` to "pyarrow". The issue is that Colab still doesn...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5776/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5776/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5775
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5775/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5775/comments
https://api.github.com/repos/huggingface/datasets/issues/5775/events
https://github.com/huggingface/datasets/issues/5775
1,677,089,901
I_kwDODunzps5j9lxt
5,775
ArrowDataset.save_to_disk lost some logic of remote
{ "login": "Zoupers", "id": 29817738, "node_id": "MDQ6VXNlcjI5ODE3NzM4", "avatar_url": "https://avatars.githubusercontent.com/u/29817738?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Zoupers", "html_url": "https://github.com/Zoupers", "followers_url": "https://api.github.com/users/Zouper...
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "We just fixed this on `main` and will do a new release soon :)" ]
2023-04-20T16:58:01
2023-04-26T12:11:36
2023-04-26T12:11:17
NONE
null
null
null
### Describe the bug https://github.com/huggingface/datasets/blob/e7ce0ac60c7efc10886471932854903a7c19f172/src/datasets/arrow_dataset.py#L1371 Here is the bug point, when I want to save from a `DatasetDict` class and the items of the instance is like `[('train', Dataset({features: ..., num_rows: ...}))]` , there ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5775/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5775/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5773
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5773/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5773/comments
https://api.github.com/repos/huggingface/datasets/issues/5773/events
https://github.com/huggingface/datasets/issues/5773
1,675,984,633
I_kwDODunzps5j5X75
5,773
train_dataset does not implement __len__
{ "login": "v-yunbin", "id": 38179632, "node_id": "MDQ6VXNlcjM4MTc5NjMy", "avatar_url": "https://avatars.githubusercontent.com/u/38179632?v=4", "gravatar_id": "", "url": "https://api.github.com/users/v-yunbin", "html_url": "https://github.com/v-yunbin", "followers_url": "https://api.github.com/users/v-y...
[]
open
false
null
[]
null
[ "Thanks for reporting, @v-yunbin.\r\n\r\nCould you please give more details, the steps to reproduce the bug, the complete error back trace and the environment information (`datasets-cli env`)?", "this is a detail error info from transformers:\r\n```\r\nTraceback (most recent call last):\r\n File \"finetune.py\",...
2023-04-20T04:37:05
2023-07-19T20:33:13
null
NONE
null
null
null
when train using data precessored by the datasets, I get follow warning and it leads to that I can not set epoch numbers: `ValueError: The train_dataset does not implement __len__, max_steps has to be specified. The number of steps needs to be known in advance for the learning rate scheduler.`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5773/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5773/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5771
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5771/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5771/comments
https://api.github.com/repos/huggingface/datasets/issues/5771/events
https://github.com/huggingface/datasets/issues/5771
1,674,828,380
I_kwDODunzps5j09pc
5,771
Support cloud storage for loading datasets
{ "login": "eli-osherovich", "id": 2437102, "node_id": "MDQ6VXNlcjI0MzcxMDI=", "avatar_url": "https://avatars.githubusercontent.com/u/2437102?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eli-osherovich", "html_url": "https://github.com/eli-osherovich", "followers_url": "https://api.gith...
[ { "id": 1935892865, "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate", "name": "duplicate", "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists" }, { "id": 1935892871, "...
closed
false
null
[]
null
[ "A duplicate of https://github.com/huggingface/datasets/issues/5281" ]
2023-04-19T12:43:53
2023-05-07T17:47:41
2023-05-07T17:47:41
CONTRIBUTOR
null
null
null
### Feature request It seems that the the current implementation supports cloud storage only for `load_from_disk`. It would be nice if a similar functionality existed in `load_dataset`. ### Motivation Motivation is pretty clear -- let users work with datasets located in the cloud. ### Your contribution ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5771/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5771/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5769
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5769/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5769/comments
https://api.github.com/repos/huggingface/datasets/issues/5769/events
https://github.com/huggingface/datasets/issues/5769
1,673,441,182
I_kwDODunzps5jvq-e
5,769
Tiktoken tokenizers are not pickable
{ "login": "markovalexander", "id": 22663468, "node_id": "MDQ6VXNlcjIyNjYzNDY4", "avatar_url": "https://avatars.githubusercontent.com/u/22663468?v=4", "gravatar_id": "", "url": "https://api.github.com/users/markovalexander", "html_url": "https://github.com/markovalexander", "followers_url": "https://api...
[]
closed
false
null
[]
null
[ "Thanks for reporting, @markovalexander.\r\n\r\nUnfortunately, I'm not able to reproduce the issue: the `tiktoken` tokenizer can be used within `Dataset.map`, both in my local machine and in a Colab notebook: https://colab.research.google.com/drive/1DhJroZgk0sNFJ2Mrz-jYgrmh9jblXaCG?usp=sharing\r\n\r\nAre you sure y...
2023-04-18T16:07:40
2023-05-04T18:55:57
2023-05-04T18:55:57
NONE
null
null
null
### Describe the bug Since tiktoken tokenizer is not pickable, it is not possible to use it inside `dataset.map()` with multiprocessing enabled. However, you [made](https://github.com/huggingface/datasets/issues/5536) tiktoken's tokenizers pickable in `datasets==2.10.0` for caching. For some reason, this logic does no...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5769/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5769/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5768
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5768/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5768/comments
https://api.github.com/repos/huggingface/datasets/issues/5768/events
https://github.com/huggingface/datasets/issues/5768
1,672,494,561
I_kwDODunzps5jsD3h
5,768
load_dataset("squad") doesn't work in 2.7.1 and 2.10.1
{ "login": "yaseen157", "id": 57412770, "node_id": "MDQ6VXNlcjU3NDEyNzcw", "avatar_url": "https://avatars.githubusercontent.com/u/57412770?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yaseen157", "html_url": "https://github.com/yaseen157", "followers_url": "https://api.github.com/users/...
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Thanks for reporting, @yaseen157.\r\n\r\nCould you please give the complete error stack trace?", "I am not able to reproduce your issue: the dataset loads perfectly on my local machine and on a Colab notebook: https://colab.research.google.com/drive/1Fbdoa1JdNz8DOdX6gmIsOK1nCT8Abj4O?usp=sharing\r\n```python\r\nI...
2023-04-18T07:10:56
2023-04-20T10:27:23
2023-04-20T10:27:22
NONE
null
null
null
### Describe the bug There is an issue that seems to be unique to the "squad" dataset, in which it cannot be loaded using standard methods. This issue is most quickly reproduced from the command line, using the HF examples to verify a dataset is loaded properly. This is not a problem with "squad_v2" dataset for e...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5768/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5768/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5767
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5767/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5767/comments
https://api.github.com/repos/huggingface/datasets/issues/5767/events
https://github.com/huggingface/datasets/issues/5767
1,672,433,979
I_kwDODunzps5jr1E7
5,767
How to use Distill-BERT with different datasets?
{ "login": "sauravtii", "id": 109907638, "node_id": "U_kgDOBo0Otg", "avatar_url": "https://avatars.githubusercontent.com/u/109907638?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sauravtii", "html_url": "https://github.com/sauravtii", "followers_url": "https://api.github.com/users/saurav...
[]
closed
false
null
[]
null
[ "Closing this one in favor of the same issue opened in the `transformers` repo." ]
2023-04-18T06:25:12
2023-04-20T16:52:05
2023-04-20T16:52:05
NONE
null
null
null
### Describe the bug - `transformers` version: 4.11.3 - Platform: Linux-5.4.0-58-generic-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyTorch version (GPU?): 1.12.0+cu102 (True) - Tensorflow version (GPU?): 2.10.0 (True) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxL...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5767/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5767/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5766
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5766/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5766/comments
https://api.github.com/repos/huggingface/datasets/issues/5766/events
https://github.com/huggingface/datasets/issues/5766
1,671,485,882
I_kwDODunzps5joNm6
5,766
Support custom feature types
{ "login": "jmontalt", "id": 37540982, "node_id": "MDQ6VXNlcjM3NTQwOTgy", "avatar_url": "https://avatars.githubusercontent.com/u/37540982?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmontalt", "html_url": "https://github.com/jmontalt", "followers_url": "https://api.github.com/users/jmo...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Hi ! Interesting :) What kind of new types would you like to use ?\r\n\r\nNote that you can already implement your own decoding by using `set_transform` that can decode data on-the-fly when rows are accessed", "An interesting proposal indeed. \r\n\r\nPandas and Polars have the \"extension API\", so doing somethi...
2023-04-17T15:46:41
2024-03-10T11:11:22
null
NONE
null
null
null
### Feature request I think it would be nice to allow registering custom feature types with the πŸ€— Datasets library. For example, allow to do something along the following lines: ``` from datasets.features import register_feature_type # this would be a new function @register_feature_type class CustomFeature...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5766/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5766/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5765
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5765/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5765/comments
https://api.github.com/repos/huggingface/datasets/issues/5765/events
https://github.com/huggingface/datasets/issues/5765
1,671,388,824
I_kwDODunzps5jn16Y
5,765
ValueError: You should supply an encoding or a list of encodings to this method that includes input_ids, but you provided ['text']
{ "login": "sauravtii", "id": 109907638, "node_id": "U_kgDOBo0Otg", "avatar_url": "https://avatars.githubusercontent.com/u/109907638?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sauravtii", "html_url": "https://github.com/sauravtii", "followers_url": "https://api.github.com/users/saurav...
[]
open
false
null
[]
null
[ "You need to remove the `text` and `text_en` columns before passing the dataset to the `DataLoader` to avoid this error:\r\n```python\r\ntokenized_datasets = tokenized_datasets.remove_columns([\"text\", \"text_en\"])\r\n```\r\n", "Thanks @mariosasko. Now I am getting this error:\r\n\r\n```\r\nTraceback (most rece...
2023-04-17T15:00:50
2023-04-25T13:50:45
null
NONE
null
null
null
### Describe the bug Following is my code that I am trying to run, but facing an error (have attached the whole error below): My code: ``` from collections import OrderedDict import warnings import flwr as fl import torch import numpy as np import random from torch.utils.data import DataLoader from...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5765/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5765/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5764
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5764/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5764/comments
https://api.github.com/repos/huggingface/datasets/issues/5764/events
https://github.com/huggingface/datasets/issues/5764
1,670,740,198
I_kwDODunzps5jlXjm
5,764
ConnectionError: Couldn't reach https://www.dropbox.com/s/zts98j4vkqtsns6/aclImdb_v2.tar?dl=1
{ "login": "sauravtii", "id": 109907638, "node_id": "U_kgDOBo0Otg", "avatar_url": "https://avatars.githubusercontent.com/u/109907638?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sauravtii", "html_url": "https://github.com/sauravtii", "followers_url": "https://api.github.com/users/saurav...
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Thanks for reporting, @sauravtii.\r\n\r\nUnfortunately, I'm not able to reproduce the issue:\r\n```python\r\nIn [1]: from datasets import load_dataset\r\n\r\nIn [2]: ds = load_dataset(\"josianem/imdb\")\r\n\r\nIn [2]: ds\r\nOut[2]: \r\nDatasetDict({\r\n train: Dataset({\r\n features: ['text', 'label'],\r...
2023-04-17T09:08:18
2023-04-18T07:18:20
2023-04-18T07:18:20
NONE
null
null
null
### Describe the bug I want to use this (https://huggingface.co/datasets/josianem/imdb) dataset therefore I am trying to load it using the following code: ``` dataset = load_dataset("josianem/imdb") ``` The dataset is not getting loaded and gives the error message as the following: ``` Traceback (most rece...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5764/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5764/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5762
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5762/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5762/comments
https://api.github.com/repos/huggingface/datasets/issues/5762/events
https://github.com/huggingface/datasets/issues/5762
1,670,326,470
I_kwDODunzps5jjyjG
5,762
Not able to load the pile
{ "login": "surya-narayanan", "id": 17240858, "node_id": "MDQ6VXNlcjE3MjQwODU4", "avatar_url": "https://avatars.githubusercontent.com/u/17240858?v=4", "gravatar_id": "", "url": "https://api.github.com/users/surya-narayanan", "html_url": "https://github.com/surya-narayanan", "followers_url": "https://api...
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Thanks for reporting, @surya-narayanan.\r\n\r\nI see you already started a discussion about this on the Community tab of the corresponding dataset: https://huggingface.co/datasets/EleutherAI/the_pile/discussions/10\r\nLet's continue the discussion there!" ]
2023-04-17T03:09:10
2023-04-17T09:37:27
2023-04-17T09:37:27
NONE
null
null
null
### Describe the bug Got this error when I am trying to load the pile dataset ``` TypeError: Couldn't cast array of type struct<file: string, id: string> to {'id': Value(dtype='string', id=None)} ``` ### Steps to reproduce the bug Please visit the following sample notebook https://colab.research.goo...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5762/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5762/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5761
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5761/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5761/comments
https://api.github.com/repos/huggingface/datasets/issues/5761/events
https://github.com/huggingface/datasets/issues/5761
1,670,034,582
I_kwDODunzps5jirSW
5,761
One or several metadata.jsonl were found, but not in the same directory or in a parent directory
{ "login": "blghtr", "id": 69686152, "node_id": "MDQ6VXNlcjY5Njg2MTUy", "avatar_url": "https://avatars.githubusercontent.com/u/69686152?v=4", "gravatar_id": "", "url": "https://api.github.com/users/blghtr", "html_url": "https://github.com/blghtr", "followers_url": "https://api.github.com/users/blghtr/fo...
[]
open
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Also, when generated from a zip archive, the dataset contains only a few images. In my case, 20 versus 2000+ contained in the archive. The generation from folders works as expected.", "Thanks for reporting, @blghtr.\r\n\r\nYou should include the `metadata.jsonl` in your ZIP archives, at the root level directory....
2023-04-16T16:21:55
2023-04-19T11:53:24
null
NONE
null
null
null
### Describe the bug An attempt to generate a dataset from a zip archive using imagefolder and metadata.jsonl does not lead to the expected result. Tried all possible locations of the json file: the file in the archive is ignored (generated dataset contains only images), the file next to the archive like [here](http...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5761/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5761/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5760
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5760/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5760/comments
https://api.github.com/repos/huggingface/datasets/issues/5760/events
https://github.com/huggingface/datasets/issues/5760
1,670,028,072
I_kwDODunzps5jipso
5,760
Multi-image loading in Imagefolder dataset
{ "login": "vvvm23", "id": 44398246, "node_id": "MDQ6VXNlcjQ0Mzk4MjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/44398246?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vvvm23", "html_url": "https://github.com/vvvm23", "followers_url": "https://api.github.com/users/vvvm23/fo...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[ { "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https:...
null
[ "Supporting this could be useful (I remember a use-case for this on the Hub). Do you agree @polinaeterna? \r\n\r\nImplementing this should be possible if we iterate over metadata files and build image/audio file paths instead of iterating over image/audio files and looking for the corresponding entries in metadata ...
2023-04-16T16:01:05
2023-11-30T12:06:20
null
NONE
null
null
null
### Feature request Extend the `imagefolder` dataloading script to support loading multiple images per dataset entry. This only really makes sense if a metadata file is present. Currently you can use the following format (example `metadata.jsonl`: ``` {'file_name': 'path_to_image.png', 'metadata': ...} ... `...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5760/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5760/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5759
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5759/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5759/comments
https://api.github.com/repos/huggingface/datasets/issues/5759/events
https://github.com/huggingface/datasets/issues/5759
1,669,977,848
I_kwDODunzps5jidb4
5,759
Can I load in list of list of dict format?
{ "login": "LZY-the-boys", "id": 72137647, "node_id": "MDQ6VXNlcjcyMTM3NjQ3", "avatar_url": "https://avatars.githubusercontent.com/u/72137647?v=4", "gravatar_id": "", "url": "https://api.github.com/users/LZY-the-boys", "html_url": "https://github.com/LZY-the-boys", "followers_url": "https://api.github.c...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Thanks for reporting, @LZY-the-boys.\r\n\r\nCould you please give more details about what is your intended dataset structure? What are the names of the columns and the value of each row?\r\n\r\nCurrently, the JSON-Lines format is supported:\r\n- Each line correspond to one row of the dataset\r\n- Each line is comp...
2023-04-16T13:50:14
2023-04-19T12:04:36
null
NONE
null
null
null
### Feature request my jsonl dataset has following format: ``` [{'input':xxx, 'output':xxx},{'input:xxx,'output':xxx},...] [{'input':xxx, 'output':xxx},{'input:xxx,'output':xxx},...] ``` I try to use `datasets.load_dataset('json', data_files=path)` or `datasets.Dataset.from_json`, it raises ``` File "site-p...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5759/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5759/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5757
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5757/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5757/comments
https://api.github.com/repos/huggingface/datasets/issues/5757/events
https://github.com/huggingface/datasets/issues/5757
1,669,910,503
I_kwDODunzps5jiM_n
5,757
Tilde (~) is not supported
{ "login": "eli-osherovich", "id": 2437102, "node_id": "MDQ6VXNlcjI0MzcxMDI=", "avatar_url": "https://avatars.githubusercontent.com/u/2437102?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eli-osherovich", "html_url": "https://github.com/eli-osherovich", "followers_url": "https://api.gith...
[]
closed
false
null
[]
null
[]
2023-04-16T11:48:10
2023-04-20T15:30:51
2023-04-20T15:30:51
CONTRIBUTOR
null
null
null
### Describe the bug It seems that `~` is not recognized correctly in local paths. Whenever I try to use it I get an exception ### Steps to reproduce the bug ```python load_dataset("imagefolder", data_dir="~/data/my_dataset") ``` Will generate the following error: ``` EmptyDatasetError: The directory at ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5757/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5757/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5756
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5756/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5756/comments
https://api.github.com/repos/huggingface/datasets/issues/5756/events
https://github.com/huggingface/datasets/issues/5756
1,669,678,080
I_kwDODunzps5jhUQA
5,756
Calling shuffle on a IterableDataset with streaming=True, gives "ValueError: cannot reshape array"
{ "login": "rohfle", "id": 21077341, "node_id": "MDQ6VXNlcjIxMDc3MzQx", "avatar_url": "https://avatars.githubusercontent.com/u/21077341?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rohfle", "html_url": "https://github.com/rohfle", "followers_url": "https://api.github.com/users/rohfle/fo...
[]
closed
false
null
[]
null
[ "Hi! I've merged a PR on the Hub with a fix: https://huggingface.co/datasets/fashion_mnist/discussions/3", "Thanks, this appears to have fixed the issue.\r\n\r\nI've created a PR for the same change in the mnist dataset: https://huggingface.co/datasets/mnist/discussions/3/files" ]
2023-04-16T04:59:47
2023-04-18T03:40:56
2023-04-18T03:40:56
NONE
null
null
null
### Describe the bug When calling shuffle on a IterableDataset with streaming=True, I get the following error: ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/administrator/Documents/Projects/huggingface/jax-diffusers-sprint-consistency-models/virtualenv/lib/python3.1...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5756/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5756/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5755
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5755/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5755/comments
https://api.github.com/repos/huggingface/datasets/issues/5755/events
https://github.com/huggingface/datasets/issues/5755
1,669,048,438
I_kwDODunzps5je6h2
5,755
ImportError: cannot import name 'DeprecatedEnum' from 'datasets.utils.deprecation_utils'
{ "login": "fivejjs", "id": 1405491, "node_id": "MDQ6VXNlcjE0MDU0OTE=", "avatar_url": "https://avatars.githubusercontent.com/u/1405491?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fivejjs", "html_url": "https://github.com/fivejjs", "followers_url": "https://api.github.com/users/fivejjs/...
[]
closed
false
null
[]
null
[ "update the version. fix" ]
2023-04-14T23:28:54
2023-04-14T23:36:19
2023-04-14T23:36:19
NONE
null
null
null
### Describe the bug The module moved to new place? ### Steps to reproduce the bug in the import step, ```python from datasets.utils.deprecation_utils import DeprecatedEnum ``` error: ``` ImportError: cannot import name 'DeprecatedEnum' from 'datasets.utils.deprecation_utils' ``` ### Expected behavior...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5755/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5755/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5753
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5753/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5753/comments
https://api.github.com/repos/huggingface/datasets/issues/5753/events
https://github.com/huggingface/datasets/issues/5753
1,668,659,536
I_kwDODunzps5jdblQ
5,753
[IterableDatasets] Add column followed by interleave datasets gives bogus outputs
{ "login": "sanchit-gandhi", "id": 93869735, "node_id": "U_kgDOBZhWpw", "avatar_url": "https://avatars.githubusercontent.com/u/93869735?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sanchit-gandhi", "html_url": "https://github.com/sanchit-gandhi", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "Problem with the code snippet! Using global vars and functions was not a good idea with iterable datasets!\r\n\r\nIf we update to:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\noriginal_dataset = load_dataset(\"librispeech_asr\", \"clean\", split=\"validation\", streaming=True)\r\n\r\n# now add a new co...
2023-04-14T17:32:31
2023-04-14T17:45:52
2023-04-14T17:36:37
CONTRIBUTOR
null
null
null
### Describe the bug If we add a new column to our iterable dataset using the hack described in #5752, when we then interleave datasets the new column is pinned to one value. ### Steps to reproduce the bug What we're going to do here is: 1. Load an iterable dataset in streaming mode (`original_dataset`) 2. A...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5753/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5753/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5752
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5752/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5752/comments
https://api.github.com/repos/huggingface/datasets/issues/5752/events
https://github.com/huggingface/datasets/issues/5752
1,668,574,209
I_kwDODunzps5jdGwB
5,752
Streaming dataset looses `.feature` method after `.add_column`
{ "login": "sanchit-gandhi", "id": 93869735, "node_id": "U_kgDOBZhWpw", "avatar_url": "https://avatars.githubusercontent.com/u/93869735?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sanchit-gandhi", "html_url": "https://github.com/sanchit-gandhi", "followers_url": "https://api.github.com...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
[ "I believe the issue resides in this line:\r\nhttps://github.com/huggingface/datasets/blob/7c3a9b057c476c40d157bd7a5d57f49066239df0/src/datasets/iterable_dataset.py#L1415\r\n\r\nIf we pass the **new** features of the dataset to the `.map` method we can return the features after adding a column, e.g.:\r\n```python\r...
2023-04-14T16:39:50
2024-01-18T10:15:20
null
CONTRIBUTOR
null
null
null
### Describe the bug After appending a new column to a streaming dataset using `.add_column`, we can no longer access the list of dataset features using the `.feature` method. ### Steps to reproduce the bug ```python from datasets import load_dataset original_dataset = load_dataset("librispeech_asr", "clean", sp...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5752/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5752/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5750
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5750/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5750/comments
https://api.github.com/repos/huggingface/datasets/issues/5750/events
https://github.com/huggingface/datasets/issues/5750
1,668,289,067
I_kwDODunzps5jcBIr
5,750
Fail to create datasets from a generator when using Google Big Query
{ "login": "ivanprado", "id": 895720, "node_id": "MDQ6VXNlcjg5NTcyMA==", "avatar_url": "https://avatars.githubusercontent.com/u/895720?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ivanprado", "html_url": "https://github.com/ivanprado", "followers_url": "https://api.github.com/users/ivan...
[]
closed
false
null
[]
null
[ "`from_generator` expects a generator function, not a generator object, so this should work:\r\n```python\r\nfrom datasets import Dataset\r\nfrom google.cloud import bigquery\r\n\r\nclient = bigquery.Client()\r\n\r\ndef gen()\r\n # Perform a query.\r\n QUERY = (\r\n 'SELECT name FROM `bigquery-public-d...
2023-04-14T13:50:59
2023-04-17T12:20:43
2023-04-17T12:20:43
NONE
null
null
null
### Describe the bug Creating a dataset from a generator using `Dataset.from_generator()` fails if the generator is the [Google Big Query Python client](https://cloud.google.com/python/docs/reference/bigquery/latest). The problem is that the Big Query client is not pickable. And the function `create_config_id` tries t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5750/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5750/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5749
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5749/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5749/comments
https://api.github.com/repos/huggingface/datasets/issues/5749/events
https://github.com/huggingface/datasets/issues/5749
1,668,016,321
I_kwDODunzps5ja-jB
5,749
AttributeError: 'Version' object has no attribute 'match'
{ "login": "gulnaz-zh", "id": 54584290, "node_id": "MDQ6VXNlcjU0NTg0Mjkw", "avatar_url": "https://avatars.githubusercontent.com/u/54584290?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gulnaz-zh", "html_url": "https://github.com/gulnaz-zh", "followers_url": "https://api.github.com/users/...
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "I got the same error, and the official website for visual genome is down. Did you solve this problem? ", "I am in the same situation now :( ", "Thanks for reporting, @gulnaz-zh.\r\n\r\nI am investigating it.", "The host server is down: https://visualgenome.org/\r\n\r\nWe are contacting the dataset authors.",...
2023-04-14T10:48:06
2023-06-30T11:31:17
2023-04-18T12:57:08
NONE
null
null
null
### Describe the bug When I run from datasets import load_dataset data = load_dataset("visual_genome", 'region_descriptions_v1.2.0') AttributeError: 'Version' object has no attribute 'match' ### Steps to reproduce the bug from datasets import load_dataset data = load_dataset("visual_genome", 'region_descripti...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5749/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5749/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5744
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5744/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5744/comments
https://api.github.com/repos/huggingface/datasets/issues/5744/events
https://github.com/huggingface/datasets/issues/5744
1,667,076,620
I_kwDODunzps5jXZIM
5,744
[BUG] With Pandas 2.0.0, `load_dataset` raises `TypeError: read_csv() got an unexpected keyword argument 'mangle_dupe_cols'`
{ "login": "keyboardAnt", "id": 15572698, "node_id": "MDQ6VXNlcjE1NTcyNjk4", "avatar_url": "https://avatars.githubusercontent.com/u/15572698?v=4", "gravatar_id": "", "url": "https://api.github.com/users/keyboardAnt", "html_url": "https://github.com/keyboardAnt", "followers_url": "https://api.github.com/...
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Thanks for reporting, @keyboardAnt.\r\n\r\nWe haven't noticed any crash in our CI tests. Could you please indicate specifically the `load_dataset` command that crashes in your side, so that we can reproduce it?", "This has been fixed in `datasets` 2.11", "I am still getting this bug with the latest pandas and ...
2023-04-13T20:21:28
2024-03-16T01:15:27
2023-07-06T17:01:59
NONE
null
null
null
The `load_dataset` function with Pandas `1.5.3` has no issue (just a FutureWarning) but crashes with Pandas `2.0.0`. For your convenience, I opened a draft Pull Request to fix it quickly: https://github.com/huggingface/datasets/pull/5745 --- * The FutureWarning mentioned above: ``` FutureWarning: the 'mangle_...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5744/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5744/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5743
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5743/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5743/comments
https://api.github.com/repos/huggingface/datasets/issues/5743/events
https://github.com/huggingface/datasets/issues/5743
1,666,843,832
I_kwDODunzps5jWgS4
5,743
dataclass.py in virtual environment is overriding the stdlib module "dataclasses"
{ "login": "syedabdullahhassan", "id": 71216295, "node_id": "MDQ6VXNlcjcxMjE2Mjk1", "avatar_url": "https://avatars.githubusercontent.com/u/71216295?v=4", "gravatar_id": "", "url": "https://api.github.com/users/syedabdullahhassan", "html_url": "https://github.com/syedabdullahhassan", "followers_url": "ht...
[]
closed
false
null
[]
null
[ "We no longer depend on `dataclasses` (for almost a year), so I don't think our package is the problematic one. \r\n\r\nI think it makes more sense to raise this issue in the `dataclasses` repo: https://github.com/ericvsmith/dataclasses." ]
2023-04-13T17:28:33
2023-04-17T12:23:18
2023-04-17T12:23:18
NONE
null
null
null
### Describe the bug "e:\Krish_naik\FSDSRegression\venv\Lib\dataclasses.py" is overriding the stdlib module "dataclasses" ### Steps to reproduce the bug module issue ### Expected behavior overriding the stdlib module "dataclasses" ### Environment info VS code
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5743/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5743/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5739
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5739/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5739/comments
https://api.github.com/repos/huggingface/datasets/issues/5739/events
https://github.com/huggingface/datasets/issues/5739
1,663,762,901
I_kwDODunzps5jKwHV
5,739
weird result during dataset split when data path starts with `/data`
{ "login": "ericxsun", "id": 1772912, "node_id": "MDQ6VXNlcjE3NzI5MTI=", "avatar_url": "https://avatars.githubusercontent.com/u/1772912?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ericxsun", "html_url": "https://github.com/ericxsun", "followers_url": "https://api.github.com/users/ericx...
[]
open
false
null
[]
null
[ "Same problem.", "hi! \r\nI think you can run python from `/data/train/raw/` directory and load dataset as `load_dataset(\"code_contests\")` to mitigate this issue as a workaround. \r\n@ericxsun Do you want to open a PR to fix the regex? As you already found the solution :) ", "> hi! I think you can run python ...
2023-04-12T04:51:35
2023-04-21T14:20:59
null
NONE
null
null
null
### Describe the bug The regex defined here https://github.com/huggingface/datasets/blob/f2607935c4e45c70c44fcb698db0363ca7ba83d4/src/datasets/utils/py_utils.py#L158 will cause a weird result during dataset split when data path starts with `/data` ### Steps to reproduce the bug 1. clone dataset into local path ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5739/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5739/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5738
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5738/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5738/comments
https://api.github.com/repos/huggingface/datasets/issues/5738/events
https://github.com/huggingface/datasets/issues/5738
1,663,477,690
I_kwDODunzps5jJqe6
5,738
load_dataset("text","dataset.txt") loads the wrong dataset!
{ "login": "Tylersuard", "id": 41713505, "node_id": "MDQ6VXNlcjQxNzEzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/41713505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Tylersuard", "html_url": "https://github.com/Tylersuard", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "You need to provide a text file as `data_files`, not as a configuration:\r\n\r\n```python\r\nmy_dataset = load_dataset(\"text\", data_files=\"TextFile.txt\")\r\n```\r\n\r\nOtherwise, since `data_files` is `None`, it picks up Colab's sample datasets from the `content` dir." ]
2023-04-12T01:07:46
2023-04-19T12:08:27
2023-04-19T12:08:27
NONE
null
null
null
### Describe the bug I am trying to load my own custom text dataset using the load_dataset function. My dataset is a bunch of ordered text, think along the lines of shakespeare plays. However, after I load the dataset and I inspect it, the dataset is a table with a bunch of latitude and longitude values! What in th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5738/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5738/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5737
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5737/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5737/comments
https://api.github.com/repos/huggingface/datasets/issues/5737/events
https://github.com/huggingface/datasets/issues/5737
1,662,919,811
I_kwDODunzps5jHiSD
5,737
ClassLabel Error
{ "login": "mrcaelumn", "id": 10896776, "node_id": "MDQ6VXNlcjEwODk2Nzc2", "avatar_url": "https://avatars.githubusercontent.com/u/10896776?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mrcaelumn", "html_url": "https://github.com/mrcaelumn", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi, you can use the `cast_column` function to change the feature type from a `Value(int64)` to `ClassLabel`:\r\n\r\n```py\r\ndataset = dataset.cast_column(\"label\", ClassLabel(names=[\"label_1\", \"label_2\", \"label_3\"]))\r\nprint(dataset.features)\r\n{'text': Value(dtype='string', id=None),\r\n 'label': ClassL...
2023-04-11T17:14:13
2023-04-13T16:49:57
2023-04-13T16:49:57
NONE
null
null
null
### Describe the bug I still getting the error "call() takes 1 positional argument but 2 were given" even after ensuring that the value being passed to the label object is a single value and that the ClassLabel object has been created with the correct number of label classes ### Steps to reproduce the bug from...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5737/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5737/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5736
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5736/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5736/comments
https://api.github.com/repos/huggingface/datasets/issues/5736/events
https://github.com/huggingface/datasets/issues/5736
1,662,286,061
I_kwDODunzps5jFHjt
5,736
FORCE_REDOWNLOAD raises "Directory not empty" exception on second run
{ "login": "rcasero", "id": 1219084, "node_id": "MDQ6VXNlcjEyMTkwODQ=", "avatar_url": "https://avatars.githubusercontent.com/u/1219084?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rcasero", "html_url": "https://github.com/rcasero", "followers_url": "https://api.github.com/users/rcasero/...
[]
open
false
null
[]
null
[ "Hi ! I couldn't reproduce your issue :/\r\n\r\nIt seems that `shutil.rmtree` failed. It is supposed to work even if the directory is not empty, but you still end up with `OSError: [Errno 39] Directory not empty:`. Can you make sure another process is not using this directory at the same time ?", "I have the same...
2023-04-11T11:29:15
2023-11-30T07:16:58
null
NONE
null
null
null
### Describe the bug Running `load_dataset(..., download_mode=datasets.DownloadMode.FORCE_REDOWNLOAD)` twice raises a `Directory not empty` exception on the second run. ### Steps to reproduce the bug I cannot test this on datasets v2.11.0 due to #5711, but this happens in v2.10.1. 1. Set up a script `my_dataset.p...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5736/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5736/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5734
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5734/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5734/comments
https://api.github.com/repos/huggingface/datasets/issues/5734/events
https://github.com/huggingface/datasets/issues/5734
1,662,058,028
I_kwDODunzps5jEP4s
5,734
Remove temporary pin of fsspec
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[]
2023-04-11T09:04:17
2023-04-11T11:04:52
2023-04-11T11:04:52
MEMBER
null
null
null
Once root cause is found and fixed, remove the temporary pin introduced by: - #5731
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5734/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5734/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5732
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5732/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5732/comments
https://api.github.com/repos/huggingface/datasets/issues/5732/events
https://github.com/huggingface/datasets/issues/5732
1,662,020,571
I_kwDODunzps5jEGvb
5,732
Enwik8 should support the standard split
{ "login": "lucaslingle", "id": 10287371, "node_id": "MDQ6VXNlcjEwMjg3Mzcx", "avatar_url": "https://avatars.githubusercontent.com/u/10287371?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lucaslingle", "html_url": "https://github.com/lucaslingle", "followers_url": "https://api.github.com/...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
{ "login": "lucaslingle", "id": 10287371, "node_id": "MDQ6VXNlcjEwMjg3Mzcx", "avatar_url": "https://avatars.githubusercontent.com/u/10287371?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lucaslingle", "html_url": "https://github.com/lucaslingle", "followers_url": "https://api.github.com/...
[ { "login": "lucaslingle", "id": 10287371, "node_id": "MDQ6VXNlcjEwMjg3Mzcx", "avatar_url": "https://avatars.githubusercontent.com/u/10287371?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lucaslingle", "html_url": "https://github.com/lucaslingle", "followers_url": "htt...
null
[ "#self-assign", "The Enwik8 pipeline is not present in this codebase, and is hosted elsewhere. I have opened a PR [there](https://huggingface.co/datasets/enwik8/discussions/4) instead. " ]
2023-04-11T08:38:53
2023-04-11T09:28:17
2023-04-11T09:28:16
NONE
null
null
null
### Feature request The HuggingFace Datasets library currently supports two BuilderConfigs for Enwik8. One config yields individual lines as examples, while the other config yields the entire dataset as a single example. Both support only a monolithic split: it is all grouped as "train". The HuggingFace Datasets l...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5732/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5732/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5730
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5730/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5730/comments
https://api.github.com/repos/huggingface/datasets/issues/5730/events
https://github.com/huggingface/datasets/issues/5730
1,662,007,926
I_kwDODunzps5jEDp2
5,730
CI is broken: ValueError: Name (mock) already in the registry and clobber is False
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[]
2023-04-11T08:29:46
2023-04-11T08:47:56
2023-04-11T08:47:56
MEMBER
null
null
null
CI is broken for `test_py310`. See: https://github.com/huggingface/datasets/actions/runs/4665326892/jobs/8258580948 ``` =========================== short test summary info ============================ ERROR tests/test_builder.py::test_builder_with_filesystem_download_and_prepare - ValueError: Name (mock) already ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5730/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5730/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5728
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5728/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5728/comments
https://api.github.com/repos/huggingface/datasets/issues/5728/events
https://github.com/huggingface/datasets/issues/5728
1,661,925,932
I_kwDODunzps5jDvos
5,728
The order of data split names is nondeterministic
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[]
2023-04-11T07:31:25
2023-04-26T15:05:13
2023-04-26T15:05:13
MEMBER
null
null
null
After this CI error: https://github.com/huggingface/datasets/actions/runs/4639528358/jobs/8210492953?pr=5718 ``` FAILED tests/test_data_files.py::test_get_data_files_patterns[data_file_per_split4] - AssertionError: assert ['random', 'train'] == ['train', 'random'] At index 0 diff: 'random' != 'train' Full diff:...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5728/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5728/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5727
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5727/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5727/comments
https://api.github.com/repos/huggingface/datasets/issues/5727/events
https://github.com/huggingface/datasets/issues/5727
1,661,536,363
I_kwDODunzps5jCQhr
5,727
load_dataset fails with FileNotFound error on Windows
{ "login": "joelkowalewski", "id": 122648572, "node_id": "U_kgDOB093_A", "avatar_url": "https://avatars.githubusercontent.com/u/122648572?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joelkowalewski", "html_url": "https://github.com/joelkowalewski", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Hi! Can you please paste the entire error stack trace, not only the last few lines?", "`----> 1 dataset = datasets.load_dataset(\"glue\", \"ax\")\r\n\r\nFile ~\\anaconda3\\envs\\huggingface\\Lib\\site-packages\\datasets\\load.py:1767, in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, ...
2023-04-10T23:21:12
2023-07-21T14:08:20
2023-07-21T14:08:19
NONE
null
null
null
### Describe the bug Although I can import and run the datasets library in a Colab environment, I cannot successfully load any data on my own machine (Windows 10) despite following the install steps: (1) create conda environment (2) activate environment (3) install with: ``conda` install -c huggingface -c conda-...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5727/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5727/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5726
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5726/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5726/comments
https://api.github.com/repos/huggingface/datasets/issues/5726/events
https://github.com/huggingface/datasets/issues/5726
1,660,944,807
I_kwDODunzps5jAAGn
5,726
Fallback JSON Dataset loading does not load all values when features specified manually
{ "login": "myluki2000", "id": 3610788, "node_id": "MDQ6VXNlcjM2MTA3ODg=", "avatar_url": "https://avatars.githubusercontent.com/u/3610788?v=4", "gravatar_id": "", "url": "https://api.github.com/users/myluki2000", "html_url": "https://github.com/myluki2000", "followers_url": "https://api.github.com/users...
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Thanks for reporting, @myluki2000.\r\n\r\nI am working on a fix." ]
2023-04-10T15:22:14
2023-04-21T06:35:28
2023-04-21T06:35:28
NONE
null
null
null
### Describe the bug The fallback JSON dataset loader located here: https://github.com/huggingface/datasets/blob/1c4ec00511868bd881e84a6f7e0333648d833b8e/src/datasets/packaged_modules/json/json.py#L130-L153 does not load the values of features correctly when features are specified manually and not all features...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5726/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5726/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5725
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5725/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5725/comments
https://api.github.com/repos/huggingface/datasets/issues/5725/events
https://github.com/huggingface/datasets/issues/5725
1,660,455,202
I_kwDODunzps5i-Iki
5,725
How to limit the number of examples in dataset, for testing?
{ "login": "ndvbd", "id": 845175, "node_id": "MDQ6VXNlcjg0NTE3NQ==", "avatar_url": "https://avatars.githubusercontent.com/u/845175?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ndvbd", "html_url": "https://github.com/ndvbd", "followers_url": "https://api.github.com/users/ndvbd/followers"...
[]
closed
false
null
[]
null
[ "Hi! You can use the `nrows` parameter for this:\r\n```python\r\ndata = load_dataset(\"json\", data_files=data_path, nrows=10)\r\n```", "@mariosasko I get:\r\n\r\n`TypeError: __init__() got an unexpected keyword argument 'nrows'`", "I misread the format in which the dataset is stored - the `nrows` parameter wo...
2023-04-10T08:41:43
2023-04-21T06:16:24
2023-04-21T06:16:24
NONE
null
null
null
### Describe the bug I am using this command: `data = load_dataset("json", data_files=data_path)` However, I want to add a parameter, to limit the number of loaded examples to be 10, for development purposes, but can't find this simple parameter. ### Steps to reproduce the bug In the description. ### Expected beh...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5725/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5725/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5724
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5724/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5724/comments
https://api.github.com/repos/huggingface/datasets/issues/5724/events
https://github.com/huggingface/datasets/issues/5724
1,659,938,135
I_kwDODunzps5i8KVX
5,724
Error after shuffling streaming IterableDatasets with downloaded dataset
{ "login": "szxiangjn", "id": 41177966, "node_id": "MDQ6VXNlcjQxMTc3OTY2", "avatar_url": "https://avatars.githubusercontent.com/u/41177966?v=4", "gravatar_id": "", "url": "https://api.github.com/users/szxiangjn", "html_url": "https://github.com/szxiangjn", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Moving `\"en\"` to the end of the path instead of passing it as a config name should fix the error:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset('/path/to/your/data/dir/en', streaming=True, split='train')\r\ndataset = dataset.shuffle(buffer_size=10_000, seed=42)\r\nnext(iter(dataset))\r\n```\...
2023-04-09T16:58:44
2023-04-20T20:37:30
2023-04-20T20:37:30
NONE
null
null
null
### Describe the bug I downloaded the C4 dataset, and used streaming IterableDatasets to read it. Everything went normal until I used `dataset = dataset.shuffle(seed=42, buffer_size=10_000)` to shuffle the dataset. Shuffled dataset will throw the following error when it is used by `next(iter(dataset))`: ``` File "/d...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5724/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5724/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5722
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5722/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5722/comments
https://api.github.com/repos/huggingface/datasets/issues/5722/events
https://github.com/huggingface/datasets/issues/5722
1,659,837,510
I_kwDODunzps5i7xxG
5,722
Distributed Training Error on Customized Dataset
{ "login": "wlhgtc", "id": 16603773, "node_id": "MDQ6VXNlcjE2NjAzNzcz", "avatar_url": "https://avatars.githubusercontent.com/u/16603773?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wlhgtc", "html_url": "https://github.com/wlhgtc", "followers_url": "https://api.github.com/users/wlhgtc/fo...
[]
closed
false
null
[]
null
[ "Hmm the error doesn't seem related to data loading.\r\n\r\nRegarding `split_dataset_by_node`: it's generally used to split an iterable dataset (e.g. when streaming) in pytorch DDP. It's not needed if you use a regular dataset since the pytorch DataLoader already assigns a subset of the dataset indices to each node...
2023-04-09T11:04:59
2023-07-24T14:50:46
2023-07-24T14:50:46
NONE
null
null
null
Hi guys, recently I tried to use `datasets` to train a dual encoder. I finish my own datasets according to the nice [tutorial](https://huggingface.co/docs/datasets/v2.11.0/en/dataset_script) Here are my code: ```python class RetrivalDataset(datasets.GeneratorBasedBuilder): """CrossEncoder dataset.""" B...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5722/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5722/timeline
null
completed
https://api.github.com/repos/huggingface/datasets/issues/5721
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5721/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5721/comments
https://api.github.com/repos/huggingface/datasets/issues/5721/events
https://github.com/huggingface/datasets/issues/5721
1,659,680,682
I_kwDODunzps5i7Leq
5,721
Calling datasets.load_dataset("text" ...) results in a wrong split.
{ "login": "cyrilzakka", "id": 1841186, "node_id": "MDQ6VXNlcjE4NDExODY=", "avatar_url": "https://avatars.githubusercontent.com/u/1841186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cyrilzakka", "html_url": "https://github.com/cyrilzakka", "followers_url": "https://api.github.com/users...
[]
open
false
null
[]
null
[]
2023-04-08T23:55:12
2023-04-08T23:55:12
null
NONE
null
null
null
### Describe the bug When creating a text dataset, the training split should have the bulk of the examples by default. Currently, testing does. ### Steps to reproduce the bug I have a folder with 18K text files in it. Each text file essentially consists in a document or article scraped from online. Calling the follo...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5721/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5721/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5720
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5720/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5720/comments
https://api.github.com/repos/huggingface/datasets/issues/5720/events
https://github.com/huggingface/datasets/issues/5720
1,659,610,705
I_kwDODunzps5i66ZR
5,720
Streaming IterableDatasets do not work with torch DataLoaders
{ "login": "jlehrer1", "id": 29244648, "node_id": "MDQ6VXNlcjI5MjQ0NjQ4", "avatar_url": "https://avatars.githubusercontent.com/u/29244648?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jlehrer1", "html_url": "https://github.com/jlehrer1", "followers_url": "https://api.github.com/users/jle...
[]
open
false
null
[]
null
[ "Edit: This behavior is true even without `.take/.set`", "I'm experiencing the same problem that @jlehrer1. I was able to reproduce it with a very small example:\r\n\r\n```py\r\nfrom datasets import Dataset, load_dataset, load_dataset_builder\r\nfrom torch.utils.data import DataLoader\r\n\r\n\r\ndef my_gen():\r\n...
2023-04-08T18:45:48
2023-05-27T12:57:08
null
NONE
null
null
null
### Describe the bug When using streaming datasets set up with train/val split using `.skip()` and `.take()`, the following error occurs when iterating over a torch dataloader: ``` File "/Users/julian/miniconda3/envs/sims/lib/python3.9/site-packages/torch/utils/data/dataloader.py", line 363, in __iter__ self....
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5720/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5720/timeline
null
null
https://api.github.com/repos/huggingface/datasets/issues/5719
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5719/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5719/comments
https://api.github.com/repos/huggingface/datasets/issues/5719/events
https://github.com/huggingface/datasets/issues/5719
1,659,203,222
I_kwDODunzps5i5W6W
5,719
Array2D feature creates a list of list instead of a numpy array
{ "login": "offchan42", "id": 15215732, "node_id": "MDQ6VXNlcjE1MjE1NzMy", "avatar_url": "https://avatars.githubusercontent.com/u/15215732?v=4", "gravatar_id": "", "url": "https://api.github.com/users/offchan42", "html_url": "https://github.com/offchan42", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi! \r\n\r\nYou need to set the format to `np` before indexing the dataset to get NumPy arrays:\r\n```python\r\nfeatures = Features(dict(seq=Array2D((2,2), 'float32'))) \r\nds = Dataset.from_dict(dict(seq=[np.random.rand(2,2)]), features=features)\r\nds.set_format(\"np\")\r\na = ds[0]['seq']\r\n```\r\n\r\n> I th...
2023-04-07T21:04:08
2023-04-20T15:34:41
2023-04-20T15:34:41
NONE
null
null
null
### Describe the bug I'm not sure if this is expected behavior or not. When I create a 2D array using `Array2D`, the data has list type instead of numpy array. I think it should not be the expected behavior especially when I feed a numpy array as input to the data creation function. Why is it converting my array int...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/5719/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/5719/timeline
null
completed