url stringlengths 58 61 | repository_url stringclasses 1
value | labels_url stringlengths 72 75 | comments_url stringlengths 67 70 | events_url stringlengths 65 68 | html_url stringlengths 48 51 | id int64 600M 2.19B | node_id stringlengths 18 24 | number int64 2 6.73k | title stringlengths 1 290 | user dict | labels listlengths 0 4 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 4 | milestone dict | comments listlengths 0 30 | created_at timestamp[s] | updated_at timestamp[s] | closed_at timestamp[s] | author_association stringclasses 3
values | active_lock_reason null | draft null | pull_request null | body stringlengths 0 228k ⌀ | reactions dict | timeline_url stringlengths 67 70 | performed_via_github_app null | state_reason stringclasses 3
values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/6624 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6624/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6624/comments | https://api.github.com/repos/huggingface/datasets/issues/6624/events | https://github.com/huggingface/datasets/issues/6624 | 2,103,950,718 | I_kwDODunzps59Z71- | 6,624 | How to download the laion-coco dataset | {
"login": "vanpersie32",
"id": 15981416,
"node_id": "MDQ6VXNlcjE1OTgxNDE2",
"avatar_url": "https://avatars.githubusercontent.com/u/15981416?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vanpersie32",
"html_url": "https://github.com/vanpersie32",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | [
"Hi, this dataset has been disabled by the authors, so unfortunately it's no longer possible to download it."
] | 2024-01-28T03:56:05 | 2024-02-06T09:43:31 | 2024-02-06T09:43:31 | NONE | null | null | null | The laion coco dataset is not available now. How to download it
https://huggingface.co/datasets/laion/laion-coco | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6624/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6624/timeline | null | not_planned |
https://api.github.com/repos/huggingface/datasets/issues/6623 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6623/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6623/comments | https://api.github.com/repos/huggingface/datasets/issues/6623/events | https://github.com/huggingface/datasets/issues/6623 | 2,103,870,123 | I_kwDODunzps59ZoKr | 6,623 | streaming datasets doesn't work properly with multi-node | {
"login": "rohitgr7",
"id": 30778939,
"node_id": "MDQ6VXNlcjMwNzc4OTM5",
"avatar_url": "https://avatars.githubusercontent.com/u/30778939?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rohitgr7",
"html_url": "https://github.com/rohitgr7",
"followers_url": "https://api.github.com/users/roh... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | [
"@mariosasko, @lhoestq, @albertvillanova\r\nhey guys! can anyone help? or can you guys suggest who can help with this?",
"Hi ! \r\n\r\n1. When the dataset is running of of examples, the last batches received by the GPU can be incomplete or empty/missing. We haven't implemented yet a way to ignore the last batch. ... | 2024-01-27T23:46:13 | 2024-03-08T14:27:08 | null | NONE | null | null | null | ### Feature request
Let’s say I have a dataset with 5 samples with values [1, 2, 3, 4, 5], with 2 GPUs (for DDP) and batch size of 2. This dataset is an `IterableDataset` since I am streaming it.
Now I split the dataset using `split_dataset_by_node` to ensure it doesn’t get repeated. And since it’s already splitt... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6623/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6623/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6622 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6622/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6622/comments | https://api.github.com/repos/huggingface/datasets/issues/6622/events | https://github.com/huggingface/datasets/issues/6622 | 2,103,780,697 | I_kwDODunzps59ZSVZ | 6,622 | multi-GPU map does not work | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | [
"This should now be fixed by https://github.com/huggingface/datasets/pull/6550 and updated with https://github.com/huggingface/datasets/pull/6646\r\n\r\nFeel free to re-open if you're still having issues :)"
] | 2024-01-27T20:06:08 | 2024-02-08T11:18:21 | 2024-02-08T11:18:21 | NONE | null | null | null | ### Describe the bug
Here is the code for single-GPU processing: https://pastebin.com/bfmEeK2y
Here is the code for multi-GPU processing: https://pastebin.com/gQ7i5AQy
Here is the video showing that the multi-GPU mapping does not work as expected (there are so many things wrong here, it's better to watch the 3-min... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6622/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6622/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6621 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6621/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6621/comments | https://api.github.com/repos/huggingface/datasets/issues/6621/events | https://github.com/huggingface/datasets/issues/6621 | 2,103,675,294 | I_kwDODunzps59Y4me | 6,621 | deleted | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | [] | 2024-01-27T16:59:58 | 2024-01-27T17:14:43 | 2024-01-27T17:14:43 | NONE | null | null | null | ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6621/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6621/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6620 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6620/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6620/comments | https://api.github.com/repos/huggingface/datasets/issues/6620/events | https://github.com/huggingface/datasets/issues/6620 | 2,103,110,536 | I_kwDODunzps59WuuI | 6,620 | wiki_dpr.py error (ID mismatch between lines {id} and vector {vec_id} | {
"login": "kiehls90",
"id": 101498700,
"node_id": "U_kgDOBgy_TA",
"avatar_url": "https://avatars.githubusercontent.com/u/101498700?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kiehls90",
"html_url": "https://github.com/kiehls90",
"followers_url": "https://api.github.com/users/kiehls90/... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Thanks for reporting, @kiehls90.\r\n\r\nAs this seems an issue with the specific \"wiki_dpr\" dataset, I am transferring the issue to the corresponding dataset page: https://huggingface.co/datasets/wiki_dpr/discussions/13"
] | 2024-01-27T01:00:09 | 2024-02-06T09:40:19 | 2024-02-06T09:40:19 | NONE | null | null | null | ### Describe the bug
I'm trying to run a rag example, and the dataset is wiki_dpr.
wiki_dpr download and extracting have been completed successfully.
However, at the generating train split stage, an error from wiki_dpr.py keeps popping up.
Especially in "_generate_examples" :
1. The following error occurs in the... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6620/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6620/timeline | null | not_planned |
https://api.github.com/repos/huggingface/datasets/issues/6618 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6618/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6618/comments | https://api.github.com/repos/huggingface/datasets/issues/6618/events | https://github.com/huggingface/datasets/issues/6618 | 2,101,868,198 | I_kwDODunzps59R_am | 6,618 | While importing load_dataset from datasets | {
"login": "Era-cell",
"id": 77973415,
"node_id": "MDQ6VXNlcjc3OTczNDE1",
"avatar_url": "https://avatars.githubusercontent.com/u/77973415?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Era-cell",
"html_url": "https://github.com/Era-cell",
"followers_url": "https://api.github.com/users/Era... | [] | closed | false | null | [] | null | [
"Hi! Can you please share the error's stack trace so we can see where it comes from?",
"We cannot reproduce the issue and we do not have enough information: environment info (need to run `datasets-cli env`), stack trace,...\r\n\r\nI am closing the issue. Feel free to reopen it (with additional information) if the... | 2024-01-26T09:21:57 | 2024-02-06T10:57:01 | 2024-02-06T09:25:54 | NONE | null | null | null | ### Describe the bug
cannot import name 'DEFAULT_CIPHERS' from 'urllib3.util.ssl_' this is the error i received
### Steps to reproduce the bug
from datasets import load_dataset
### Expected behavior
No errors
### Environment info
python 3.11.5 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6618/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6618/timeline | null | not_planned |
https://api.github.com/repos/huggingface/datasets/issues/6615 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6615/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6615/comments | https://api.github.com/repos/huggingface/datasets/issues/6615/events | https://github.com/huggingface/datasets/issues/6615 | 2,098,951,409 | I_kwDODunzps59G3Tx | 6,615 | ... | {
"login": "ftkeys",
"id": 22179777,
"node_id": "MDQ6VXNlcjIyMTc5Nzc3",
"avatar_url": "https://avatars.githubusercontent.com/u/22179777?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ftkeys",
"html_url": "https://github.com/ftkeys",
"followers_url": "https://api.github.com/users/ftkeys/fo... | [] | closed | false | null | [] | null | [
"Sorry I posted in the wrong repo, please delete.. thanks!"
] | 2024-01-24T19:37:03 | 2024-01-24T19:42:30 | 2024-01-24T19:40:11 | NONE | null | null | null | ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6615/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6615/timeline | null | not_planned |
https://api.github.com/repos/huggingface/datasets/issues/6614 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6614/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6614/comments | https://api.github.com/repos/huggingface/datasets/issues/6614/events | https://github.com/huggingface/datasets/issues/6614 | 2,098,884,520 | I_kwDODunzps59Gm-o | 6,614 | `datasets/downloads` cleanup tool | {
"login": "stas00",
"id": 10676103,
"node_id": "MDQ6VXNlcjEwNjc2MTAz",
"avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/stas00",
"html_url": "https://github.com/stas00",
"followers_url": "https://api.github.com/users/stas00/fo... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | [] | 2024-01-24T18:52:10 | 2024-01-24T18:55:09 | null | CONTRIBUTOR | null | null | null | ### Feature request
Splitting off https://github.com/huggingface/huggingface_hub/issues/1997 - currently `huggingface-cli delete-cache` doesn't take care of cleaning `datasets` temp files
e.g. I discovered having millions of files under `datasets/downloads` cache, I had to do:
```
sudo find /data/huggingface/... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6614/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6614/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6612 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6612/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6612/comments | https://api.github.com/repos/huggingface/datasets/issues/6612/events | https://github.com/huggingface/datasets/issues/6612 | 2,098,078,210 | I_kwDODunzps59DiIC | 6,612 | cnn_dailymail repeats itself | {
"login": "KeremZaman",
"id": 8274752,
"node_id": "MDQ6VXNlcjgyNzQ3NTI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8274752?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/KeremZaman",
"html_url": "https://github.com/KeremZaman",
"followers_url": "https://api.github.com/users... | [] | closed | false | null | [] | null | [
"Hi ! We recently updated `cnn_dailymail` and now `datasets>=2.14` is needed to load it.\r\n\r\nYou can update `datasets` with\r\n\r\n```\r\npip install -U datasets\r\n```"
] | 2024-01-24T11:38:25 | 2024-02-01T08:14:50 | 2024-02-01T08:14:50 | NONE | null | null | null | ### Describe the bug
When I try to load `cnn_dailymail` dataset, it takes longer than usual and when I checked the dataset it's 3x bigger than it's supposed to be.
Check https://huggingface.co/datasets/cnn_dailymail: it says 287k rows for train. But when I check length of train split it says 861339.
Also I che... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6612/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6612/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6611 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6611/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6611/comments | https://api.github.com/repos/huggingface/datasets/issues/6611/events | https://github.com/huggingface/datasets/issues/6611 | 2,096,004,858 | I_kwDODunzps587n76 | 6,611 | `load_from_disk` with large dataset from S3 runs into `botocore.exceptions.ClientError` | {
"login": "zotroneneis",
"id": 15320635,
"node_id": "MDQ6VXNlcjE1MzIwNjM1",
"avatar_url": "https://avatars.githubusercontent.com/u/15320635?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zotroneneis",
"html_url": "https://github.com/zotroneneis",
"followers_url": "https://api.github.com/... | [] | open | false | null | [] | null | [] | 2024-01-23T12:37:57 | 2024-01-23T12:37:57 | null | NONE | null | null | null | ### Describe the bug
When loading a large dataset (>1000GB) from S3 I run into the following error:
```
Traceback (most recent call last):
File "/home/alp/.local/lib/python3.10/site-packages/s3fs/core.py", line 113, in _error_wrapper
return await func(*args, **kwargs)
File "/home/alp/.local/lib/python3.... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6611/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6611/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6610 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6610/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6610/comments | https://api.github.com/repos/huggingface/datasets/issues/6610/events | https://github.com/huggingface/datasets/issues/6610 | 2,095,643,711 | I_kwDODunzps586Pw_ | 6,610 | cast_column to Sequence(subfeatures_dict) has err | {
"login": "neiblegy",
"id": 16574677,
"node_id": "MDQ6VXNlcjE2NTc0Njc3",
"avatar_url": "https://avatars.githubusercontent.com/u/16574677?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/neiblegy",
"html_url": "https://github.com/neiblegy",
"followers_url": "https://api.github.com/users/nei... | [] | closed | false | null | [] | null | [
"Hi! You are passing the wrong feature type to `cast_column`. This is the fixed call:\r\n```python\r\nais_dataset = ais_dataset.cast_column(\"my_labeled_bbox\", {\"bbox\": Sequence(Value(dtype=\"int64\")), \"label\": ClassLabel(names=[\"cat\", \"dog\"])})\r\n```",
"> Hi! You are passing the wrong feature type to ... | 2024-01-23T09:32:32 | 2024-01-25T02:15:23 | 2024-01-25T02:15:23 | NONE | null | null | null | ### Describe the bug
I am working with the following demo code:
```
from datasets import load_dataset
from datasets.features import Sequence, Value, ClassLabel, Features
ais_dataset = load_dataset("/data/ryan.gao/ais_dataset_cache/raw/1978/")
ais_dataset = ais_dataset["train"]
def add_class(example):
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6610/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6610/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6609 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6609/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6609/comments | https://api.github.com/repos/huggingface/datasets/issues/6609/events | https://github.com/huggingface/datasets/issues/6609 | 2,095,085,650 | I_kwDODunzps584HhS | 6,609 | Wrong path for cache directory in offline mode | {
"login": "je-santos",
"id": 42117435,
"node_id": "MDQ6VXNlcjQyMTE3NDM1",
"avatar_url": "https://avatars.githubusercontent.com/u/42117435?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/je-santos",
"html_url": "https://github.com/je-santos",
"followers_url": "https://api.github.com/users/... | [] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"+1",
"same error in 2.16.1",
"@kongjiellx any luck with the issue?",
"I opened https://github.com/huggingface/datasets/pull/6632 to fix this issue. Once it's merged we'll do a new release of `datasets`",
"Thanks @lhoestq !"
] | 2024-01-23T01:47:19 | 2024-02-06T17:21:25 | 2024-02-06T17:21:25 | NONE | null | null | null | ### Describe the bug
Dear huggingfacers,
I'm trying to use a subset of the-stack dataset. When I run the command the first time
```
dataset = load_dataset(
path='bigcode/the-stack',
data_dir='data/fortran',
split='train' )
```
It downloads the files and caches them normally.
Nevertheless, ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6609/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6609/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6605 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6605/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6605/comments | https://api.github.com/repos/huggingface/datasets/issues/6605/events | https://github.com/huggingface/datasets/issues/6605 | 2,090,188,376 | I_kwDODunzps58lb5Y | 6,605 | ELI5 no longer available, but referenced in example code | {
"login": "drdsgvo",
"id": 81480344,
"node_id": "MDQ6VXNlcjgxNDgwMzQ0",
"avatar_url": "https://avatars.githubusercontent.com/u/81480344?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/drdsgvo",
"html_url": "https://github.com/drdsgvo",
"followers_url": "https://api.github.com/users/drdsgv... | [] | closed | false | null | [] | null | [
"Addressed in https://github.com/huggingface/transformers/pull/28715."
] | 2024-01-19T10:21:52 | 2024-02-01T17:58:23 | 2024-02-01T17:58:22 | NONE | null | null | null | Here, an example code is given:
https://huggingface.co/docs/transformers/tasks/language_modeling
This code + article references the ELI5 dataset.
ELI5 is no longer available, as the ELI5 dataset page states: https://huggingface.co/datasets/eli5
"Defunct: Dataset "eli5" is defunct and no longer accessible due to u... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6605/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6605/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6604 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6604/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6604/comments | https://api.github.com/repos/huggingface/datasets/issues/6604/events | https://github.com/huggingface/datasets/issues/6604 | 2,089,713,945 | I_kwDODunzps58joEZ | 6,604 | Transform fingerprint collisions due to setting fixed random seed | {
"login": "normster",
"id": 6687910,
"node_id": "MDQ6VXNlcjY2ODc5MTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/6687910?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/normster",
"html_url": "https://github.com/normster",
"followers_url": "https://api.github.com/users/norms... | [] | closed | false | null | [] | null | [
"I've opened a PR with a fix.",
"I don't think the PR fixes the root cause, since it still relies on the `random` library which will often have its seed fixed. I think the builtin `uuid.uuid4()` is a better choice: https://docs.python.org/3/library/uuid.html"
] | 2024-01-19T06:32:25 | 2024-01-26T15:05:35 | 2024-01-26T15:05:35 | NONE | null | null | null | ### Describe the bug
The transform fingerprinting logic relies on the `random` library for random bits when the function is not hashable (e.g. bound methods as used in `trl`: https://github.com/huggingface/trl/blob/main/trl/trainer/dpo_trainer.py#L356). This causes collisions when the training code sets a fixed random... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6604/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6604/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6603 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6603/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6603/comments | https://api.github.com/repos/huggingface/datasets/issues/6603/events | https://github.com/huggingface/datasets/issues/6603 | 2,089,230,766 | I_kwDODunzps58hyGu | 6,603 | datasets map `cache_file_name` does not work | {
"login": "ChenchaoZhao",
"id": 35147961,
"node_id": "MDQ6VXNlcjM1MTQ3OTYx",
"avatar_url": "https://avatars.githubusercontent.com/u/35147961?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ChenchaoZhao",
"html_url": "https://github.com/ChenchaoZhao",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | null | [
"Unfortunately, I'm unable to reproduce this error. Can you share the reproducer?",
"```\r\nds = datasets.Dataset.from_dict(dict(a=[i for i in range(100)]))\r\nds.map(lambda item: dict(b=item['a'] * 2), cache_file_name=\"/tmp/whatever-fn\") # this worked\r\nds.map(lambda item: dict(b=item['a'] * 2), cache_file_na... | 2024-01-18T23:08:30 | 2024-01-28T04:01:15 | null | NONE | null | null | null | ### Describe the bug
In the documentation `datasets.Dataset.map` arg `cache_file_name` is said to be a string, but it doesn't work.
### Steps to reproduce the bug
1. pick a dataset
2. write a map function
3. do `ds.map(..., cache_file_name='some_filename')`
4. it crashes
### Expected behavior
It will tell you t... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6603/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6603/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6602 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6602/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6602/comments | https://api.github.com/repos/huggingface/datasets/issues/6602/events | https://github.com/huggingface/datasets/issues/6602 | 2,089,217,483 | I_kwDODunzps58hu3L | 6,602 | Index error when data is large | {
"login": "ChenchaoZhao",
"id": 35147961,
"node_id": "MDQ6VXNlcjM1MTQ3OTYx",
"avatar_url": "https://avatars.githubusercontent.com/u/35147961?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ChenchaoZhao",
"html_url": "https://github.com/ChenchaoZhao",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | null | [] | 2024-01-18T23:00:47 | 2024-01-18T23:00:47 | null | NONE | null | null | null | ### Describe the bug
At `save_to_disk` step, the `max_shard_size` by default is `500MB`. However, one row of the dataset might be larger than `500MB` then the saving will throw an index error. Without looking at the source code, the bug is due to wrong calculation of number of shards which i think is
`total_size / m... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6602/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6602/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6600 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6600/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6600/comments | https://api.github.com/repos/huggingface/datasets/issues/6600/events | https://github.com/huggingface/datasets/issues/6600 | 2,088,446,385 | I_kwDODunzps58eymx | 6,600 | Loading CSV exported dataset has unexpected format | {
"login": "OrianeN",
"id": 59572247,
"node_id": "MDQ6VXNlcjU5NTcyMjQ3",
"avatar_url": "https://avatars.githubusercontent.com/u/59572247?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/OrianeN",
"html_url": "https://github.com/OrianeN",
"followers_url": "https://api.github.com/users/Oriane... | [] | open | false | null | [] | null | [
"Hi! Parquet is the only format that supports complex/nested features such as `Translation`. So, this should work:\r\n```python\r\ntest_dataset = load_dataset(\"opus100\", name=\"en-fr\", split=\"test\")\r\n\r\n# Save with .to_parquet()\r\ntest_parquet_path = \"try_testset_save.parquet\"\r\ntest_dataset.to_parquet(... | 2024-01-18T14:48:27 | 2024-01-23T14:42:32 | null | NONE | null | null | null | ### Describe the bug
I wanted to be able to save a HF dataset for translations and load it again in another script, but I'm a bit confused with the documentation and the result I've got so I'm opening this issue to ask if this behavior is as expected.
### Steps to reproduce the bug
The documentation I've mainly cons... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6600/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6600/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6599 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6599/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6599/comments | https://api.github.com/repos/huggingface/datasets/issues/6599/events | https://github.com/huggingface/datasets/issues/6599 | 2,086,684,664 | I_kwDODunzps58YEf4 | 6,599 | Easy way to segment into 30s snippets given an m4a file and a vtt file | {
"login": "RonanKMcGovern",
"id": 78278410,
"node_id": "MDQ6VXNlcjc4Mjc4NDEw",
"avatar_url": "https://avatars.githubusercontent.com/u/78278410?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RonanKMcGovern",
"html_url": "https://github.com/RonanKMcGovern",
"followers_url": "https://api.gi... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | null | [] | null | [
"Hi! Non-generic data processing is out of this library's scope, so it's downstream libraries/users' responsibility to implement such logic.",
"That's fair. Thanks"
] | 2024-01-17T17:51:40 | 2024-01-23T10:42:17 | 2024-01-22T15:35:49 | NONE | null | null | null | ### Feature request
Uploading datasets is straightforward thanks to the ability to push Audio to hub. However, it would be nice if the data (text and audio) could be segmented when being pushed (if not possible already).
### Motivation
It's easy to create a vtt file from an audio file. If there could be auto-segment... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6599/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6599/timeline | null | not_planned |
https://api.github.com/repos/huggingface/datasets/issues/6598 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6598/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6598/comments | https://api.github.com/repos/huggingface/datasets/issues/6598/events | https://github.com/huggingface/datasets/issues/6598 | 2,084,236,605 | I_kwDODunzps58Ou09 | 6,598 | Unexpected keyword argument 'hf' when downloading CSV dataset from S3 | {
"login": "dguenms",
"id": 5592111,
"node_id": "MDQ6VXNlcjU1OTIxMTE=",
"avatar_url": "https://avatars.githubusercontent.com/u/5592111?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dguenms",
"html_url": "https://github.com/dguenms",
"followers_url": "https://api.github.com/users/dguenms/... | [] | open | false | null | [] | null | [
"I am facing similar issue while reading a csv file from s3. Wondering if somebody has found a workaround. ",
"same thing happened to other formats like parquet",
"I am facing similar issue while reading a parquet file from s3.\r\ni try with every version between 2.14 to 2.16.1 but it dosen't work ",
"Re-def... | 2024-01-16T15:16:01 | 2024-03-09T02:32:01 | null | NONE | null | null | null | ### Describe the bug
I receive this error message when using `load_dataset` with "csv" path and `dataset_files=s3://...`:
```
TypeError: Session.__init__() got an unexpected keyword argument 'hf'
```
I found a similar issue here: https://stackoverflow.com/questions/77596258/aws-issue-load-dataset-from-s3-fails-w... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6598/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6598/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6597 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6597/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6597/comments | https://api.github.com/repos/huggingface/datasets/issues/6597/events | https://github.com/huggingface/datasets/issues/6597 | 2,083,708,521 | I_kwDODunzps58Mt5p | 6,597 | Dataset.push_to_hub of a canonical dataset creates an additional dataset under the user namespace | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"It is caused by these code lines: https://github.com/huggingface/datasets/blob/9d6d16117a30ba345b0236407975f701c5b288d4/src/datasets/dataset_dict.py#L1688-L1694",
"Also note the information in the docstring: https://github.com/huggingface/datasets/blob/9d6d16117a30ba345b0236407975f701c5b288d4/src/datasets/datase... | 2024-01-16T11:27:07 | 2024-02-05T12:29:37 | 2024-02-05T12:29:37 | MEMBER | null | null | null | While using `Dataset.push_to_hub` of a canonical dataset, an additional dataset was created under my user namespace.
## Steps to reproduce the bug
The command:
```python
commit_info = ds.push_to_hub(
"caner",
config_name="default",
commit_message="Convert dataset to Parquet",
commit_descriptio... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6597/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6597/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6595 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6595/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6595/comments | https://api.github.com/repos/huggingface/datasets/issues/6595/events | https://github.com/huggingface/datasets/issues/6595 | 2,082,896,148 | I_kwDODunzps58JnkU | 6,595 | Loading big dataset raises pyarrow.lib.ArrowNotImplementedError 2 | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | [
"Hi ! I think the issue comes from the \"float16\" features that are not supported yet in Parquet\r\n\r\nFeel free to open an issue in `pyarrow` about this. In the meantime, I'd encourage you to use \"float32\" for your \"pooled_prompt_embeds\" and \"prompt_embeds\" features.\r\n\r\nYou can cast them to \"float32\"... | 2024-01-16T02:03:09 | 2024-01-27T18:26:33 | 2024-01-26T02:28:32 | NONE | null | null | null | ### Describe the bug
I'm aware of the issue #5695 .
I'm using a modified SDXL trainer: https://github.com/kopyl/diffusers/blob/5e70f604155aeecee254a5c63c5e4236ad4a0d3d/examples/text_to_image/train_text_to_image_sdxl.py#L1027C16-L1027C16
So i
1. Map dataset
2. Save to disk
3. Try to upload:
```
import data... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6595/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6595/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6594 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6594/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6594/comments | https://api.github.com/repos/huggingface/datasets/issues/6594/events | https://github.com/huggingface/datasets/issues/6594 | 2,082,748,275 | I_kwDODunzps58JDdz | 6,594 | IterableDataset sharding logic needs improvement | {
"login": "rwightman",
"id": 5702664,
"node_id": "MDQ6VXNlcjU3MDI2NjQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/5702664?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rwightman",
"html_url": "https://github.com/rwightman",
"followers_url": "https://api.github.com/users/rw... | [] | open | false | null | [] | null | [] | 2024-01-15T22:22:36 | 2024-01-15T22:25:10 | null | NONE | null | null | null | ### Describe the bug
The sharding of IterableDatasets with respect to distributed and dataloader worker processes appears problematic with significant performance traps and inconsistencies wrt to distributed train processes vs worker processes.
Splitting across num_workers (per train process loader processes) and... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6594/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6594/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6592 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6592/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6592/comments | https://api.github.com/repos/huggingface/datasets/issues/6592/events | https://github.com/huggingface/datasets/issues/6592 | 2,082,410,257 | I_kwDODunzps58Hw8R | 6,592 | Logs are delayed when doing .map when `docker logs` | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | [
"Hi! `tqdm` doesn't work well in non-interactive environments, so there isn't much we can do about this. It's best to [disable it](https://huggingface.co/docs/datasets/v2.16.1/en/package_reference/utilities#datasets.disable_progress_bars) in such environments and instead use logging to track progress."
] | 2024-01-15T17:05:21 | 2024-02-12T17:35:21 | 2024-02-12T17:35:21 | NONE | null | null | null | ### Describe the bug
When I run my SD training in a Docker image and then listen to logs like `docker logs train -f`, the progress bar is delayed.
It's updating every few percent.
When you have a large dataset that has to be mapped (like 1+ million samples), it's crucial to see the updates in real-time, not every co... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6592/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6592/timeline | null | not_planned |
https://api.github.com/repos/huggingface/datasets/issues/6591 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6591/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6591/comments | https://api.github.com/repos/huggingface/datasets/issues/6591/events | https://github.com/huggingface/datasets/issues/6591 | 2,082,378,957 | I_kwDODunzps58HpTN | 6,591 | The datasets models housed in Dropbox can't support a lot of users downloading them | {
"login": "RDaneelOlivav",
"id": 4933774,
"node_id": "MDQ6VXNlcjQ5MzM3NzQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/4933774?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RDaneelOlivav",
"html_url": "https://github.com/RDaneelOlivav",
"followers_url": "https://api.github.... | [] | closed | false | null | [] | null | [
"Hi! Indeed, Dropbox is not a reliable host. I've just merged https://huggingface.co/datasets/PolyAI/minds14/discussions/24 to fix this by hosting the data files inside the repo."
] | 2024-01-15T16:43:38 | 2024-01-22T23:18:09 | 2024-01-22T23:18:09 | NONE | null | null | null | ### Describe the bug
I'm using the datasets
```
from datasets import load_dataset, Audio
dataset = load_dataset("PolyAI/minds14", name="en-US", split="train")
```
And it seems that sometimes when I imagine a lot of users are accessing the same resources, the Dropbox host fails:
`raise ConnectionError(... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6591/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6591/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6590 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6590/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6590/comments | https://api.github.com/repos/huggingface/datasets/issues/6590/events | https://github.com/huggingface/datasets/issues/6590 | 2,082,000,084 | I_kwDODunzps58GMzU | 6,590 | Feature request: Multi-GPU dataset mapping for SDXL training | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | [] | 2024-01-15T13:06:06 | 2024-01-15T13:07:07 | null | NONE | null | null | null | ### Feature request
We need to speed up SDXL dataset pre-process. Please make it possible to use multiple GPUs for the [official SDXL trainer](https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_sdxl.py) :)
### Motivation
Pre-computing 3 million of images takes around ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6590/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6590/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6589 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6589/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6589/comments | https://api.github.com/repos/huggingface/datasets/issues/6589/events | https://github.com/huggingface/datasets/issues/6589 | 2,081,358,619 | I_kwDODunzps58DwMb | 6,589 | After `2.16.0` version, there are `PermissionError` when users use shared cache_dir | {
"login": "minhopark-neubla",
"id": 106717516,
"node_id": "U_kgDOBlxhTA",
"avatar_url": "https://avatars.githubusercontent.com/u/106717516?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/minhopark-neubla",
"html_url": "https://github.com/minhopark-neubla",
"followers_url": "https://api.gi... | [] | closed | false | null | [] | null | [
"We'll do a new release of `datasets` in the coming days with a fix !",
"@lhoestq Thank you very much!"
] | 2024-01-15T06:46:27 | 2024-02-02T07:55:38 | 2024-01-30T15:28:38 | NONE | null | null | null | ### Describe the bug
- We use shared `cache_dir` using `HF_HOME="{shared_directory}"`
- After dataset version 2.16.0, datasets uses `filelock` package for file locking #6445
- But, `filelock` package make `.lock` file with `644` permission
- Dataset is not available to other users except the user who created the ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6589/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6589/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6588 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6588/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6588/comments | https://api.github.com/repos/huggingface/datasets/issues/6588/events | https://github.com/huggingface/datasets/issues/6588 | 2,081,284,253 | I_kwDODunzps58DeCd | 6,588 | fix os.listdir return name is empty string | {
"login": "d710055071",
"id": 12895488,
"node_id": "MDQ6VXNlcjEyODk1NDg4",
"avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/d710055071",
"html_url": "https://github.com/d710055071",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [] | 2024-01-15T05:34:36 | 2024-01-24T10:08:29 | 2024-01-24T10:08:29 | CONTRIBUTOR | null | null | null | ### Describe the bug
xlistdir return name is empty string
Overloaded os.listdir
### Steps to reproduce the bug
```python
from datasets.download.streaming_download_manager import xjoin
from datasets.download.streaming_download_manager import xlistdir
config = DownloadConfig(storage_options=options)
manger = Str... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6588/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6588/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6585 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6585/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6585/comments | https://api.github.com/repos/huggingface/datasets/issues/6585/events | https://github.com/huggingface/datasets/issues/6585 | 2,078,874,005 | I_kwDODunzps576RmV | 6,585 | losing DatasetInfo in Dataset.map when num_proc > 1 | {
"login": "JochenSiegWork",
"id": 135010976,
"node_id": "U_kgDOCAwaoA",
"avatar_url": "https://avatars.githubusercontent.com/u/135010976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JochenSiegWork",
"html_url": "https://github.com/JochenSiegWork",
"followers_url": "https://api.github.c... | [] | open | false | {
"login": "JochenSiegWork",
"id": 135010976,
"node_id": "U_kgDOCAwaoA",
"avatar_url": "https://avatars.githubusercontent.com/u/135010976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JochenSiegWork",
"html_url": "https://github.com/JochenSiegWork",
"followers_url": "https://api.github.c... | [
{
"login": "JochenSiegWork",
"id": 135010976,
"node_id": "U_kgDOCAwaoA",
"avatar_url": "https://avatars.githubusercontent.com/u/135010976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JochenSiegWork",
"html_url": "https://github.com/JochenSiegWork",
"followers_url": "... | null | [
"Hi ! This issue comes from the fact that `map()` with `num_proc>1` shards the dataset in multiple chunks to be processed (one per process) and merges them. The DatasetInfos of each chunk are then merged together, but for some fields like `dataset_name` it's not been implemented and default to None.\r\n\r\nThe Data... | 2024-01-12T13:39:19 | 2024-01-12T14:08:24 | null | CONTRIBUTOR | null | null | null | ### Describe the bug
Hello and thanks for developing this package!
When I process a Dataset with the map function using multiple processors some set attributes of the DatasetInfo get lost and are None in the resulting Dataset.
### Steps to reproduce the bug
```python
from datasets import Dataset, DatasetInfo... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6585/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6585/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6584 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6584/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6584/comments | https://api.github.com/repos/huggingface/datasets/issues/6584/events | https://github.com/huggingface/datasets/issues/6584 | 2,078,454,878 | I_kwDODunzps574rRe | 6,584 | np.fromfile not supported | {
"login": "d710055071",
"id": 12895488,
"node_id": "MDQ6VXNlcjEyODk1NDg4",
"avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/d710055071",
"html_url": "https://github.com/d710055071",
"followers_url": "https://api.github.com/use... | [] | open | false | null | [] | null | [
"@lhoestq\r\nCan you provide me with some ideas?",
"Hi ! What's the error ?",
"@lhoestq \r\n```\r\nTraceback (most recent call last):\r\n File \"/home/dongzf/miniconda3/envs/dataset_ai/lib/python3.11/runpy.py\", line 198, in _run_module_as_main\r\n return _run_code(code, main_globals, None,\r\n ^^... | 2024-01-12T09:46:17 | 2024-01-15T05:20:50 | null | CONTRIBUTOR | null | null | null | How to do np.fromfile to use it like np.load
```python
def xnumpy_fromfile(filepath_or_buffer, *args, download_config: Optional[DownloadConfig] = None, **kwargs):
import numpy as np
if hasattr(filepath_or_buffer, "read"):
return np.fromfile(filepath_or_buffer, *args, **kwargs)
else:
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6584/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6584/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6580 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6580/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6580/comments | https://api.github.com/repos/huggingface/datasets/issues/6580/events | https://github.com/huggingface/datasets/issues/6580 | 2,075,645,042 | I_kwDODunzps57t9Ry | 6,580 | dataset cache only stores one config of the dataset in parquet dir, and uses that for all other configs resulting in showing same data in all configs. | {
"login": "kartikgupta321",
"id": 78641018,
"node_id": "MDQ6VXNlcjc4NjQxMDE4",
"avatar_url": "https://avatars.githubusercontent.com/u/78641018?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kartikgupta321",
"html_url": "https://github.com/kartikgupta321",
"followers_url": "https://api.gi... | [] | closed | false | null | [] | null | [] | 2024-01-11T03:14:18 | 2024-01-20T12:46:16 | 2024-01-20T12:46:16 | NONE | null | null | null | ### Describe the bug
ds = load_dataset("ai2_arc", "ARC-Easy"), i have tried to force redownload, delete cache and changing the cache dir.
### Steps to reproduce the bug
dataset = []
dataset_name = "ai2_arc"
possible_configs = [
'ARC-Challenge',
'ARC-Easy'
]
for config in possible_configs:
data... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6580/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6580/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6579 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6579/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6579/comments | https://api.github.com/repos/huggingface/datasets/issues/6579/events | https://github.com/huggingface/datasets/issues/6579 | 2,075,407,473 | I_kwDODunzps57tDRx | 6,579 | Unable to load `eli5` dataset with streaming | {
"login": "haok1402",
"id": 89672451,
"node_id": "MDQ6VXNlcjg5NjcyNDUx",
"avatar_url": "https://avatars.githubusercontent.com/u/89672451?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/haok1402",
"html_url": "https://github.com/haok1402",
"followers_url": "https://api.github.com/users/hao... | [] | closed | false | null | [] | null | [
"Hi @haok1402, I have created an issue in the Discussion tab of the corresponding dataset: https://huggingface.co/datasets/eli5/discussions/7\r\nLet's continue the discussion there!"
] | 2024-01-10T23:44:20 | 2024-01-11T09:19:18 | 2024-01-11T09:19:17 | NONE | null | null | null | ### Describe the bug
Unable to load `eli5` dataset with streaming.
### Steps to reproduce the bug
This fails with FileNotFoundError: https://files.pushshift.io/reddit/submissions
```
from datasets import load_dataset
load_dataset("eli5", streaming=True)
```
This works correctly.
```
from datasets import lo... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6579/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6579/timeline | null | not_planned |
https://api.github.com/repos/huggingface/datasets/issues/6577 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6577/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6577/comments | https://api.github.com/repos/huggingface/datasets/issues/6577/events | https://github.com/huggingface/datasets/issues/6577 | 2,074,790,848 | I_kwDODunzps57qsvA | 6,577 | 502 Server Errors when streaming large dataset | {
"login": "sanchit-gandhi",
"id": 93869735,
"node_id": "U_kgDOBZhWpw",
"avatar_url": "https://avatars.githubusercontent.com/u/93869735?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sanchit-gandhi",
"html_url": "https://github.com/sanchit-gandhi",
"followers_url": "https://api.github.com... | [
{
"id": 3287858981,
"node_id": "MDU6TGFiZWwzMjg3ODU4OTgx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/streaming",
"name": "streaming",
"color": "fef2c0",
"default": false,
"description": ""
}
] | closed | false | null | [] | null | [
"cc @mariosasko @lhoestq ",
"Hi! We should be able to avoid this error by retrying to read the data when it happens. I'll open a PR in `huggingface_hub` to address this.",
"Thanks for the fix @mariosasko! Just wondering whether \"500 error\" should also be excluded? I got these errors overnight:\r\n\r\n```\r\nh... | 2024-01-10T16:59:36 | 2024-02-12T11:46:03 | 2024-01-15T16:05:44 | CONTRIBUTOR | null | null | null | ### Describe the bug
When streaming a [large ASR dataset](https://huggingface.co/datasets/sanchit-gandhi/concatenated-train-set) from the Hug (~3TB) I often encounter 502 Server Errors seemingly randomly during streaming:
```
huggingface_hub.utils._errors.HfHubHTTPError: 502 Server Error: Bad Gateway for url: http... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6577/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6577/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6576 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6576/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6576/comments | https://api.github.com/repos/huggingface/datasets/issues/6576/events | https://github.com/huggingface/datasets/issues/6576 | 2,073,710,124 | I_kwDODunzps57mk4s | 6,576 | document page 404 not found after redirection | {
"login": "annahung31",
"id": 39179888,
"node_id": "MDQ6VXNlcjM5MTc5ODg4",
"avatar_url": "https://avatars.githubusercontent.com/u/39179888?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/annahung31",
"html_url": "https://github.com/annahung31",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [
"Thanks for reporting! I've opened a PR with a fix."
] | 2024-01-10T06:48:14 | 2024-01-17T14:01:31 | 2024-01-17T14:01:31 | NONE | null | null | null | ### Describe the bug
The redirected page encountered 404 not found.
### Steps to reproduce the bug
1. In this tutorial: https://huggingface.co/learn/nlp-course/chapter5/4?fw=pt
original md: https://github.com/huggingface/course/blob/2c733c2246b8b7e0e6f19a9e5d15bb12df43b2a3/chapters/en/chapter5/4.mdx#L49
`... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6576/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6576/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6571 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6571/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6571/comments | https://api.github.com/repos/huggingface/datasets/issues/6571/events | https://github.com/huggingface/datasets/issues/6571 | 2,072,111,000 | I_kwDODunzps57geeY | 6,571 | Make DatasetDict.column_names return a list instead of dict | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | [] | 2024-01-09T10:45:17 | 2024-01-09T10:45:17 | null | MEMBER | null | null | null | Currently, `DatasetDict.column_names` returns a dict, with each split name as keys and the corresponding list of column names as values.
However, by construction, all splits have the same column names.
I think it makes more sense to return a single list with the column names, which is the same for all the split k... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6571/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6571/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6570 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6570/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6570/comments | https://api.github.com/repos/huggingface/datasets/issues/6570/events | https://github.com/huggingface/datasets/issues/6570 | 2,071,805,265 | I_kwDODunzps57fT1R | 6,570 | No online docs for 2.16 release | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 1935892861,
"node_id": "MDU6TGFiZWwxOTM1ODk... | closed | false | null | [] | null | [
"Though the `build / build_main_documentation` CI job ran for 2.16.0: https://github.com/huggingface/datasets/actions/runs/7300836845/job/19896275099 🤔 ",
"Yes, I saw it. Maybe @mishig25 can give us some hint...",
"fixed https://huggingface.co/docs/datasets/v2.16.0/en/index",
"Still missing 2.16.1.",
"> St... | 2024-01-09T07:43:30 | 2024-01-09T16:45:50 | 2024-01-09T16:45:50 | MEMBER | null | null | null | We do not have the online docs for the latest minor release 2.16 (2.16.0 nor 2.16.1).
In the online docs, the latest version appearing is 2.15.0: https://huggingface.co/docs/datasets/index
..... | 2024-01-08T08:03:58 | 2024-01-13T04:53:04 | null | NONE | null | null | null | UPD: [Fixed](https://github.com/huggingface/datasets/issues/6568#issuecomment-1880817794) . But a new issue came up :( | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6568/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6568/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6567 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6567/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6567/comments | https://api.github.com/repos/huggingface/datasets/issues/6567/events | https://github.com/huggingface/datasets/issues/6567 | 2,069,808,842 | I_kwDODunzps57XsbK | 6,567 | AttributeError: 'str' object has no attribute 'to' | {
"login": "andysingal",
"id": 20493493,
"node_id": "MDQ6VXNlcjIwNDkzNDkz",
"avatar_url": "https://avatars.githubusercontent.com/u/20493493?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/andysingal",
"html_url": "https://github.com/andysingal",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [
"I think you are reporting an issue with the `transformers` library. Note this is the repository of the `datasets` library. I recommend that you open an issue in their repository: https://github.com/huggingface/transformers/issues\r\n\r\nEDIT: I have not the rights to transfer the issue\r\n~~I am transferring your ... | 2024-01-08T06:40:21 | 2024-01-08T11:56:19 | 2024-01-08T10:03:17 | NONE | null | null | null | ### Describe the bug
```
--------------------------------------------------------------------------
AttributeError Traceback (most recent call last)
[<ipython-input-6-80c6086794e8>](https://localhost:8080/#) in <cell line: 10>()
8 report_to="wandb")
9
---> 10 trainer =... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6567/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6567/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6566 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6566/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6566/comments | https://api.github.com/repos/huggingface/datasets/issues/6566/events | https://github.com/huggingface/datasets/issues/6566 | 2,069,495,429 | I_kwDODunzps57Wf6F | 6,566 | I train controlnet_sdxl in bf16 datatype, got unsupported ERROR in datasets | {
"login": "HelloWorldBeginner",
"id": 25008090,
"node_id": "MDQ6VXNlcjI1MDA4MDkw",
"avatar_url": "https://avatars.githubusercontent.com/u/25008090?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/HelloWorldBeginner",
"html_url": "https://github.com/HelloWorldBeginner",
"followers_url": "ht... | [] | open | false | null | [] | null | [
"I also see the same error and get passed it by casting that line to float. \r\n\r\nso `for x in obj.detach().cpu().numpy()` becomes `for x in obj.detach().to(torch.float).cpu().numpy()`\r\n\r\nI got the idea from [this ](https://github.com/kohya-ss/sd-webui-additional-networks/pull/128/files) PR where someone was... | 2024-01-08T02:37:03 | 2024-01-20T00:29:18 | null | NONE | null | null | null | ### Describe the bug
```
Traceback (most recent call last):
File "train_controlnet_sdxl.py", line 1252, in <module>
main(args)
File "train_controlnet_sdxl.py", line 1013, in main
train_dataset = train_dataset.map(compute_embeddings_fn, batched=True, new_fingerprint=new_fingerprint)
File "/home/mini... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6566/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6566/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6565 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6565/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6565/comments | https://api.github.com/repos/huggingface/datasets/issues/6565/events | https://github.com/huggingface/datasets/issues/6565 | 2,068,939,670 | I_kwDODunzps57UYOW | 6,565 | `drop_last_batch=True` for IterableDataset map function is ignored with multiprocessing DataLoader | {
"login": "naba89",
"id": 12119806,
"node_id": "MDQ6VXNlcjEyMTE5ODA2",
"avatar_url": "https://avatars.githubusercontent.com/u/12119806?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/naba89",
"html_url": "https://github.com/naba89",
"followers_url": "https://api.github.com/users/naba89/fo... | [] | closed | false | null | [] | null | [
"My current workaround this issue is to return `None` in the second element and then filter out samples which have `None` in them.\r\n\r\n```python\r\ndef merge_samples(batch):\r\n if len(batch['a']) == 1:\r\n batch['c'] = [batch['a'][0]]\r\n batch['d'] = [None]\r\n else:\r\n batch['c'] ... | 2024-01-07T02:46:50 | 2024-01-11T16:10:31 | 2024-01-11T16:10:31 | NONE | null | null | null | ### Describe the bug
Scenario:
- Interleaving two iterable datasets of unequal lengths (`all_exhausted`), followed by a batch mapping with batch size 2 to effectively merge the two datasets and get a sample from each dataset in a single batch, with `drop_last_batch=True` to skip the last batch in case it doesn't ha... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6565/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6565/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6564 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6564/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6564/comments | https://api.github.com/repos/huggingface/datasets/issues/6564/events | https://github.com/huggingface/datasets/issues/6564 | 2,068,893,194 | I_kwDODunzps57UM4K | 6,564 | `Dataset.filter` missing `with_rank` parameter | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | closed | false | null | [] | null | [
"Thanks for reporting! I've opened a PR with a fix",
"@mariosasko thank you very much :)"
] | 2024-01-06T23:48:13 | 2024-01-29T16:36:55 | 2024-01-29T16:36:54 | NONE | null | null | null | ### Describe the bug
The issue shall be open: https://github.com/huggingface/datasets/issues/6435
When i try to pass `with_rank` to `Dataset.filter()`, i get this:
`Dataset.filter() got an unexpected keyword argument 'with_rank'`
### Steps to reproduce the bug
Run notebook:
https://colab.research.google.com... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6564/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6564/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6563 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6563/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6563/comments | https://api.github.com/repos/huggingface/datasets/issues/6563/events | https://github.com/huggingface/datasets/issues/6563 | 2,068,302,402 | I_kwDODunzps57R8pC | 6,563 | `ImportError`: cannot import name 'insecure_hashlib' from 'huggingface_hub.utils' (.../huggingface_hub/utils/__init__.py) | {
"login": "wasertech",
"id": 79070834,
"node_id": "MDQ6VXNlcjc5MDcwODM0",
"avatar_url": "https://avatars.githubusercontent.com/u/79070834?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wasertech",
"html_url": "https://github.com/wasertech",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | [
"@Wauplin Do you happen to know what's up?",
"<del>Installing `datasets` from `main` did the trick so I guess it will be fixed in the next release.\r\n\r\nNVM https://github.com/huggingface/datasets/blob/d26abadce0b884db32382b92422d8a6aa997d40a/src/datasets/utils/info_utils.py#L5",
"@wasertech upgrading `huggin... | 2024-01-06T02:28:54 | 2024-03-14T02:59:42 | 2024-01-06T16:13:27 | NONE | null | null | null | ### Describe the bug
Yep its not [there](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/utils/__init__.py) anymore.
```text
+ python /home/trainer/sft_train.py --model_name cognitivecomputations/dolphin-2.2.1-mistral-7b --dataset_name wasertech/OneOS --load_in_4bit --use_peft --batch_... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6563/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6563/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6562 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6562/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6562/comments | https://api.github.com/repos/huggingface/datasets/issues/6562/events | https://github.com/huggingface/datasets/issues/6562 | 2,067,904,504 | I_kwDODunzps57Qbf4 | 6,562 | datasets.DownloadMode.FORCE_REDOWNLOAD use cache to download dataset features with load_dataset function | {
"login": "LsTam91",
"id": 73234162,
"node_id": "MDQ6VXNlcjczMjM0MTYy",
"avatar_url": "https://avatars.githubusercontent.com/u/73234162?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/LsTam91",
"html_url": "https://github.com/LsTam91",
"followers_url": "https://api.github.com/users/LsTam9... | [] | open | false | null | [] | null | [] | 2024-01-05T19:10:25 | 2024-01-05T19:10:25 | null | NONE | null | null | null | ### Describe the bug
I have updated my dataset by adding a new feature, and push it to the hub. When I want to download it on my machine which contain the old version by using `datasets.load_dataset("your_dataset_name", download_mode=datasets.DownloadMode.FORCE_REDOWNLOAD)` I get an error (paste bellow).
Seems that... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6562/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6562/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6561 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6561/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6561/comments | https://api.github.com/repos/huggingface/datasets/issues/6561/events | https://github.com/huggingface/datasets/issues/6561 | 2,067,404,951 | I_kwDODunzps57OhiX | 6,561 | Document YAML configuration with "data_dir" | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892861,
"node_id": "MDU6TGFiZWwxOTM1ODkyODYx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/documentation",
"name": "documentation",
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation"
}
] | open | false | null | [] | null | [
"In particular, I would like to have an example of how to replace the following configuration (from https://huggingface.co/docs/hub/datasets-manual-configuration#splits)\r\n\r\n```\r\n---\r\nconfigs:\r\n- config_name: default\r\n data_files:\r\n - split: train\r\n path: \"data/*.csv\"\r\n - split: test\r\n ... | 2024-01-05T14:03:33 | 2024-01-05T14:06:18 | null | CONTRIBUTOR | null | null | null | See https://huggingface.co/datasets/uonlp/CulturaX/discussions/15#6597e83f185db94370d6bf50 for reference | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6561/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6561/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6560 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6560/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6560/comments | https://api.github.com/repos/huggingface/datasets/issues/6560/events | https://github.com/huggingface/datasets/issues/6560 | 2,065,637,625 | I_kwDODunzps57HyD5 | 6,560 | Support Video | {
"login": "yuvalkirstain",
"id": 57996478,
"node_id": "MDQ6VXNlcjU3OTk2NDc4",
"avatar_url": "https://avatars.githubusercontent.com/u/57996478?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yuvalkirstain",
"html_url": "https://github.com/yuvalkirstain",
"followers_url": "https://api.githu... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | [] | 2024-01-04T13:10:58 | 2024-01-04T13:10:58 | null | NONE | null | null | null | ### Feature request
HF datasets are awesome in supporting text and images. Will be great to see such a support in videos :)
### Motivation
Video generation :)
### Your contribution
Will probably be limited to raising this feature request ;) | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6560/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6560/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6559 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6559/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6559/comments | https://api.github.com/repos/huggingface/datasets/issues/6559/events | https://github.com/huggingface/datasets/issues/6559 | 2,065,118,332 | I_kwDODunzps57FzR8 | 6,559 | Latest version 2.16.1, when load dataset error occurs. ValueError: BuilderConfig 'allenai--c4' not found. Available: ['default'] | {
"login": "zhulinJulia24",
"id": 145004780,
"node_id": "U_kgDOCKSY7A",
"avatar_url": "https://avatars.githubusercontent.com/u/145004780?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zhulinJulia24",
"html_url": "https://github.com/zhulinJulia24",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | [
"Hi ! The \"allenai--c4\" config doesn't exist (this naming schema comes from old versions of `datasets`)\r\n\r\nYou can load it this way instead:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ncache_dir = 'path/to/your/cache/directory'\r\ndataset = load_dataset('allenai/c4', data_files={'train': 'en/c4-t... | 2024-01-04T07:04:48 | 2024-01-05T01:26:26 | 2024-01-05T01:26:25 | NONE | null | null | null | ### Describe the bug
python script is:
```
from datasets import load_dataset
cache_dir = 'path/to/your/cache/directory'
dataset = load_dataset('allenai/c4','allenai--c4', data_files={'train': 'en/c4-train.00000-of-01024.json.gz'}, split='train', use_auth_token=False, cache_dir=cache_dir)
```
the script su... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6559/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6559/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6558 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6558/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6558/comments | https://api.github.com/repos/huggingface/datasets/issues/6558/events | https://github.com/huggingface/datasets/issues/6558 | 2,064,885,984 | I_kwDODunzps57E6jg | 6,558 | OSError: image file is truncated (1 bytes not processed) #28323 | {
"login": "andysingal",
"id": 20493493,
"node_id": "MDQ6VXNlcjIwNDkzNDkz",
"avatar_url": "https://avatars.githubusercontent.com/u/20493493?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/andysingal",
"html_url": "https://github.com/andysingal",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [
"You can add \r\n\r\n```python\r\nfrom PIL import ImageFile\r\nImageFile.LOAD_TRUNCATED_IMAGES = True\r\n```\r\n\r\nafter the imports to be able to read truncated images."
] | 2024-01-04T02:15:13 | 2024-02-21T00:38:12 | 2024-02-21T00:38:12 | NONE | null | null | null | ### Describe the bug
```
---------------------------------------------------------------------------
OSError Traceback (most recent call last)
Cell In[24], line 28
23 return example
25 # Filter the dataset
26 # filtered_dataset = dataset.filter(contains_number... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6558/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6558/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6554 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6554/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6554/comments | https://api.github.com/repos/huggingface/datasets/issues/6554/events | https://github.com/huggingface/datasets/issues/6554 | 2,063,839,916 | I_kwDODunzps57A7Ks | 6,554 | Parquet exports are used even if revision is passed | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"I don't think this bug is a thing ? Do you have some code that leads to this issue ?"
] | 2024-01-03T11:32:26 | 2024-02-02T10:35:29 | 2024-02-02T10:35:29 | MEMBER | null | null | null | We should not used Parquet exports if `revision` is passed.
I think this is a regression. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6554/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6554/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6553 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6553/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6553/comments | https://api.github.com/repos/huggingface/datasets/issues/6553/events | https://github.com/huggingface/datasets/issues/6553 | 2,063,474,183 | I_kwDODunzps56_h4H | 6,553 | Cannot import name 'load_dataset' from .... module ‘datasets’ | {
"login": "ciaoyizhen",
"id": 83450192,
"node_id": "MDQ6VXNlcjgzNDUwMTky",
"avatar_url": "https://avatars.githubusercontent.com/u/83450192?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ciaoyizhen",
"html_url": "https://github.com/ciaoyizhen",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [
"I don't know My conpany conputer cannot work. but in my computer, it work?",
"Do you have a folder in your working directory called datasets?"
] | 2024-01-03T08:18:21 | 2024-02-21T00:38:24 | 2024-02-21T00:38:24 | NONE | null | null | null | ### Describe the bug
use python -m pip install datasets to install
### Steps to reproduce the bug
from datasets import load_dataset
### Expected behavior
it doesn't work
### Environment info
datasets version==2.15.0
python == 3.10.12
linux version I don't know?? | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6553/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6553/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6552 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6552/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6552/comments | https://api.github.com/repos/huggingface/datasets/issues/6552/events | https://github.com/huggingface/datasets/issues/6552 | 2,063,157,187 | I_kwDODunzps56-UfD | 6,552 | Loading a dataset from Google Colab hangs at "Resolving data files". | {
"login": "KelSolaar",
"id": 99779,
"node_id": "MDQ6VXNlcjk5Nzc5",
"avatar_url": "https://avatars.githubusercontent.com/u/99779?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/KelSolaar",
"html_url": "https://github.com/KelSolaar",
"followers_url": "https://api.github.com/users/KelSolaar/... | [] | closed | false | null | [] | null | [
"This bug comes from the `huggingface_hub` library, see: https://github.com/huggingface/huggingface_hub/issues/1952\r\n\r\nA fix is provided at https://github.com/huggingface/huggingface_hub/pull/1953. Feel free to install `huggingface_hub` from this PR, or wait for it to be merged and the new version of `huggingfa... | 2024-01-03T02:18:17 | 2024-01-08T10:09:04 | 2024-01-08T10:09:04 | NONE | null | null | null | ### Describe the bug
Hello,
I'm trying to load a dataset from Google Colab but the process hangs at `Resolving data files`:

It is happening when the `_get_origin_metadata` definition is invoked:
```python
d... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6552/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6552/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6549 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6549/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6549/comments | https://api.github.com/repos/huggingface/datasets/issues/6549/events | https://github.com/huggingface/datasets/issues/6549 | 2,062,420,259 | I_kwDODunzps567gkj | 6,549 | Loading from hf hub with clearer error message | {
"login": "thomwolf",
"id": 7353373,
"node_id": "MDQ6VXNlcjczNTMzNzM=",
"avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/thomwolf",
"html_url": "https://github.com/thomwolf",
"followers_url": "https://api.github.com/users/thomw... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | [
"Maybe we can add a helper message like `Maybe try again using \"hf://path/without/resolve\"` if the path contains `/resolve/` ?\r\n\r\ne.g.\r\n\r\n```\r\nFileNotFoundError: Unable to find 'hf://datasets/HuggingFaceTB/eval_data/resolve/main/eval_data_context_and_answers.json'\r\nIt looks like you used parts of the ... | 2024-01-02T13:26:34 | 2024-01-02T14:06:49 | null | MEMBER | null | null | null | ### Feature request
Shouldn't this kinda work ?
```
Dataset.from_json("hf://datasets/HuggingFaceTB/eval_data/resolve/main/eval_data_context_and_answers.json")
```
I got an error
```
File ~/miniconda3/envs/datatrove/lib/python3.10/site-packages/datasets/data_files.py:380, in resolve_pattern(pattern, base_path, al... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6549/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6549/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6548 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6548/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6548/comments | https://api.github.com/repos/huggingface/datasets/issues/6548/events | https://github.com/huggingface/datasets/issues/6548 | 2,061,047,984 | I_kwDODunzps562Riw | 6,548 | Skip if a dataset has issues | {
"login": "hadianasliwa",
"id": 143214684,
"node_id": "U_kgDOCIlIXA",
"avatar_url": "https://avatars.githubusercontent.com/u/143214684?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hadianasliwa",
"html_url": "https://github.com/hadianasliwa",
"followers_url": "https://api.github.com/use... | [] | open | false | null | [] | null | [
"It looks like a transient DNS issue. It should work fine now if you try again.\r\n\r\nThere is no parameter in load_dataset to skip failed downloads. In your case it would have skipped every single subsequent download until the DNS issue was resolved anyway."
] | 2023-12-31T12:41:26 | 2024-01-02T10:33:17 | null | NONE | null | null | null | ### Describe the bug
Hello everyone,
I'm using **load_datasets** from **huggingface** to download the datasets and I'm facing an issue, the download starts but it reaches some state and then fails with the following error:
Couldn't reach https://huggingface.co/datasets/wikimedia/wikipedia/resolve/4cb9b0d719291f1a10... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6548/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6548/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6545 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6545/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6545/comments | https://api.github.com/repos/huggingface/datasets/issues/6545/events | https://github.com/huggingface/datasets/issues/6545 | 2,060,789,507 | I_kwDODunzps561ScD | 6,545 | `image` column not automatically inferred if image dataset only contains 1 image | {
"login": "apolinario",
"id": 788417,
"node_id": "MDQ6VXNlcjc4ODQxNw==",
"avatar_url": "https://avatars.githubusercontent.com/u/788417?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/apolinario",
"html_url": "https://github.com/apolinario",
"followers_url": "https://api.github.com/users/a... | [] | closed | false | null | [] | null | [] | 2023-12-30T16:17:29 | 2024-01-09T13:06:31 | 2024-01-09T13:06:31 | NONE | null | null | null | ### Describe the bug
By default, the standard Image Dataset maps out `file_name` to `image` when loading an Image Dataset.
However, if the dataset contains only 1 image, this does not take place
### Steps to reproduce the bug
Input
(dataset with one image `multimodalart/repro_1_image`)
```py
from data... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6545/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6545/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6542 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6542/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6542/comments | https://api.github.com/repos/huggingface/datasets/issues/6542/events | https://github.com/huggingface/datasets/issues/6542 | 2,059,198,575 | I_kwDODunzps56vOBv | 6,542 | Datasets : wikipedia 20220301.en error | {
"login": "ppx666",
"id": 53203620,
"node_id": "MDQ6VXNlcjUzMjAzNjIw",
"avatar_url": "https://avatars.githubusercontent.com/u/53203620?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ppx666",
"html_url": "https://github.com/ppx666",
"followers_url": "https://api.github.com/users/ppx666/fo... | [] | closed | false | null | [] | null | [
"Hi ! We now recommend using the `wikimedia/wikipedia` dataset, can you try loading this one instead ?\r\n\r\n```python\r\nwiki_dataset = load_dataset(\"wikimedia/wikipedia\", \"20231101.en\")\r\n```",
"This bug has been fixed in `2.16.1` thanks to https://github.com/huggingface/datasets/pull/6544, feel free to ... | 2023-12-29T08:34:51 | 2024-01-02T13:21:06 | 2024-01-02T13:20:30 | NONE | null | null | null | ### Describe the bug
When I used load_dataset to download this data set, the following error occurred. The main problem was that the target data did not exist.
### Steps to reproduce the bug
1.I tried downloading directly.
```python
wiki_dataset = load_dataset("wikipedia", "20220301.en")
```
An exception occurre... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6542/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6542/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6541 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6541/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6541/comments | https://api.github.com/repos/huggingface/datasets/issues/6541/events | https://github.com/huggingface/datasets/issues/6541 | 2,058,983,826 | I_kwDODunzps56uZmS | 6,541 | Dataset not loading successfully. | {
"login": "hi-sushanta",
"id": 93595990,
"node_id": "U_kgDOBZQpVg",
"avatar_url": "https://avatars.githubusercontent.com/u/93595990?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hi-sushanta",
"html_url": "https://github.com/hi-sushanta",
"followers_url": "https://api.github.com/users/hi... | [] | closed | false | null | [] | null | [
"This is a problem with your environment. You should be able to fix it by upgrading `numpy` based on [this](https://github.com/numpy/numpy/issues/23570) issue.",
"Bro I already update numpy package.",
"Then, this shouldn't throw an error on your machine:\r\n```python\r\nimport numpy\r\nnumpy._no_nep50_warning\r... | 2023-12-29T01:35:47 | 2024-01-17T00:40:46 | 2024-01-17T00:40:45 | NONE | null | null | null | ### Describe the bug
When I run down the below code shows this error: AttributeError: module 'numpy' has no attribute '_no_nep50_warning'
I also added this issue in transformers library please check out: [link](https://github.com/huggingface/transformers/issues/28099)
### Steps to reproduce the bug
## Reproduction
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6541/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6541/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6540 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6540/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6540/comments | https://api.github.com/repos/huggingface/datasets/issues/6540/events | https://github.com/huggingface/datasets/issues/6540 | 2,058,965,157 | I_kwDODunzps56uVCl | 6,540 | Extreme inefficiency for `save_to_disk` when merging datasets | {
"login": "KatarinaYuan",
"id": 43512683,
"node_id": "MDQ6VXNlcjQzNTEyNjgz",
"avatar_url": "https://avatars.githubusercontent.com/u/43512683?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/KatarinaYuan",
"html_url": "https://github.com/KatarinaYuan",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | null | [
"Concatenating datasets doesn't create any indices mapping - so flattening indices is not needed (unless you shuffle the dataset).\r\nCan you share the snippet of code you are using to merge your datasets and save them to disk ?"
] | 2023-12-29T00:44:35 | 2023-12-30T15:05:48 | null | NONE | null | null | null | ### Describe the bug
Hi, I tried to merge in total 22M sequences of data, where each sequence is of maximum length 2000. I found that merging these datasets and then `save_to_disk` is extremely slow because of flattening the indices. Wondering if you have any suggestions or guidance on this. Thank you very much!
###... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6540/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6540/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6539 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6539/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6539/comments | https://api.github.com/repos/huggingface/datasets/issues/6539/events | https://github.com/huggingface/datasets/issues/6539 | 2,058,493,960 | I_kwDODunzps56siAI | 6,539 | 'Repo card metadata block was not found' when loading a pragmeval dataset | {
"login": "lambdaofgod",
"id": 3647577,
"node_id": "MDQ6VXNlcjM2NDc1Nzc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3647577?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lambdaofgod",
"html_url": "https://github.com/lambdaofgod",
"followers_url": "https://api.github.com/us... | [] | open | false | null | [] | null | [] | 2023-12-28T14:18:25 | 2023-12-28T14:18:37 | null | NONE | null | null | null | ### Describe the bug
I can't load dataset subsets of 'pragmeval'.
The funny thing is I ran the dataset author's [colab notebook](https://colab.research.google.com/drive/1sg--LF4z7XR1wxAOfp0-3d4J6kQ9nj_A?usp=sharing) and it works just fine. I tried to install exactly the same packages that are installed on colab usi... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6539/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6539/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6538 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6538/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6538/comments | https://api.github.com/repos/huggingface/datasets/issues/6538/events | https://github.com/huggingface/datasets/issues/6538 | 2,057,377,630 | I_kwDODunzps56oRde | 6,538 | ImportError: cannot import name 'SchemaInferenceError' from 'datasets.arrow_writer' (/opt/conda/lib/python3.10/site-packages/datasets/arrow_writer.py) | {
"login": "Sonali-Behera-TRT",
"id": 131662185,
"node_id": "U_kgDOB9kBaQ",
"avatar_url": "https://avatars.githubusercontent.com/u/131662185?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Sonali-Behera-TRT",
"html_url": "https://github.com/Sonali-Behera-TRT",
"followers_url": "https://api... | [] | closed | false | null | [] | null | [
"Hi ! Are you sure you have `datasets` 2.16 ? I just checked and on 2.16 I can run `from datasets.arrow_writer import SchemaInferenceError` without error",
"I have the same issue - using with datasets version 2.16.1. Also this is on a kaggle notebook - other people with the same issue also seem to be having it on... | 2023-12-27T13:31:16 | 2024-01-03T10:06:47 | 2024-01-03T10:04:58 | NONE | null | null | null | ### Describe the bug
While importing from packages getting the error
Code:
```
import os
import torch
from datasets import load_dataset, Dataset
from transformers import (
AutoModelForCausalLM,
AutoTokenizer,
BitsAndBytesConfig,
HfArgumentParser,
TrainingArguments,
pipeline,
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6538/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6538/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6537 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6537/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6537/comments | https://api.github.com/repos/huggingface/datasets/issues/6537/events | https://github.com/huggingface/datasets/issues/6537 | 2,057,132,173 | I_kwDODunzps56nViN | 6,537 | Adding support for netCDF (*.nc) files | {
"login": "shermansiu",
"id": 12627125,
"node_id": "MDQ6VXNlcjEyNjI3MTI1",
"avatar_url": "https://avatars.githubusercontent.com/u/12627125?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shermansiu",
"html_url": "https://github.com/shermansiu",
"followers_url": "https://api.github.com/use... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | [
"Related to #3113 ",
"Conceptually, we can use xarray to load the netCDF file, then xarray -> pandas -> pyarrow.",
"I'd still need to verify that such a conversion would be lossless, especially for multi-dimensional data."
] | 2023-12-27T09:27:29 | 2023-12-27T20:46:53 | null | NONE | null | null | null | ### Feature request
netCDF (*.nc) is a file format for storing multidimensional scientific data, which is used by packages like `xarray` (labelled multi-dimensional arrays in Python). It would be nice to have native support for netCDF in `datasets`.
### Motivation
When uploading *.nc files onto Huggingface Hub throu... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6537/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6537/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6536 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6536/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6536/comments | https://api.github.com/repos/huggingface/datasets/issues/6536/events | https://github.com/huggingface/datasets/issues/6536 | 2,056,863,239 | I_kwDODunzps56mT4H | 6,536 | datasets.load_dataset raises FileNotFoundError for datasets==2.16.0 | {
"login": "ArvinZhuang",
"id": 46237844,
"node_id": "MDQ6VXNlcjQ2MjM3ODQ0",
"avatar_url": "https://avatars.githubusercontent.com/u/46237844?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ArvinZhuang",
"html_url": "https://github.com/ArvinZhuang",
"followers_url": "https://api.github.com/... | [] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"Hi ! Thanks for reporting\r\n\r\nThis is a bug in 2.16.0 for some datasets when `cache_dir` is a relative path. I opened https://github.com/huggingface/datasets/pull/6543 to fix this",
"We just released 2.16.1 with a fix:\r\n\r\n```\r\npip install -U datasets\r\n```"
] | 2023-12-27T03:15:48 | 2023-12-30T18:58:04 | 2023-12-30T15:54:00 | NONE | null | null | null | ### Describe the bug
Seems `datasets.load_dataset` raises FileNotFoundError for some hub datasets with the latest `datasets==2.16.0`
### Steps to reproduce the bug
For example `pip install datasets==2.16.0`
then
```python
import datasets
datasets.load_dataset("wentingzhao/anthropic-hh-first-prompt", cache_di... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6536/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6536/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6535 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6535/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6535/comments | https://api.github.com/repos/huggingface/datasets/issues/6535/events | https://github.com/huggingface/datasets/issues/6535 | 2,056,264,339 | I_kwDODunzps56kBqT | 6,535 | IndexError: Invalid key: 47682 is out of bounds for size 0 while using PEFT | {
"login": "MahavirDabas18",
"id": 57484266,
"node_id": "MDQ6VXNlcjU3NDg0MjY2",
"avatar_url": "https://avatars.githubusercontent.com/u/57484266?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MahavirDabas18",
"html_url": "https://github.com/MahavirDabas18",
"followers_url": "https://api.gi... | [] | open | false | null | [] | null | [
"@sabman @pvl @kashif @vigsterkr ",
"This is surely the same issue as https://discuss.huggingface.co/t/indexerror-invalid-key-16-is-out-of-bounds-for-size-0/14298/25 that comes from the `transformers` `Trainer`. You should add `remove_unused_columns=False` to `TrainingArguments`\r\n\r\nAlso check your logs: the `... | 2023-12-26T10:14:33 | 2024-02-05T08:42:31 | null | NONE | null | null | null | ### Describe the bug
I am trying to fine-tune the t5 model on the paraphrasing task. While running the same code without-
model = get_peft_model(model, config)
the model trains without any issues. However, using the model returned from get_peft_model raises the following error due to datasets-
IndexError: Inv... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6535/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6535/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6534 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6534/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6534/comments | https://api.github.com/repos/huggingface/datasets/issues/6534/events | https://github.com/huggingface/datasets/issues/6534 | 2,056,002,548 | I_kwDODunzps56jBv0 | 6,534 | How to configure multiple folders in the same zip package | {
"login": "d710055071",
"id": 12895488,
"node_id": "MDQ6VXNlcjEyODk1NDg4",
"avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/d710055071",
"html_url": "https://github.com/d710055071",
"followers_url": "https://api.github.com/use... | [] | open | false | null | [] | null | [
"@albertvillanova"
] | 2023-12-26T03:56:20 | 2023-12-26T06:31:16 | null | CONTRIBUTOR | null | null | null | How should I write "config" in readme when all the data, such as train test, is in a zip file
train floder and test floder in data.zip | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6534/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6534/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6533 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6533/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6533/comments | https://api.github.com/repos/huggingface/datasets/issues/6533/events | https://github.com/huggingface/datasets/issues/6533 | 2,055,929,101 | I_kwDODunzps56iv0N | 6,533 | ted_talks_iwslt | Error: Config name is missing | {
"login": "rayliuca",
"id": 35850903,
"node_id": "MDQ6VXNlcjM1ODUwOTAz",
"avatar_url": "https://avatars.githubusercontent.com/u/35850903?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rayliuca",
"html_url": "https://github.com/rayliuca",
"followers_url": "https://api.github.com/users/ray... | [] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"Hi ! Thanks for reporting. I opened https://github.com/huggingface/datasets/pull/6544 to fix this",
"We just released 2.16.1 with a fix:\r\n\r\n```\r\npip install -U datasets\r\n```"
] | 2023-12-26T00:38:18 | 2023-12-30T18:58:21 | 2023-12-30T16:09:50 | NONE | null | null | null | ### Describe the bug
Running load_dataset using the newest `datasets` library like below on the ted_talks_iwslt using year pair data will throw an error "Config name is missing"
see also:
https://huggingface.co/datasets/ted_talks_iwslt/discussions/3
likely caused by #6493, where the `and not config_kwargs` part... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6533/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6533/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6532 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6532/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6532/comments | https://api.github.com/repos/huggingface/datasets/issues/6532/events | https://github.com/huggingface/datasets/issues/6532 | 2,055,631,201 | I_kwDODunzps56hnFh | 6,532 | [Feature request] Indexing datasets by a customly-defined id field to enable random access dataset items via the id | {
"login": "Yu-Shi",
"id": 3377221,
"node_id": "MDQ6VXNlcjMzNzcyMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/3377221?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Yu-Shi",
"html_url": "https://github.com/Yu-Shi",
"followers_url": "https://api.github.com/users/Yu-Shi/foll... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | [
"You can simply use a python dict as index:\r\n\r\n```python\r\n>>> from datasets import load_dataset\r\n>>> ds = load_dataset(\"BeIR/dbpedia-entity\", \"corpus\", split=\"corpus\")\r\n>>> index = {key: idx for idx, key in enumerate(ds[\"_id\"])}\r\n>>> ds[index[\"<dbpedia:Pikachu>\"]]\r\n{'_id': '<dbpedia:Pikachu>... | 2023-12-25T11:37:10 | 2024-01-02T13:52:05 | null | NONE | null | null | null | ### Feature request
Some datasets may contain an id-like field, for example the `id` field in [wikimedia/wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia) and the `_id` field in [BeIR/dbpedia-entity](https://huggingface.co/datasets/BeIR/dbpedia-entity). HF datasets support efficient random access via r... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6532/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6532/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6530 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6530/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6530/comments | https://api.github.com/repos/huggingface/datasets/issues/6530/events | https://github.com/huggingface/datasets/issues/6530 | 2,054,817,609 | I_kwDODunzps56egdJ | 6,530 | Impossible to save a mapped dataset to disk | {
"login": "kopyl",
"id": 17604849,
"node_id": "MDQ6VXNlcjE3NjA0ODQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kopyl",
"html_url": "https://github.com/kopyl",
"followers_url": "https://api.github.com/users/kopyl/follow... | [] | open | false | null | [] | null | [
"I solved it with `train_dataset.with_format(None)`\r\nBut then faced some more issues (which i later solved too).\r\n\r\nHuggingface does not seem to care, so I do. Here is an updated training script which saves a pre-processed (mapped) dataset to your local directory if you specify `--save_precomputed_data_dir=DI... | 2023-12-23T15:18:27 | 2023-12-24T09:40:30 | null | NONE | null | null | null | ### Describe the bug
I want to play around with different hyperparameters when training but don't want to re-map my dataset with 3 million samples each time for tens of hours when I [fully fine-tune SDXL](https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_sdxl.py).
After... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6530/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6530/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6529 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6529/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6529/comments | https://api.github.com/repos/huggingface/datasets/issues/6529/events | https://github.com/huggingface/datasets/issues/6529 | 2,054,209,449 | I_kwDODunzps56cL-p | 6,529 | Impossible to only download a test split | {
"login": "ysig",
"id": 28439529,
"node_id": "MDQ6VXNlcjI4NDM5NTI5",
"avatar_url": "https://avatars.githubusercontent.com/u/28439529?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ysig",
"html_url": "https://github.com/ysig",
"followers_url": "https://api.github.com/users/ysig/followers"... | [] | open | false | null | [] | null | [
"The only way right now is to load with streaming=True",
"This feature has been proposed for a long time. I'm looking forward to the implementation. On clusters `streaming=True` is not an option since we do not have Internet on compute nodes. See: https://github.com/huggingface/datasets/discussions/1896#discussio... | 2023-12-22T16:56:32 | 2024-02-02T00:05:04 | null | NONE | null | null | null | I've spent a significant amount of time trying to locate the split object inside my _split_generators() custom function.
Then after diving [in the code](https://github.com/huggingface/datasets/blob/5ff3670c18ed34fa8ddfa70a9aa403ae6cc9ad54/src/datasets/load.py#L2558) I realized that `download_and_prepare` is executed b... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6529/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/huggingface/datasets/issues/6529/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6524 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6524/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6524/comments | https://api.github.com/repos/huggingface/datasets/issues/6524/events | https://github.com/huggingface/datasets/issues/6524 | 2,053,076,311 | I_kwDODunzps56X3VX | 6,524 | Streaming the Pile: Missing Files | {
"login": "FelixLabelle",
"id": 23347756,
"node_id": "MDQ6VXNlcjIzMzQ3NzU2",
"avatar_url": "https://avatars.githubusercontent.com/u/23347756?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/FelixLabelle",
"html_url": "https://github.com/FelixLabelle",
"followers_url": "https://api.github.c... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Hello @FelixLabelle,\r\n\r\nAs you can see in the Community tab of the corresponding dataset, it is a known issue: https://huggingface.co/datasets/EleutherAI/pile/discussions/15\r\n\r\nThe data has been taken down due to reported copyright infringement.\r\n\r\nFeel free to continue the discussion there."
] | 2023-12-21T21:25:09 | 2023-12-22T09:17:05 | 2023-12-22T09:17:05 | NONE | null | null | null | ### Describe the bug
The pile does not stream, a "File not Found error" is returned. It looks like the Pile's files have been moved.
### Steps to reproduce the bug
To reproduce run the following code:
```
from datasets import load_dataset
dataset = load_dataset('EleutherAI/pile', 'en', split='train', streamin... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6524/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6524/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6522 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6522/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6522/comments | https://api.github.com/repos/huggingface/datasets/issues/6522/events | https://github.com/huggingface/datasets/issues/6522 | 2,052,332,528 | I_kwDODunzps56VBvw | 6,522 | Loading HF Hub Dataset (private org repo) fails to load all features | {
"login": "versipellis",
"id": 6579034,
"node_id": "MDQ6VXNlcjY1NzkwMzQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/6579034?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/versipellis",
"html_url": "https://github.com/versipellis",
"followers_url": "https://api.github.com/us... | [] | open | false | null | [] | null | [] | 2023-12-21T12:26:35 | 2023-12-21T13:24:31 | null | NONE | null | null | null | ### Describe the bug
When pushing a `Dataset` with multiple `Features` (`input`, `output`, `tags`) to Huggingface Hub (private org repo), and later downloading the `Dataset`, only `input` and `output` load - I believe the expected behavior is for all `Features` to be loaded by default?
### Steps to reproduce the ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6522/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6522/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6521 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6521/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6521/comments | https://api.github.com/repos/huggingface/datasets/issues/6521/events | https://github.com/huggingface/datasets/issues/6521 | 2,052,229,538 | I_kwDODunzps56Uomi | 6,521 | The order of the splits is not preserved | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"After investigation, I think the issue was introduced by the use of the Parquet export:\r\n- #6448\r\n\r\nI am proposing a fix.\r\n\r\nCC: @lhoestq "
] | 2023-12-21T11:17:27 | 2023-12-22T11:36:15 | 2023-12-22T11:36:15 | MEMBER | null | null | null | We had a regression and the order of the splits is not preserved. They are alphabetically sorted, instead of preserving original "train", "validation", "test" order.
Check: In branch "main"
```python
In [9]: dataset = load_dataset("adversarial_qa", '"adversarialQA")
In [10]: dataset
Out[10]:
DatasetDict({
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6521/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6521/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6517 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6517/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6517/comments | https://api.github.com/repos/huggingface/datasets/issues/6517/events | https://github.com/huggingface/datasets/issues/6517 | 2,050,121,588 | I_kwDODunzps56Ml90 | 6,517 | Bug get_metadata_patterns arg error | {
"login": "d710055071",
"id": 12895488,
"node_id": "MDQ6VXNlcjEyODk1NDg4",
"avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/d710055071",
"html_url": "https://github.com/d710055071",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [] | 2023-12-20T08:56:44 | 2023-12-22T00:24:23 | 2023-12-22T00:24:23 | CONTRIBUTOR | null | null | null | https://github.com/huggingface/datasets/blob/3f149204a2a5948287adcade5e90707aa5207a92/src/datasets/load.py#L1240C1-L1240C69
metadata_patterns = get_metadata_patterns(base_path, download_config=self.download_config) | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6517/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6517/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6515 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6515/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6515/comments | https://api.github.com/repos/huggingface/datasets/issues/6515/events | https://github.com/huggingface/datasets/issues/6515 | 2,049,724,251 | I_kwDODunzps56LE9b | 6,515 | Why call http_head() when fsspec_head() succeeds | {
"login": "d710055071",
"id": 12895488,
"node_id": "MDQ6VXNlcjEyODk1NDg4",
"avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/d710055071",
"html_url": "https://github.com/d710055071",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [] | 2023-12-20T02:25:51 | 2023-12-26T05:35:46 | 2023-12-26T05:35:46 | CONTRIBUTOR | null | null | null | https://github.com/huggingface/datasets/blob/a91582de288d98e94bcb5ab634ca1cfeeff544c5/src/datasets/utils/file_utils.py#L510C1-L523C14 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6515/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6515/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6513 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6513/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6513/comments | https://api.github.com/repos/huggingface/datasets/issues/6513/events | https://github.com/huggingface/datasets/issues/6513 | 2,048,869,151 | I_kwDODunzps56H0Mf | 6,513 | Support huggingface-hub 0.20.0 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 2023-12-19T15:15:46 | 2023-12-20T08:44:45 | 2023-12-20T08:44:45 | MEMBER | null | null | null | CI to test the support of `huggingface-hub` 0.20.0: https://github.com/huggingface/datasets/compare/main...ci-test-huggingface-hub-v0.20.0.rc1
We need to merge:
- #6510
- #6512
- #6516 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6513/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6513/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6507 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6507/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6507/comments | https://api.github.com/repos/huggingface/datasets/issues/6507/events | https://github.com/huggingface/datasets/issues/6507 | 2,045,152,928 | I_kwDODunzps555o6g | 6,507 | where is glue_metric.py> @Frankie123421 what was the resolution to this? | {
"login": "Mcccccc1024",
"id": 119146162,
"node_id": "U_kgDOBxoGsg",
"avatar_url": "https://avatars.githubusercontent.com/u/119146162?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Mcccccc1024",
"html_url": "https://github.com/Mcccccc1024",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | [] | 2023-12-17T09:58:25 | 2023-12-18T11:42:49 | 2023-12-18T11:42:49 | NONE | null | null | null | > @Frankie123421 what was the resolution to this?
use glue_metric.py instead of glue.py in load_metric
_Originally posted by @Frankie123421 in https://github.com/huggingface/datasets/issues/2117#issuecomment-905093763_
| {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6507/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6507/timeline | null | not_planned |
https://api.github.com/repos/huggingface/datasets/issues/6506 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6506/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6506/comments | https://api.github.com/repos/huggingface/datasets/issues/6506/events | https://github.com/huggingface/datasets/issues/6506 | 2,044,975,038 | I_kwDODunzps5549e- | 6,506 | Incorrect test set labels for RTE and CoLA datasets via load_dataset | {
"login": "emreonal11",
"id": 73316684,
"node_id": "MDQ6VXNlcjczMzE2Njg0",
"avatar_url": "https://avatars.githubusercontent.com/u/73316684?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/emreonal11",
"html_url": "https://github.com/emreonal11",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [
"As this is a specific issue of the \"glue\" dataset, I have transferred it to the dataset Discussion page: https://huggingface.co/datasets/glue/discussions/15\r\n\r\nLet's continue the discussion there!"
] | 2023-12-16T22:06:08 | 2023-12-21T09:57:57 | 2023-12-21T09:57:57 | NONE | null | null | null | ### Describe the bug
The test set labels for the RTE and CoLA datasets when loading via datasets load_dataset are all -1.
Edit: It appears this is also the case for every other dataset except for MRPC (stsb, sst2, qqp, mnli (both matched and mismatched), qnli, wnli, ax). Is this intended behavior to safeguard the t... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6506/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6506/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6505 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6505/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6505/comments | https://api.github.com/repos/huggingface/datasets/issues/6505/events | https://github.com/huggingface/datasets/issues/6505 | 2,044,721,288 | I_kwDODunzps553_iI | 6,505 | Got stuck when I trying to load a dataset | {
"login": "yirenpingsheng",
"id": 18232551,
"node_id": "MDQ6VXNlcjE4MjMyNTUx",
"avatar_url": "https://avatars.githubusercontent.com/u/18232551?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yirenpingsheng",
"html_url": "https://github.com/yirenpingsheng",
"followers_url": "https://api.gi... | [] | open | false | null | [] | null | [
"I ran into the same problem when I used a server cluster (Slurm system managed) that couldn't load any of the huggingface datasets or models, but it worked on my laptop. I suspected some system configuration-related problem, but I had no idea. \r\nMy problems are consistent with [issue #2618](https://github.com/hu... | 2023-12-16T11:51:07 | 2024-02-08T21:52:05 | null | NONE | null | null | null | ### Describe the bug
Hello, everyone. I met a problem when I am trying to load a data file using load_dataset method on a Debian 10 system. The data file is not very large, only 1.63MB with 600 records.
Here is my code:
from datasets import load_dataset
dataset = load_dataset('json', data_files='mypath/oaast_r... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6505/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6505/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6504 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6504/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6504/comments | https://api.github.com/repos/huggingface/datasets/issues/6504/events | https://github.com/huggingface/datasets/issues/6504 | 2,044,541,154 | I_kwDODunzps553Tji | 6,504 | Error Pushing to Hub | {
"login": "Jiayi-Pan",
"id": 55055083,
"node_id": "MDQ6VXNlcjU1MDU1MDgz",
"avatar_url": "https://avatars.githubusercontent.com/u/55055083?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Jiayi-Pan",
"html_url": "https://github.com/Jiayi-Pan",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | [] | 2023-12-16T01:05:22 | 2023-12-16T06:20:53 | 2023-12-16T06:20:53 | NONE | null | null | null | ### Describe the bug
Error when trying to push a dataset in a special format to hub
### Steps to reproduce the bug
```
import datasets
from datasets import Dataset
dataset_dict = {
"filename": ["apple", "banana"],
"token": [[[1,2],[3,4]],[[1,2],[3,4]]],
"label": [0, 1],
}
dataset = Dataset.from_d... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6504/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6504/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6501 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6501/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6501/comments | https://api.github.com/repos/huggingface/datasets/issues/6501/events | https://github.com/huggingface/datasets/issues/6501 | 2,043,377,240 | I_kwDODunzps55y3ZY | 6,501 | OverflowError: value too large to convert to int32_t | {
"login": "zhangfan-algo",
"id": 47747764,
"node_id": "MDQ6VXNlcjQ3NzQ3NzY0",
"avatar_url": "https://avatars.githubusercontent.com/u/47747764?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zhangfan-algo",
"html_url": "https://github.com/zhangfan-algo",
"followers_url": "https://api.githu... | [] | open | false | null | [] | null | [] | 2023-12-15T10:10:21 | 2023-12-15T10:10:21 | null | NONE | null | null | null | ### Describe the bug

### Steps to reproduce the bug
just loading datasets
### Expected behavior
how can I fix it
### Environment info
pip install /mnt/cluster/zhangfan/study_info/LLaMA-Factory/peft-0.6.0-py3... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6501/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6501/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6497 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6497/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6497/comments | https://api.github.com/repos/huggingface/datasets/issues/6497/events | https://github.com/huggingface/datasets/issues/6497 | 2,041,994,274 | I_kwDODunzps55tlwi | 6,497 | Support setting a default config name in push_to_hub | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 2023-12-14T15:59:03 | 2023-12-18T11:50:04 | 2023-12-18T11:50:04 | MEMBER | null | null | null | In order to convert script-datasets to no-script datasets, we need to support setting a default config name for those scripts that set one. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6497/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6497/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6496 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6496/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6496/comments | https://api.github.com/repos/huggingface/datasets/issues/6496/events | https://github.com/huggingface/datasets/issues/6496 | 2,041,589,386 | I_kwDODunzps55sC6K | 6,496 | Error when writing a dataset to HF Hub: A commit has happened since. Please refresh and try again. | {
"login": "GeorgesLorre",
"id": 35808396,
"node_id": "MDQ6VXNlcjM1ODA4Mzk2",
"avatar_url": "https://avatars.githubusercontent.com/u/35808396?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/GeorgesLorre",
"html_url": "https://github.com/GeorgesLorre",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | null | [
"I transferred from datasets-server, since the issue is more about `datasets` and the integration with `huggingface_hub`."
] | 2023-12-14T11:24:54 | 2023-12-14T12:22:21 | null | NONE | null | null | null | **Describe the bug**
Getting a `412 Client Error: Precondition Failed` when trying to write a dataset to the HF hub.
```
huggingface_hub.utils._errors.HfHubHTTPError: 412 Client Error: Precondition Failed for url: https://huggingface.co/api/datasets/GLorr/test-dask/commit/main (Request ID: Root=1-657ae26f-3bd92b... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6496/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6496/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6494 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6494/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6494/comments | https://api.github.com/repos/huggingface/datasets/issues/6494/events | https://github.com/huggingface/datasets/issues/6494 | 2,039,684,839 | I_kwDODunzps55kx7n | 6,494 | Image Data loaded Twice | {
"login": "ArcaneLex",
"id": 28867010,
"node_id": "MDQ6VXNlcjI4ODY3MDEw",
"avatar_url": "https://avatars.githubusercontent.com/u/28867010?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ArcaneLex",
"html_url": "https://github.com/ArcaneLex",
"followers_url": "https://api.github.com/users/... | [] | open | false | null | [] | null | [] | 2023-12-13T13:11:42 | 2023-12-13T13:11:42 | null | NONE | null | null | null | ### Describe the bug

When I learn from https://huggingface.co/docs/datasets/image_load and try to load image data from a folder. I noticed that the image was read twice in the returned data. As you can see i... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6494/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6494/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6495 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6495/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6495/comments | https://api.github.com/repos/huggingface/datasets/issues/6495/events | https://github.com/huggingface/datasets/issues/6495 | 2,039,708,529 | I_kwDODunzps55k3tx | 6,495 | Newline characters don't behave as expected when calling dataset.info | {
"login": "gerald-wrona",
"id": 32300890,
"node_id": "MDQ6VXNlcjMyMzAwODkw",
"avatar_url": "https://avatars.githubusercontent.com/u/32300890?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gerald-wrona",
"html_url": "https://github.com/gerald-wrona",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | null | [] | 2023-12-12T23:07:51 | 2023-12-13T13:24:22 | null | NONE | null | null | null | ### System Info
- `transformers` version: 4.32.1
- Platform: Windows-10-10.0.19045-SP0
- Python version: 3.11.5
- Huggingface_hub version: 0.15.1
- Safetensors version: 0.3.2
- Accelerate version: not installed
- Accelerate config: not found
- PyTorch version (GPU?): 2.1.1+cpu (False)
- Tensorflow version (GPU... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6495/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6495/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6490 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6490/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6490/comments | https://api.github.com/repos/huggingface/datasets/issues/6490/events | https://github.com/huggingface/datasets/issues/6490 | 2,037,204,892 | I_kwDODunzps55bUec | 6,490 | `load_dataset(...,save_infos=True)` not working without loading script | {
"login": "morganveyret",
"id": 114978051,
"node_id": "U_kgDOBtptAw",
"avatar_url": "https://avatars.githubusercontent.com/u/114978051?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/morganveyret",
"html_url": "https://github.com/morganveyret",
"followers_url": "https://api.github.com/use... | [] | open | false | null | [] | null | [
"Also, once the README.md exists in the python environment it is used when loading another dataset in the same format (e.g. json) since it always resolves the path to the same directory.\r\nThe consequence here is any other dataset won't load because of infos mismatch.\r\nTo reproduce this aspect:\r\n1. Do a `load_... | 2023-12-12T08:09:18 | 2023-12-12T08:36:22 | null | NONE | null | null | null | ### Describe the bug
It seems that saving a dataset infos back into the card file is not working for datasets without a loading script.
After tracking the problem a bit it looks like saving the infos uses `Builder.get_imported_module_dir()` as its destination directory.
Internally this is a call to `inspect.getfil... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6490/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6490/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6489 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6489/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6489/comments | https://api.github.com/repos/huggingface/datasets/issues/6489/events | https://github.com/huggingface/datasets/issues/6489 | 2,036,743,777 | I_kwDODunzps55Zj5h | 6,489 | load_dataset imageflder for aws s3 path | {
"login": "segalinc",
"id": 9353106,
"node_id": "MDQ6VXNlcjkzNTMxMDY=",
"avatar_url": "https://avatars.githubusercontent.com/u/9353106?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/segalinc",
"html_url": "https://github.com/segalinc",
"followers_url": "https://api.github.com/users/segal... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | null | [] | 2023-12-12T00:08:43 | 2023-12-12T00:09:27 | null | NONE | null | null | null | ### Feature request
I would like to load a dataset from S3 using the imagefolder option
something like
`dataset = datasets.load_dataset('imagefolder', data_dir='s3://.../lsun/train/bedroom', fs=S3FileSystem(), streaming=True) `
### Motivation
no need of data_files
### Your contribution
no experience... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6489/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6489/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6488 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6488/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6488/comments | https://api.github.com/repos/huggingface/datasets/issues/6488/events | https://github.com/huggingface/datasets/issues/6488 | 2,035,899,898 | I_kwDODunzps55WV36 | 6,488 | 429 Client Error | {
"login": "sasaadi",
"id": 7882383,
"node_id": "MDQ6VXNlcjc4ODIzODM=",
"avatar_url": "https://avatars.githubusercontent.com/u/7882383?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sasaadi",
"html_url": "https://github.com/sasaadi",
"followers_url": "https://api.github.com/users/sasaadi/... | [] | open | false | null | [] | null | [
"Transferring repos as this is a datasets issue ",
"I'm getting a similar issue even though I've already downloaded the dataset 😅 \r\n\r\n```\r\nhuggingface_hub.utils._errors.HfHubHTTPError: 429 Client Error: Too Many Requests for url: https://huggingface.co/api/datasets/HuggingFaceM4/WebSight\r\n```"
] | 2023-12-11T15:06:01 | 2024-01-18T02:05:15 | null | NONE | null | null | null | Hello, I was downloading the following dataset and after 20% of data was downloaded, I started getting error 429. It is not resolved since a few days. How should I resolve it?
Thanks
Dataset:
https://huggingface.co/datasets/cerebras/SlimPajama-627B
Error:
`requests.exceptions.HTTPError: 429 Client Error: Too M... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6488/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6488/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6485 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6485/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6485/comments | https://api.github.com/repos/huggingface/datasets/issues/6485/events | https://github.com/huggingface/datasets/issues/6485 | 2,035,141,884 | I_kwDODunzps55Tcz8 | 6,485 | FileNotFoundError: [Errno 2] No such file or directory: 'nul' | {
"login": "amanyara",
"id": 73683903,
"node_id": "MDQ6VXNlcjczNjgzOTAz",
"avatar_url": "https://avatars.githubusercontent.com/u/73683903?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/amanyara",
"html_url": "https://github.com/amanyara",
"followers_url": "https://api.github.com/users/ama... | [] | closed | false | null | [] | null | [
"Hi! It seems like the problem is your environment. Maybe this issue can help: https://github.com/pytest-dev/pytest/issues/9519. "
] | 2023-12-11T08:52:13 | 2023-12-14T08:09:08 | 2023-12-14T08:09:08 | NONE | null | null | null | ### Describe the bug
it seems that sth wrong with my terrible "bug body" life, When i run this code, "import datasets"
i meet this error FileNotFoundError: [Errno 2] No such file or directory: 'nul'

\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n# load LS in streaming mode\r\ndataset = load_dataset(\"librispeech_asr\", \"clean\", split=\"validation\", streaming=True)\r\... | 2023-12-08T16:11:30 | 2023-12-08T16:27:16 | 2023-12-08T16:27:04 | CONTRIBUTOR | null | null | null | ### Describe the bug
Suppose I have a two iterable datasets, one with the features:
* `{"audio", "text", "column_a"}`
And the other with the features:
* `{"audio", "sentence", "column_b"}`
I want to combine both datasets using `interleave_datasets`, which requires me to unify the column names. I would typic... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6483/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6483/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6484 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6484/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6484/comments | https://api.github.com/repos/huggingface/datasets/issues/6484/events | https://github.com/huggingface/datasets/issues/6484 | 2,033,333,294 | I_kwDODunzps55MjQu | 6,484 | [Feature Request] Dataset versioning | {
"login": "kenfus",
"id": 47979198,
"node_id": "MDQ6VXNlcjQ3OTc5MTk4",
"avatar_url": "https://avatars.githubusercontent.com/u/47979198?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kenfus",
"html_url": "https://github.com/kenfus",
"followers_url": "https://api.github.com/users/kenfus/fo... | [] | open | false | null | [] | null | [
"Hello @kenfus, this is meant to be possible to do yes. Let me ping @lhoestq or @mariosasko from the `datasets` team (`huggingface_hub` is only the underlying library to download files from the Hub but here it looks more like a `datasets` problem). ",
"Hi! https://github.com/huggingface/datasets/pull/6459 will fi... | 2023-12-08T16:01:35 | 2023-12-11T19:13:46 | null | NONE | null | null | null | **Is your feature request related to a problem? Please describe.**
I am working on a project, where I would like to test different preprocessing methods for my ML-data. Thus, I would like to work a lot with revisions and compare them. Currently, I was not able to make it work with the revision keyword because it was n... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6484/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6484/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6481 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6481/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6481/comments | https://api.github.com/repos/huggingface/datasets/issues/6481/events | https://github.com/huggingface/datasets/issues/6481 | 2,032,650,003 | I_kwDODunzps55J8cT | 6,481 | using torchrun, save_to_disk suddenly shows SIGTERM | {
"login": "Ariya12138",
"id": 85916625,
"node_id": "MDQ6VXNlcjg1OTE2NjI1",
"avatar_url": "https://avatars.githubusercontent.com/u/85916625?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Ariya12138",
"html_url": "https://github.com/Ariya12138",
"followers_url": "https://api.github.com/use... | [] | open | false | null | [] | null | [] | 2023-12-08T13:22:03 | 2023-12-08T13:22:03 | null | NONE | null | null | null | ### Describe the bug
When I run my code using the "torchrun" command, when the code reaches the "save_to_disk" part, suddenly I get the following warning and error messages:
Because the dataset is too large, the "save_to_disk" function splits it into 70 parts for saving. However, an error occurs suddenly when it reac... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6481/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6481/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6478 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6478/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6478/comments | https://api.github.com/repos/huggingface/datasets/issues/6478/events | https://github.com/huggingface/datasets/issues/6478 | 2,028,071,596 | I_kwDODunzps544eqs | 6,478 | How to load data from lakefs | {
"login": "d710055071",
"id": 12895488,
"node_id": "MDQ6VXNlcjEyODk1NDg4",
"avatar_url": "https://avatars.githubusercontent.com/u/12895488?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/d710055071",
"html_url": "https://github.com/d710055071",
"followers_url": "https://api.github.com/use... | [] | open | false | null | [] | null | [
"You can create a `pandas` DataFrame following [this](https://lakefs.io/data-version-control/dvc-using-python/) tutorial, and then convert this DataFrame to a `Dataset` with `datasets.Dataset.from_pandas`. For larger datasets (to memory map them), you can use `Dataset.from_generator` with a generator function that ... | 2023-12-06T09:04:11 | 2023-12-07T02:19:44 | null | CONTRIBUTOR | null | null | null | My dataset is stored on the company's lakefs server. How can I write code to load the dataset? It would be great if I could provide code examples or provide some references
| {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6478/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6478/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6476 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6476/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6476/comments | https://api.github.com/repos/huggingface/datasets/issues/6476/events | https://github.com/huggingface/datasets/issues/6476 | 2,028,018,596 | I_kwDODunzps544Ruk | 6,476 | CI on windows is broken: PermissionError | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 2023-12-06T08:32:53 | 2023-12-06T09:17:53 | 2023-12-06T09:17:53 | MEMBER | null | null | null | See: https://github.com/huggingface/datasets/actions/runs/7104781624/job/19340572394
```
FAILED tests/test_load.py::test_loading_from_the_datasets_hub - NotADirectoryError: [WinError 267] The directory name is invalid: 'C:\\Users\\RUNNER~1\\AppData\\Local\\Temp\\tmpfcnps56i\\hf-internal-testing___dataset_with_script\... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6476/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6476/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6475 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6475/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6475/comments | https://api.github.com/repos/huggingface/datasets/issues/6475/events | https://github.com/huggingface/datasets/issues/6475 | 2,027,373,734 | I_kwDODunzps5410Sm | 6,475 | laion2B-en failed to load on Windows with PrefetchVirtualMemory failed | {
"login": "doctorpangloss",
"id": 2229300,
"node_id": "MDQ6VXNlcjIyMjkzMDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2229300?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/doctorpangloss",
"html_url": "https://github.com/doctorpangloss",
"followers_url": "https://api.gith... | [] | open | false | null | [] | null | [
"~~You will see this error if the cache dir filepath contains relative `..` paths. Use `os.path.realpath(_CACHE_DIR)` before passing it to the `load_dataset` function.~~",
"This is a real issue and not related to paths.",
"Based on the StackOverflow answer, this causes the error to go away:\r\n```diff\r\ndiff -... | 2023-12-06T00:07:34 | 2023-12-06T23:26:23 | null | NONE | null | null | null | ### Describe the bug
I have downloaded laion2B-en, and I'm receiving the following error trying to load it:
```
Resolving data files: 100%|██████████| 128/128 [00:00<00:00, 1173.79it/s]
Traceback (most recent call last):
File "D:\Art-Workspace\src\artworkspace\tokeneval\compute_frequencies.py", line 31, in <mo... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6475/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6475/timeline | null | reopened |
https://api.github.com/repos/huggingface/datasets/issues/6472 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6472/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6472/comments | https://api.github.com/repos/huggingface/datasets/issues/6472/events | https://github.com/huggingface/datasets/issues/6472 | 2,026,493,439 | I_kwDODunzps54ydX_ | 6,472 | CI quality is broken | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 4296013012,
"node_id": "LA_kwDODunzps8AAAAB... | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 2023-12-05T15:35:34 | 2023-12-06T08:17:34 | 2023-12-05T18:08:43 | MEMBER | null | null | null | See: https://github.com/huggingface/datasets/actions/runs/7100835633/job/19327734359
```
Would reformat: src/datasets/features/image.py
1 file would be reformatted, 253 files left unchanged
``` | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6472/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6472/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6470 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6470/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6470/comments | https://api.github.com/repos/huggingface/datasets/issues/6470/events | https://github.com/huggingface/datasets/issues/6470 | 2,024,724,319 | I_kwDODunzps54rtdf | 6,470 | If an image in a dataset is corrupted, we get unescapable error | {
"login": "chigozienri",
"id": 14337872,
"node_id": "MDQ6VXNlcjE0MzM3ODcy",
"avatar_url": "https://avatars.githubusercontent.com/u/14337872?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chigozienri",
"html_url": "https://github.com/chigozienri",
"followers_url": "https://api.github.com/... | [] | open | false | null | [] | null | [] | 2023-12-04T20:58:49 | 2023-12-04T20:58:49 | null | NONE | null | null | null | ### Describe the bug
Example discussed in detail here: https://huggingface.co/datasets/sasha/birdsnap/discussions/1
### Steps to reproduce the bug
```
from datasets import load_dataset, VerificationMode
dataset = load_dataset(
'sasha/birdsnap',
split="train",
verification_mode=VerificationMode.ALL_C... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6470/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6470/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6467 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6467/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6467/comments | https://api.github.com/repos/huggingface/datasets/issues/6467/events | https://github.com/huggingface/datasets/issues/6467 | 2,023,174,233 | I_kwDODunzps54lzBZ | 6,467 | New version release request | {
"login": "LZHgrla",
"id": 36994684,
"node_id": "MDQ6VXNlcjM2OTk0Njg0",
"avatar_url": "https://avatars.githubusercontent.com/u/36994684?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/LZHgrla",
"html_url": "https://github.com/LZHgrla",
"followers_url": "https://api.github.com/users/LZHgrl... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | null | [] | null | [
"We will publish it soon (we usually do it in intervals of 1-2 months, so probably next week)",
"Thanks!"
] | 2023-12-04T07:08:26 | 2023-12-04T15:42:22 | 2023-12-04T15:42:22 | CONTRIBUTOR | null | null | null | ### Feature request
Hi!
I am using `datasets` in library `xtuner` and am highly interested in the features introduced since v2.15.0.
To avoid installation from source in our pypi wheels, we are eagerly waiting for the new release. So, Does your team have a new release plan for v2.15.1 and could you please share ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6467/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6467/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6466 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6466/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6466/comments | https://api.github.com/repos/huggingface/datasets/issues/6466/events | https://github.com/huggingface/datasets/issues/6466 | 2,022,601,176 | I_kwDODunzps54jnHY | 6,466 | Can't align optional features of struct | {
"login": "Dref360",
"id": 8976546,
"node_id": "MDQ6VXNlcjg5NzY1NDY=",
"avatar_url": "https://avatars.githubusercontent.com/u/8976546?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Dref360",
"html_url": "https://github.com/Dref360",
"followers_url": "https://api.github.com/users/Dref360/... | [] | closed | false | null | [] | null | [
"Friendly bump, I would be happy to work on this issue once I get the go-ahead from the dev team. ",
"Thanks for the PR!\r\n\r\nI'm struggling with this as well and would love to see this PR merged. My case is slightly different, with keys completely missing rather than being `None`:\r\n\r\n```\r\nds = Dataset.fr... | 2023-12-03T15:57:07 | 2024-02-15T15:19:33 | 2024-02-08T14:38:34 | CONTRIBUTOR | null | null | null | ### Describe the bug
Hello!
I'm currently experiencing an issue where I can't concatenate datasets if an inner field of a Feature is Optional.
I have a column named `speaker`, and this holds some information about a speaker.
```python
@dataclass
class Speaker:
name: str
email: Optional[str]
```
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6466/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6466/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6465 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6465/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6465/comments | https://api.github.com/repos/huggingface/datasets/issues/6465/events | https://github.com/huggingface/datasets/issues/6465 | 2,022,212,468 | I_kwDODunzps54iIN0 | 6,465 | `load_dataset` uses out-of-date cache instead of re-downloading a changed dataset | {
"login": "mnoukhov",
"id": 3391297,
"node_id": "MDQ6VXNlcjMzOTEyOTc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3391297?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mnoukhov",
"html_url": "https://github.com/mnoukhov",
"followers_url": "https://api.github.com/users/mnouk... | [] | open | false | null | [] | null | [
"Hi, thanks for reporting! https://github.com/huggingface/datasets/pull/6459 will fix this."
] | 2023-12-02T21:35:17 | 2023-12-04T16:13:10 | null | NONE | null | null | null | ### Describe the bug
When a dataset is updated on the hub, using `load_dataset` will load the locally cached dataset instead of re-downloading the updated dataset
### Steps to reproduce the bug
Here is a minimal example script to
1. create an initial dataset and upload
2. download it so it is stored in cache
3. c... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6465/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6465/timeline | null | null |
https://api.github.com/repos/huggingface/datasets/issues/6460 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6460/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6460/comments | https://api.github.com/repos/huggingface/datasets/issues/6460/events | https://github.com/huggingface/datasets/issues/6460 | 2,017,433,899 | I_kwDODunzps54P5kr | 6,460 | jsonlines files don't load with `load_dataset` | {
"login": "serenalotreck",
"id": 41377532,
"node_id": "MDQ6VXNlcjQxMzc3NTMy",
"avatar_url": "https://avatars.githubusercontent.com/u/41377532?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/serenalotreck",
"html_url": "https://github.com/serenalotreck",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | [
"Hi @serenalotreck,\r\n\r\nWe use Apache Arrow `pyarrow` to read jsonlines and it throws an error when trying to load your data files:\r\n```python\r\nIn [1]: import pyarrow as pa\r\n\r\nIn [2]: data = pa.json.read_json(\"train.jsonl\")\r\n---------------------------------------------------------------------------\... | 2023-11-29T21:20:11 | 2023-12-29T02:58:29 | 2023-12-05T13:30:53 | NONE | null | null | null | ### Describe the bug
While [the docs](https://huggingface.co/docs/datasets/upload_dataset#upload-dataset) seem to state that `.jsonl` is a supported extension for `datasets`, loading the dataset results in a `JSONDecodeError`.
### Steps to reproduce the bug
Code:
```
from datasets import load_dataset
dset = load_... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6460/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6460/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6457 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6457/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6457/comments | https://api.github.com/repos/huggingface/datasets/issues/6457/events | https://github.com/huggingface/datasets/issues/6457 | 2,015,650,563 | I_kwDODunzps54JGMD | 6,457 | `TypeError`: huggingface_hub.hf_file_system.HfFileSystem.find() got multiple values for keyword argument 'maxdepth' | {
"login": "wasertech",
"id": 79070834,
"node_id": "MDQ6VXNlcjc5MDcwODM0",
"avatar_url": "https://avatars.githubusercontent.com/u/79070834?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wasertech",
"html_url": "https://github.com/wasertech",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | [
"Updating `fsspec>=2023.10.0` did solve the issue.",
"May be it should be pinned somewhere?",
"> Maybe this should go in datasets directly... anyways you can easily fix this error by updating datasets>=2.15.1.dev0.\r\n\r\n@lhoestq @mariosasko for what I understand this is a bug fixed in `datasets` already, righ... | 2023-11-29T01:57:36 | 2023-11-29T15:39:03 | 2023-11-29T02:02:38 | NONE | null | null | null | ### Describe the bug
Please see https://github.com/huggingface/huggingface_hub/issues/1872
### Steps to reproduce the bug
Please see https://github.com/huggingface/huggingface_hub/issues/1872
### Expected behavior
Please see https://github.com/huggingface/huggingface_hub/issues/1872
### Environment info
Please s... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6457/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6457/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6451 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6451/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6451/comments | https://api.github.com/repos/huggingface/datasets/issues/6451/events | https://github.com/huggingface/datasets/issues/6451 | 2,010,693,912 | I_kwDODunzps532MEY | 6,451 | Unable to read "marsyas/gtzan" data | {
"login": "gerald-wrona",
"id": 32300890,
"node_id": "MDQ6VXNlcjMyMzAwODkw",
"avatar_url": "https://avatars.githubusercontent.com/u/32300890?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gerald-wrona",
"html_url": "https://github.com/gerald-wrona",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | [
"Hi! We've merged a [PR](https://huggingface.co/datasets/marsyas/gtzan/discussions/1) that fixes the script's path logic on Windows.",
"I have transferred the discussion to the corresponding dataset: https://huggingface.co/datasets/marsyas/gtzan/discussions/2\r\n\r\nLet's continue there.",
"@mariosasko @albertv... | 2023-11-25T15:13:17 | 2023-12-01T12:53:46 | 2023-11-27T09:36:25 | NONE | null | null | null | Hi, this is my code and the error:
```
from datasets import load_dataset
gtzan = load_dataset("marsyas/gtzan", "all")
```
[error_trace.txt](https://github.com/huggingface/datasets/files/13464397/error_trace.txt)
[audio_yml.txt](https://github.com/huggingface/datasets/files/13464410/audio_yml.txt)
Python 3.11.5
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6451/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6451/timeline | null | completed |
https://api.github.com/repos/huggingface/datasets/issues/6450 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6450/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6450/comments | https://api.github.com/repos/huggingface/datasets/issues/6450/events | https://github.com/huggingface/datasets/issues/6450 | 2,009,491,386 | I_kwDODunzps53xme6 | 6,450 | Support multiple image/audio columns in ImageFolder/AudioFolder | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892865,
"node_id": "MDU6TGFiZWwxOTM1ODkyODY1",
"url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate",
"name": "duplicate",
"color": "cfd3d7",
"default": true,
"description": "This issue or pull request already exists"
},
{
"id": 1935892871,
"... | closed | false | null | [] | null | [
"A duplicate of https://github.com/huggingface/datasets/issues/5760"
] | 2023-11-24T10:34:09 | 2023-11-28T11:07:17 | 2023-11-24T17:24:38 | CONTRIBUTOR | null | null | null | ### Feature request
Have a metadata.csv file with multiple columns that point to relative image or audio files.
### Motivation
Currently, ImageFolder allows one column, called `file_name`, pointing to relative image files. On the same model, AudioFolder allows one column, called `file_name`, pointing to relative aud... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/6450/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/6450/timeline | null | completed |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.