url stringlengths 61 61 | repository_url stringclasses 1
value | labels_url stringlengths 75 75 | comments_url stringlengths 70 70 | events_url stringlengths 68 68 | html_url stringlengths 49 51 | id int64 818M 2.44B | node_id stringlengths 18 32 | number int64 1.96k 7.08k | title stringlengths 1 290 | user dict | labels listlengths 0 4 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 4 | comments sequencelengths 2 2 | created_at timestamp[s] | updated_at timestamp[s] | closed_at timestamp[s] | author_association stringclasses 4
values | draft bool 2
classes | pull_request dict | body stringlengths 0 36.2k ⌀ | reactions dict | timeline_url stringlengths 70 70 | state_reason stringclasses 3
values | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/7082 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7082/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7082/comments | https://api.github.com/repos/huggingface/datasets/issues/7082/events | https://github.com/huggingface/datasets/pull/7082 | 2,437,354,975 | PR_kwDODunzps522dTJ | 7,082 | Support HTTP authentication in non-streaming mode | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-30T09:25:49 | 2024-07-30T09:59:11 | null | MEMBER | true | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7082",
"html_url": "https://github.com/huggingface/datasets/pull/7082",
"diff_url": "https://github.com/huggingface/datasets/pull/7082.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7082.patch",
"merged_at": null
} | Support HTTP authentication in non-streaming mode, by support passing HTTP storage_options in non-streaming mode.
- Note that currently, HTTP authentication is supported only in streaming mode.
For example, this is necessary if a remote HTTP host requires authentication to download the data. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7082/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7082/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7081 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7081/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7081/comments | https://api.github.com/repos/huggingface/datasets/issues/7081/events | https://github.com/huggingface/datasets/pull/7081 | 2,437,059,657 | PR_kwDODunzps521cGm | 7,081 | Set load_from_disk path type as PathLike | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-30T07:00:38 | 2024-07-30T08:30:37 | 2024-07-30T08:21:50 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7081",
"html_url": "https://github.com/huggingface/datasets/pull/7081",
"diff_url": "https://github.com/huggingface/datasets/pull/7081.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7081.patch",
"merged_at": "2024-07-30T08:21... | Set `load_from_disk` path type as `PathLike`. This way it is aligned with `save_to_disk`. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7081/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7081/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7080 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7080/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7080/comments | https://api.github.com/repos/huggingface/datasets/issues/7080/events | https://github.com/huggingface/datasets/issues/7080 | 2,434,275,664 | I_kwDODunzps6RGBlQ | 7,080 | Generating train split takes a long time | {
"login": "alexanderswerdlow",
"id": 35648800,
"node_id": "MDQ6VXNlcjM1NjQ4ODAw",
"avatar_url": "https://avatars.githubusercontent.com/u/35648800?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/alexanderswerdlow",
"html_url": "https://github.com/alexanderswerdlow",
"followers_url": "https... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-29T01:42:43 | 2024-07-29T01:42:43 | null | NONE | null | null | ### Describe the bug
Loading a simple webdataset takes ~45 minutes.
### Steps to reproduce the bug
```
from datasets import load_dataset
dataset = load_dataset("PixArt-alpha/SAM-LLaVA-Captions10M")
```
### Expected behavior
The dataset should load immediately as it does when loaded through a normal indexed WebD... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7080/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7080/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7079 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7079/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7079/comments | https://api.github.com/repos/huggingface/datasets/issues/7079/events | https://github.com/huggingface/datasets/issues/7079 | 2,433,363,298 | I_kwDODunzps6RCi1i | 7,079 | HfHubHTTPError: 500 Server Error: Internal Server Error for url: | {
"login": "neoneye",
"id": 147971,
"node_id": "MDQ6VXNlcjE0Nzk3MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/147971?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/neoneye",
"html_url": "https://github.com/neoneye",
"followers_url": "https://api.github.com/users/neoneye/fo... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-27T08:21:03 | 2024-07-27T20:06:44 | 2024-07-27T19:52:30 | NONE | null | null | ### Describe the bug
newly uploaded datasets, since yesterday, yields an error.
old datasets, works fine.
Seems like the datasets api server returns a 500
I'm getting the same error, when I invoke `load_dataset` with my dataset.
Long discussion about it here, but I'm not sure anyone from huggingface have s... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7079/reactions",
"total_count": 7,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 3,
"rocket": 0,
"eyes": 4
} | https://api.github.com/repos/huggingface/datasets/issues/7079/timeline | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/7078 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7078/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7078/comments | https://api.github.com/repos/huggingface/datasets/issues/7078/events | https://github.com/huggingface/datasets/pull/7078 | 2,433,270,271 | PR_kwDODunzps52oq4n | 7,078 | Fix CI test_convert_to_parquet | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-27T05:32:40 | 2024-07-27T05:50:57 | 2024-07-27T05:44:32 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7078",
"html_url": "https://github.com/huggingface/datasets/pull/7078",
"diff_url": "https://github.com/huggingface/datasets/pull/7078.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7078.patch",
"merged_at": "2024-07-27T05:44... | Fix `test_convert_to_parquet` by patching `HfApi.preupload_lfs_files` and revert temporary fix:
- #7074 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7078/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7078/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7077 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7077/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7077/comments | https://api.github.com/repos/huggingface/datasets/issues/7077/events | https://github.com/huggingface/datasets/issues/7077 | 2,432,345,489 | I_kwDODunzps6Q-qWR | 7,077 | column_names ignored by load_dataset() when loading CSV file | {
"login": "luismsgomes",
"id": 9130265,
"node_id": "MDQ6VXNlcjkxMzAyNjU=",
"avatar_url": "https://avatars.githubusercontent.com/u/9130265?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/luismsgomes",
"html_url": "https://github.com/luismsgomes",
"followers_url": "https://api.github.com/us... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-26T14:18:04 | 2024-07-30T07:52:26 | null | NONE | null | null | ### Describe the bug
load_dataset() ignores the column_names kwarg when loading a CSV file. Instead, it uses whatever values are on the first line of the file.
### Steps to reproduce the bug
Call `load_dataset` to load data from a CSV file and specify `column_names` kwarg.
### Expected behavior
The resulting da... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7077/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7077/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7076 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7076/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7076/comments | https://api.github.com/repos/huggingface/datasets/issues/7076/events | https://github.com/huggingface/datasets/pull/7076 | 2,432,275,393 | PR_kwDODunzps52lTDe | 7,076 | 🧪 Do not mock create_commit | {
"login": "coyotte508",
"id": 342922,
"node_id": "MDQ6VXNlcjM0MjkyMg==",
"avatar_url": "https://avatars.githubusercontent.com/u/342922?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/coyotte508",
"html_url": "https://github.com/coyotte508",
"followers_url": "https://api.github.com/users/c... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-26T13:44:42 | 2024-07-27T05:48:17 | 2024-07-27T05:48:17 | MEMBER | true | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7076",
"html_url": "https://github.com/huggingface/datasets/pull/7076",
"diff_url": "https://github.com/huggingface/datasets/pull/7076.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7076.patch",
"merged_at": null
} | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7076/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7076/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7075 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7075/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7075/comments | https://api.github.com/repos/huggingface/datasets/issues/7075/events | https://github.com/huggingface/datasets/pull/7075 | 2,432,027,412 | PR_kwDODunzps52kciD | 7,075 | Update required soxr version from pre-release to release | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-26T11:24:35 | 2024-07-26T11:46:52 | 2024-07-26T11:40:49 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7075",
"html_url": "https://github.com/huggingface/datasets/pull/7075",
"diff_url": "https://github.com/huggingface/datasets/pull/7075.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7075.patch",
"merged_at": "2024-07-26T11:40... | Update required `soxr` version from pre-release to release 0.4.0: https://github.com/dofuuz/python-soxr/releases/tag/v0.4.0 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7075/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7075/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7074 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7074/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7074/comments | https://api.github.com/repos/huggingface/datasets/issues/7074/events | https://github.com/huggingface/datasets/pull/7074 | 2,431,772,703 | PR_kwDODunzps52jkw4 | 7,074 | Fix CI by temporarily marking test_convert_to_parquet as expected to fail | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-26T09:03:33 | 2024-07-26T09:23:33 | 2024-07-26T09:16:12 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7074",
"html_url": "https://github.com/huggingface/datasets/pull/7074",
"diff_url": "https://github.com/huggingface/datasets/pull/7074.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7074.patch",
"merged_at": "2024-07-26T09:16... | As a hotfix for CI, temporarily mark test_convert_to_parquet as expected to fail.
Fix #7073.
Revert once root cause is fixed. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7074/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7074/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7073 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7073/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7073/comments | https://api.github.com/repos/huggingface/datasets/issues/7073/events | https://github.com/huggingface/datasets/issues/7073 | 2,431,706,568 | I_kwDODunzps6Q8OXI | 7,073 | CI is broken for convert_to_parquet: Invalid rev id: refs/pr/1 404 error causes RevisionNotFoundError | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-26T08:27:41 | 2024-07-27T05:48:02 | 2024-07-26T09:16:13 | MEMBER | null | null | See: https://github.com/huggingface/datasets/actions/runs/10095313567/job/27915185756
```
FAILED tests/test_hub.py::test_convert_to_parquet - huggingface_hub.utils._errors.RevisionNotFoundError: 404 Client Error. (Request ID: Root=1-66a25839-31ce7b475e70e7db1e4d44c2;b0c8870f-d5ef-4bf2-a6ff-0191f3df0f64)
Revision N... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7073/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7073/timeline | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/7072 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7072/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7072/comments | https://api.github.com/repos/huggingface/datasets/issues/7072/events | https://github.com/huggingface/datasets/issues/7072 | 2,430,577,916 | I_kwDODunzps6Q36z8 | 7,072 | nm | {
"login": "brettdavies",
"id": 26392883,
"node_id": "MDQ6VXNlcjI2MzkyODgz",
"avatar_url": "https://avatars.githubusercontent.com/u/26392883?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/brettdavies",
"html_url": "https://github.com/brettdavies",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-25T17:03:24 | 2024-07-25T20:36:11 | 2024-07-25T20:36:11 | NONE | null | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7072/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7072/timeline | not_planned | false |
https://api.github.com/repos/huggingface/datasets/issues/7071 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7071/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7071/comments | https://api.github.com/repos/huggingface/datasets/issues/7071/events | https://github.com/huggingface/datasets/issues/7071 | 2,430,313,011 | I_kwDODunzps6Q26Iz | 7,071 | Filter hangs | {
"login": "lucienwalewski",
"id": 61711045,
"node_id": "MDQ6VXNlcjYxNzExMDQ1",
"avatar_url": "https://avatars.githubusercontent.com/u/61711045?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lucienwalewski",
"html_url": "https://github.com/lucienwalewski",
"followers_url": "https://api.gi... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-25T15:29:05 | 2024-07-25T15:36:59 | null | NONE | null | null | ### Describe the bug
When trying to filter my custom dataset, the process hangs, regardless of the lambda function used. It appears to be an issue with the way the Images are being handled. The dataset in question is a preprocessed version of https://huggingface.co/datasets/danaaubakirova/patfig where notably, I hav... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7071/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7071/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7070 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7070/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7070/comments | https://api.github.com/repos/huggingface/datasets/issues/7070/events | https://github.com/huggingface/datasets/issues/7070 | 2,430,285,235 | I_kwDODunzps6Q2zWz | 7,070 | how set_transform affects batch size? | {
"login": "VafaKnm",
"id": 103993288,
"node_id": "U_kgDOBjLPyA",
"avatar_url": "https://avatars.githubusercontent.com/u/103993288?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/VafaKnm",
"html_url": "https://github.com/VafaKnm",
"followers_url": "https://api.github.com/users/VafaKnm/foll... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-25T15:19:34 | 2024-07-25T15:19:34 | null | NONE | null | null | ### Describe the bug
I am trying to fine-tune w2v-bert for ASR task. Since my dataset is so big, I preferred to use the on-the-fly method with set_transform. So i change the preprocessing function to this:
```
def prepare_dataset(batch):
input_features = processor(batch["audio"], sampling_rate=16000).input_feat... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7070/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7070/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7069 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7069/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7069/comments | https://api.github.com/repos/huggingface/datasets/issues/7069/events | https://github.com/huggingface/datasets/pull/7069 | 2,429,281,339 | PR_kwDODunzps52betB | 7,069 | Fix push_to_hub by not calling create_branch if branch exists | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-25T07:50:04 | 2024-07-30T10:56:57 | 2024-07-30T10:51:01 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7069",
"html_url": "https://github.com/huggingface/datasets/pull/7069",
"diff_url": "https://github.com/huggingface/datasets/pull/7069.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7069.patch",
"merged_at": "2024-07-30T10:51... | Fix push_to_hub by not calling create_branch if branch exists.
Note that currently create_branch raises a 403 Forbidden error even if all these conditions are met:
- exist_ok is passed
- the branch already exists
- the user does not have WRITE permission
Fix #7067.
Related issue:
- https://github.com/huggi... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7069/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7069/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7068 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7068/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7068/comments | https://api.github.com/repos/huggingface/datasets/issues/7068/events | https://github.com/huggingface/datasets/pull/7068 | 2,426,657,434 | PR_kwDODunzps52SwXS | 7,068 | Fix prepare_single_hop_path_and_storage_options | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-24T05:52:34 | 2024-07-29T07:02:07 | 2024-07-29T06:56:15 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7068",
"html_url": "https://github.com/huggingface/datasets/pull/7068",
"diff_url": "https://github.com/huggingface/datasets/pull/7068.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7068.patch",
"merged_at": "2024-07-29T06:56... | Fix `_prepare_single_hop_path_and_storage_options`:
- Do not pass HF authentication headers and HF user-agent to non-HF HTTP URLs
- Do not overwrite passed `storage_options` nested values:
- Before, when passed
```DownloadConfig(storage_options={"https": {"client_kwargs": {"raise_for_status": True}}})```,
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7068/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7068/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7067 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7067/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7067/comments | https://api.github.com/repos/huggingface/datasets/issues/7067/events | https://github.com/huggingface/datasets/issues/7067 | 2,425,460,168 | I_kwDODunzps6QkZXI | 7,067 | Convert_to_parquet fails for datasets with multiple configs | {
"login": "HuangZhen02",
"id": 97585031,
"node_id": "U_kgDOBdEHhw",
"avatar_url": "https://avatars.githubusercontent.com/u/97585031?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/HuangZhen02",
"html_url": "https://github.com/HuangZhen02",
"followers_url": "https://api.github.com/users/Hu... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-23T15:09:33 | 2024-07-30T10:51:02 | 2024-07-30T10:51:02 | NONE | null | null | If the dataset has multiple configs, when using the `datasets-cli convert_to_parquet` command to avoid issues with the data viewer caused by loading scripts, the conversion process only successfully converts the data corresponding to the first config. When it starts converting the second config, it throws an error:
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7067/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7067/timeline | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/7066 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7066/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7066/comments | https://api.github.com/repos/huggingface/datasets/issues/7066/events | https://github.com/huggingface/datasets/issues/7066 | 2,425,125,160 | I_kwDODunzps6QjHko | 7,066 | One subset per file in repo ? | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-23T12:43:59 | 2024-07-23T12:43:59 | null | MEMBER | null | null | Right now we consider all the files of a dataset to be the same data, e.g.
```
single_subset_dataset/
├── train0.jsonl
├── train1.jsonl
└── train2.jsonl
```
but in cases like this, each file is actually a different subset of the dataset and should be loaded separately
```
many_subsets_dataset/
├── animals.jso... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7066/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7066/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7065 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7065/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7065/comments | https://api.github.com/repos/huggingface/datasets/issues/7065/events | https://github.com/huggingface/datasets/issues/7065 | 2,424,734,953 | I_kwDODunzps6QhoTp | 7,065 | Cannot get item after loading from disk and then converting to iterable. | {
"login": "happyTonakai",
"id": 21305646,
"node_id": "MDQ6VXNlcjIxMzA1NjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/21305646?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/happyTonakai",
"html_url": "https://github.com/happyTonakai",
"followers_url": "https://api.github.c... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-23T09:37:56 | 2024-07-23T09:37:56 | null | NONE | null | null | ### Describe the bug
The dataset generated from local file works fine.
```py
root = "/home/data/train"
file_list1 = glob(os.path.join(root, "*part1.flac"))
file_list2 = glob(os.path.join(root, "*part2.flac"))
ds = (
Dataset.from_dict({"part1": file_list1, "part2": file_list2})
.cast_column("part1", Au... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7065/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7065/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7064 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7064/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7064/comments | https://api.github.com/repos/huggingface/datasets/issues/7064/events | https://github.com/huggingface/datasets/pull/7064 | 2,424,613,104 | PR_kwDODunzps52Lz2- | 7,064 | Add `batch` method to `Dataset` class | {
"login": "lappemic",
"id": 61876623,
"node_id": "MDQ6VXNlcjYxODc2NjIz",
"avatar_url": "https://avatars.githubusercontent.com/u/61876623?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lappemic",
"html_url": "https://github.com/lappemic",
"followers_url": "https://api.github.com/users/lap... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-23T08:40:43 | 2024-07-25T13:51:25 | 2024-07-25T13:45:20 | CONTRIBUTOR | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7064",
"html_url": "https://github.com/huggingface/datasets/pull/7064",
"diff_url": "https://github.com/huggingface/datasets/pull/7064.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7064.patch",
"merged_at": "2024-07-25T13:45... | This PR introduces a new `batch` method to the `Dataset` class, aligning its functionality with the `IterableDataset.batch()` method (implemented in #7054). The implementation uses as well the existing `map` method for efficient batching of examples.
Key changes:
- Add `batch` method to `Dataset` class in `arrow_da... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7064/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7064/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7063 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7063/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7063/comments | https://api.github.com/repos/huggingface/datasets/issues/7063/events | https://github.com/huggingface/datasets/issues/7063 | 2,424,488,648 | I_kwDODunzps6QgsLI | 7,063 | Add `batch` method to `Dataset` | {
"login": "lappemic",
"id": 61876623,
"node_id": "MDQ6VXNlcjYxODc2NjIz",
"avatar_url": "https://avatars.githubusercontent.com/u/61876623?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lappemic",
"html_url": "https://github.com/lappemic",
"followers_url": "https://api.github.com/users/lap... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-23T07:36:59 | 2024-07-25T13:45:21 | 2024-07-25T13:45:21 | CONTRIBUTOR | null | null | ### Feature request
Add a `batch` method to the Dataset class, similar to the one recently implemented for `IterableDataset` in PR #7054.
### Motivation
A batched iteration speeds up data loading significantly (see e.g. #6279)
### Your contribution
I plan to open a PR to implement this. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7063/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7063/timeline | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/7062 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7062/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7062/comments | https://api.github.com/repos/huggingface/datasets/issues/7062/events | https://github.com/huggingface/datasets/pull/7062 | 2,424,467,484 | PR_kwDODunzps52LUPR | 7,062 | Avoid calling http_head for non-HTTP URLs | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-23T07:25:09 | 2024-07-23T14:28:27 | 2024-07-23T14:21:08 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7062",
"html_url": "https://github.com/huggingface/datasets/pull/7062",
"diff_url": "https://github.com/huggingface/datasets/pull/7062.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7062.patch",
"merged_at": "2024-07-23T14:21... | Avoid calling `http_head` for non-HTTP URLs, by adding and `else` statement.
Currently, it makes an unnecessary HTTP call (which adds latency) for non-HTTP protocols, like FTP, S3,...
I discovered this while working in an unrelated issue. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7062/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7062/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7061 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7061/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7061/comments | https://api.github.com/repos/huggingface/datasets/issues/7061/events | https://github.com/huggingface/datasets/issues/7061 | 2,423,786,881 | I_kwDODunzps6QeA2B | 7,061 | Custom Dataset | Still Raise Error while handling errors in _generate_examples | {
"login": "hahmad2008",
"id": 68266028,
"node_id": "MDQ6VXNlcjY4MjY2MDI4",
"avatar_url": "https://avatars.githubusercontent.com/u/68266028?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hahmad2008",
"html_url": "https://github.com/hahmad2008",
"followers_url": "https://api.github.com/use... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-22T21:18:12 | 2024-07-22T21:18:12 | null | NONE | null | null | ### Describe the bug
I follow this [example](https://discuss.huggingface.co/t/error-handling-in-iterabledataset/72827/3) to handle errors in custom dataset. I am writing a dataset script which read jsonl files and i need to handle errors and continue reading files without raising exception and exit the execution.
`... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7061/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7061/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7060 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7060/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7060/comments | https://api.github.com/repos/huggingface/datasets/issues/7060/events | https://github.com/huggingface/datasets/pull/7060 | 2,423,188,419 | PR_kwDODunzps52G71g | 7,060 | WebDataset BuilderConfig | {
"login": "hlky",
"id": 106811348,
"node_id": "U_kgDOBl3P1A",
"avatar_url": "https://avatars.githubusercontent.com/u/106811348?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hlky",
"html_url": "https://github.com/hlky",
"followers_url": "https://api.github.com/users/hlky/followers",
"f... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-22T15:41:07 | 2024-07-23T13:28:44 | 2024-07-23T13:28:44 | NONE | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7060",
"html_url": "https://github.com/huggingface/datasets/pull/7060",
"diff_url": "https://github.com/huggingface/datasets/pull/7060.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7060.patch",
"merged_at": null
} | This PR adds `WebDatasetConfig`.
Closes #7055 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7060/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7060/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7059 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7059/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7059/comments | https://api.github.com/repos/huggingface/datasets/issues/7059/events | https://github.com/huggingface/datasets/issues/7059 | 2,422,827,892 | I_kwDODunzps6QaWt0 | 7,059 | None values are skipped when reading jsonl in subobjects | {
"login": "PonteIneptique",
"id": 1929830,
"node_id": "MDQ6VXNlcjE5Mjk4MzA=",
"avatar_url": "https://avatars.githubusercontent.com/u/1929830?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/PonteIneptique",
"html_url": "https://github.com/PonteIneptique",
"followers_url": "https://api.gith... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-22T13:02:42 | 2024-07-22T13:02:53 | null | NONE | null | null | ### Describe the bug
I have been fighting against my machine since this morning only to find out this is some kind of a bug.
When loading a dataset composed of `metadata.jsonl`, if you have nullable values (Optional[str]), they can be ignored by the parser, shifting things around.
E.g., let's take this example
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7059/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7059/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7058 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7058/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7058/comments | https://api.github.com/repos/huggingface/datasets/issues/7058/events | https://github.com/huggingface/datasets/issues/7058 | 2,422,560,355 | I_kwDODunzps6QZVZj | 7,058 | New feature type: Document | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-22T10:49:20 | 2024-07-22T10:49:20 | null | CONTRIBUTOR | null | null | It would be useful for PDF.
https://github.com/huggingface/dataset-viewer/issues/2991#issuecomment-2242656069 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7058/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7058/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7057 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7057/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7057/comments | https://api.github.com/repos/huggingface/datasets/issues/7057/events | https://github.com/huggingface/datasets/pull/7057 | 2,422,498,520 | PR_kwDODunzps52EjGC | 7,057 | Update load_hub.mdx | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-22T10:17:46 | 2024-07-22T10:34:14 | 2024-07-22T10:28:10 | CONTRIBUTOR | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7057",
"html_url": "https://github.com/huggingface/datasets/pull/7057",
"diff_url": "https://github.com/huggingface/datasets/pull/7057.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7057.patch",
"merged_at": "2024-07-22T10:28... | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7057/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7057/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7056 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7056/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7056/comments | https://api.github.com/repos/huggingface/datasets/issues/7056/events | https://github.com/huggingface/datasets/pull/7056 | 2,422,192,257 | PR_kwDODunzps52DgOu | 7,056 | Make `BufferShuffledExamplesIterable` resumable | {
"login": "yzhangcs",
"id": 18402347,
"node_id": "MDQ6VXNlcjE4NDAyMzQ3",
"avatar_url": "https://avatars.githubusercontent.com/u/18402347?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yzhangcs",
"html_url": "https://github.com/yzhangcs",
"followers_url": "https://api.github.com/users/yzh... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-22T07:50:02 | 2024-07-22T15:37:01 | null | CONTRIBUTOR | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7056",
"html_url": "https://github.com/huggingface/datasets/pull/7056",
"diff_url": "https://github.com/huggingface/datasets/pull/7056.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7056.patch",
"merged_at": null
} | This PR aims to implement a resumable `BufferShuffledExamplesIterable`.
Instead of saving the entire buffer content, which is very memory-intensive, the newly implemented `BufferShuffledExamplesIterable` saves only the minimal state necessary for recovery, e.g., the random generator states and the state of the first e... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7056/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7056/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7055 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7055/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7055/comments | https://api.github.com/repos/huggingface/datasets/issues/7055/events | https://github.com/huggingface/datasets/issues/7055 | 2,421,708,891 | I_kwDODunzps6QWFhb | 7,055 | WebDataset with different prefixes are unsupported | {
"login": "hlky",
"id": 106811348,
"node_id": "U_kgDOBl3P1A",
"avatar_url": "https://avatars.githubusercontent.com/u/106811348?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hlky",
"html_url": "https://github.com/hlky",
"followers_url": "https://api.github.com/users/hlky/followers",
"f... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-22T01:14:19 | 2024-07-24T13:26:30 | 2024-07-23T13:28:46 | NONE | null | null | ### Describe the bug
Consider a WebDataset with multiple images for each item where the number of images may vary: [example](https://huggingface.co/datasets/bigdata-pw/fashion-150k)
Due to this [code](https://github.com/huggingface/datasets/blob/87f4c2088854ff33e817e724e75179e9975c1b02/src/datasets/packaged_modules... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7055/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7055/timeline | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/7054 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7054/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7054/comments | https://api.github.com/repos/huggingface/datasets/issues/7054/events | https://github.com/huggingface/datasets/pull/7054 | 2,418,548,995 | PR_kwDODunzps514T1f | 7,054 | Add batching to `IterableDataset` | {
"login": "lappemic",
"id": 61876623,
"node_id": "MDQ6VXNlcjYxODc2NjIz",
"avatar_url": "https://avatars.githubusercontent.com/u/61876623?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lappemic",
"html_url": "https://github.com/lappemic",
"followers_url": "https://api.github.com/users/lap... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-19T10:11:47 | 2024-07-23T13:25:13 | 2024-07-23T10:34:28 | CONTRIBUTOR | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7054",
"html_url": "https://github.com/huggingface/datasets/pull/7054",
"diff_url": "https://github.com/huggingface/datasets/pull/7054.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7054.patch",
"merged_at": "2024-07-23T10:34... | I've taken a try at implementing a batched `IterableDataset` as requested in issue #6279. This PR adds a new `BatchedExamplesIterable` class and a `.batch()` method to the `IterableDataset` class.
The main changes are:
1. A new `BatchedExamplesIterable` that groups examples into batches.
2. A `.batch()` method for... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7054/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7054/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7053 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7053/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7053/comments | https://api.github.com/repos/huggingface/datasets/issues/7053/events | https://github.com/huggingface/datasets/issues/7053 | 2,416,423,791 | I_kwDODunzps6QB7Nv | 7,053 | Datasets.datafiles resolve_pattern `TypeError: can only concatenate tuple (not "str") to tuple` | {
"login": "MatthewYZhang",
"id": 48289218,
"node_id": "MDQ6VXNlcjQ4Mjg5MjE4",
"avatar_url": "https://avatars.githubusercontent.com/u/48289218?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MatthewYZhang",
"html_url": "https://github.com/MatthewYZhang",
"followers_url": "https://api.githu... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-18T13:42:35 | 2024-07-18T15:17:42 | 2024-07-18T15:16:18 | NONE | null | null | ### Describe the bug
in data_files.py, line 332,
`fs, _, _ = get_fs_token_paths(pattern, storage_options=storage_options)`
If we run the code on AWS, as fs.protocol will be a tuple like: `('file', 'local')`
So, `isinstance(fs.protocol, str) == False` and
`protocol_prefix = fs.protocol + "://" if fs.protocol != ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7053/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7053/timeline | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/7052 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7052/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7052/comments | https://api.github.com/repos/huggingface/datasets/issues/7052/events | https://github.com/huggingface/datasets/pull/7052 | 2,411,682,730 | PR_kwDODunzps51iuop | 7,052 | Adding `Music` feature for symbolic music modality (MIDI, abc) | {
"login": "Natooz",
"id": 56734983,
"node_id": "MDQ6VXNlcjU2NzM0OTgz",
"avatar_url": "https://avatars.githubusercontent.com/u/56734983?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Natooz",
"html_url": "https://github.com/Natooz",
"followers_url": "https://api.github.com/users/Natooz/fo... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-16T17:26:04 | 2024-07-29T06:47:55 | 2024-07-29T06:47:55 | NONE | true | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7052",
"html_url": "https://github.com/huggingface/datasets/pull/7052",
"diff_url": "https://github.com/huggingface/datasets/pull/7052.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7052.patch",
"merged_at": null
} | ⚠️ (WIP) ⚠️
### What this PR does
This PR adds a `Music` feature for the symbolic music modality, in particular [MIDI](https://en.wikipedia.org/wiki/Musical_Instrument_Digital_Interface) and [abc](https://en.wikipedia.org/wiki/ABC_notation) files.
### Motivations
These two file formats are widely used in th... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7052/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7052/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7051 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7051/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7051/comments | https://api.github.com/repos/huggingface/datasets/issues/7051/events | https://github.com/huggingface/datasets/issues/7051 | 2,409,353,929 | I_kwDODunzps6Pm9LJ | 7,051 | How to set_epoch with interleave_datasets? | {
"login": "jonathanasdf",
"id": 511073,
"node_id": "MDQ6VXNlcjUxMTA3Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/511073?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jonathanasdf",
"html_url": "https://github.com/jonathanasdf",
"followers_url": "https://api.github.com/u... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-15T18:24:52 | 2024-07-22T16:52:07 | null | NONE | null | null | Let's say I have dataset A which has 100k examples, and dataset B which has 100m examples.
I want to train on an interleaved dataset of A+B, with stopping_strategy='all_exhausted' so dataset B doesn't repeat any examples. But every time A is exhausted I want it to be reshuffled (eg. calling set_epoch)
Of course I... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7051/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/huggingface/datasets/issues/7051/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7050 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7050/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7050/comments | https://api.github.com/repos/huggingface/datasets/issues/7050/events | https://github.com/huggingface/datasets/pull/7050 | 2,409,048,733 | PR_kwDODunzps51Z1Yp | 7,050 | add checkpoint and resume title in docs | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-15T15:38:04 | 2024-07-15T16:06:15 | 2024-07-15T15:59:56 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7050",
"html_url": "https://github.com/huggingface/datasets/pull/7050",
"diff_url": "https://github.com/huggingface/datasets/pull/7050.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7050.patch",
"merged_at": "2024-07-15T15:59... | (minor) just to make it more prominent in the docs page for the soon-to-be-released new torchdata | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7050/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7050/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7049 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7049/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7049/comments | https://api.github.com/repos/huggingface/datasets/issues/7049/events | https://github.com/huggingface/datasets/issues/7049 | 2,408,514,366 | I_kwDODunzps6PjwM- | 7,049 | Save nparray as list | {
"login": "Sakurakdx",
"id": 48399040,
"node_id": "MDQ6VXNlcjQ4Mzk5MDQw",
"avatar_url": "https://avatars.githubusercontent.com/u/48399040?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Sakurakdx",
"html_url": "https://github.com/Sakurakdx",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-15T11:36:11 | 2024-07-18T11:33:34 | 2024-07-18T11:33:34 | NONE | null | null | ### Describe the bug
When I use the `map` function to convert images into features, datasets saves nparray as a list. Some people use the `set_format` function to convert the column back, but doesn't this lose precision?
### Steps to reproduce the bug
the map function
```python
def convert_image_to_features(inst, ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7049/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7049/timeline | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/7048 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7048/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7048/comments | https://api.github.com/repos/huggingface/datasets/issues/7048/events | https://github.com/huggingface/datasets/issues/7048 | 2,408,487,547 | I_kwDODunzps6Pjpp7 | 7,048 | ImportError: numpy.core.multiarray when using `filter` | {
"login": "kamilakesbi",
"id": 45195979,
"node_id": "MDQ6VXNlcjQ1MTk1OTc5",
"avatar_url": "https://avatars.githubusercontent.com/u/45195979?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kamilakesbi",
"html_url": "https://github.com/kamilakesbi",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-15T11:21:04 | 2024-07-16T10:11:25 | 2024-07-16T10:11:25 | NONE | null | null | ### Describe the bug
I can't apply the filter method on my dataset.
### Steps to reproduce the bug
The following snippet generates a bug:
```python
from datasets import load_dataset
ami = load_dataset('kamilakesbi/ami', 'ihm')
ami['train'].filter(
lambda example: example["file_name"] == 'EN2001a'
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7048/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7048/timeline | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/7047 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7047/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7047/comments | https://api.github.com/repos/huggingface/datasets/issues/7047/events | https://github.com/huggingface/datasets/issues/7047 | 2,406,495,084 | I_kwDODunzps6PcDNs | 7,047 | Save Dataset as Sharded Parquet | {
"login": "tom-p-reichel",
"id": 43631024,
"node_id": "MDQ6VXNlcjQzNjMxMDI0",
"avatar_url": "https://avatars.githubusercontent.com/u/43631024?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tom-p-reichel",
"html_url": "https://github.com/tom-p-reichel",
"followers_url": "https://api.githu... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-12T23:47:51 | 2024-07-17T12:07:08 | null | NONE | null | null | ### Feature request
`to_parquet` currently saves the dataset as one massive, monolithic parquet file, rather than as several small parquet files. It should shard large datasets automatically.
### Motivation
This default behavior makes me very sad because a program I ran for 6 hours saved its results using `to_... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7047/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7047/timeline | null | false |
https://api.github.com/repos/huggingface/datasets/issues/7046 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7046/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7046/comments | https://api.github.com/repos/huggingface/datasets/issues/7046/events | https://github.com/huggingface/datasets/pull/7046 | 2,405,485,582 | PR_kwDODunzps51N05n | 7,046 | Support librosa and numpy 2.0 for Python 3.10 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-12T12:42:47 | 2024-07-12T13:04:40 | 2024-07-12T12:58:17 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7046",
"html_url": "https://github.com/huggingface/datasets/pull/7046",
"diff_url": "https://github.com/huggingface/datasets/pull/7046.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7046.patch",
"merged_at": "2024-07-12T12:58... | Support librosa and numpy 2.0 for Python 3.10 by installing soxr 0.4.0b1 pre-release:
- https://github.com/dofuuz/python-soxr/releases/tag/v0.4.0b1
- https://github.com/dofuuz/python-soxr/issues/28 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7046/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7046/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7045 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7045/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7045/comments | https://api.github.com/repos/huggingface/datasets/issues/7045/events | https://github.com/huggingface/datasets/pull/7045 | 2,405,447,858 | PR_kwDODunzps51Nsie | 7,045 | Fix tensorflow min version depending on Python version | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-12T12:20:23 | 2024-07-12T12:38:53 | 2024-07-12T12:33:00 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7045",
"html_url": "https://github.com/huggingface/datasets/pull/7045",
"diff_url": "https://github.com/huggingface/datasets/pull/7045.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7045.patch",
"merged_at": "2024-07-12T12:33... | Fix tensorflow min version depending on Python version.
Related to:
- #6991 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7045/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7045/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7044 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7044/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7044/comments | https://api.github.com/repos/huggingface/datasets/issues/7044/events | https://github.com/huggingface/datasets/pull/7044 | 2,405,002,987 | PR_kwDODunzps51MLbh | 7,044 | Mark tests that require librosa | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-12T08:06:59 | 2024-07-12T09:06:32 | 2024-07-12T09:00:09 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7044",
"html_url": "https://github.com/huggingface/datasets/pull/7044",
"diff_url": "https://github.com/huggingface/datasets/pull/7044.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7044.patch",
"merged_at": "2024-07-12T09:00... | Mark tests that require `librosa`.
Note that `librosa` is an optional dependency (installed with `audio` option) and we should be able to test environments without that library installed. This is the case if we want to test Numpy 2.0, which is currently incompatible with `librosa` due to its dependency on `soxr`:
-... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7044/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7044/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7043 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7043/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7043/comments | https://api.github.com/repos/huggingface/datasets/issues/7043/events | https://github.com/huggingface/datasets/pull/7043 | 2,404,951,714 | PR_kwDODunzps51MAN0 | 7,043 | Add decorator as explicit test dependency | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-12T07:35:23 | 2024-07-12T08:12:55 | 2024-07-12T08:07:10 | MEMBER | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7043",
"html_url": "https://github.com/huggingface/datasets/pull/7043",
"diff_url": "https://github.com/huggingface/datasets/pull/7043.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7043.patch",
"merged_at": "2024-07-12T08:07... | Add decorator as explicit test dependency.
We use `decorator` library in our CI test since PR:
- #4845
However we did not add it as an explicit test requirement, and we depended on it indirectly through other libraries' dependencies.
I discovered this while testing Numpy 2.0 and removing incompatible librarie... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7043/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7043/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7042 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7042/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7042/comments | https://api.github.com/repos/huggingface/datasets/issues/7042/events | https://github.com/huggingface/datasets/pull/7042 | 2,404,605,836 | PR_kwDODunzps51K8CM | 7,042 | Improved the tutorial by adding a link for loading datasets | {
"login": "AmboThom",
"id": 41874659,
"node_id": "MDQ6VXNlcjQxODc0NjU5",
"avatar_url": "https://avatars.githubusercontent.com/u/41874659?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/AmboThom",
"html_url": "https://github.com/AmboThom",
"followers_url": "https://api.github.com/users/Amb... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-12T03:49:54 | 2024-07-12T03:49:54 | null | NONE | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7042",
"html_url": "https://github.com/huggingface/datasets/pull/7042",
"diff_url": "https://github.com/huggingface/datasets/pull/7042.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7042.patch",
"merged_at": null
} | Improved the tutorial by letting readers know about loading datasets with common files and including a link. I left the local files section alone because the methods were already listed with code snippets. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7042/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7042/timeline | null | true |
https://api.github.com/repos/huggingface/datasets/issues/7041 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/7041/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/7041/comments | https://api.github.com/repos/huggingface/datasets/issues/7041/events | https://github.com/huggingface/datasets/issues/7041 | 2,404,576,038 | I_kwDODunzps6PUusm | 7,041 | `sort` after `filter` unreasonably slow | {
"login": "Tobin-rgb",
"id": 56711045,
"node_id": "MDQ6VXNlcjU2NzExMDQ1",
"avatar_url": "https://avatars.githubusercontent.com/u/56711045?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Tobin-rgb",
"html_url": "https://github.com/Tobin-rgb",
"followers_url": "https://api.github.com/users/... | [] | open | false | null | [] | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 2024-07-12T03:29:27 | 2024-07-22T13:55:17 | null | NONE | null | null | ### Describe the bug
as the tittle says ...
### Steps to reproduce the bug
`sort` seems to be normal.
```python
from datasets import Dataset
import random
nums = [{"k":random.choice(range(0,1000))} for _ in range(100000)]
ds = Dataset.from_list(nums)
print("start sort")
ds = ds.sort("k")
print("f... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/7041/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/7041/timeline | null | false |
End of preview. Expand in Data Studio
Dataset Card for GitHub Issues
Dataset Summary
GitHub Issues is a dataset consisting of GitHub issues and pull requests associated with the 🤗 Datasets repository. It can be used for semantic search or multilabel text classification.
- Downloads last month
- 4