id int64 1.9B 3.25B | title stringlengths 2 244 | state stringclasses 2
values | body stringlengths 3 58.6k ⌀ | created_at timestamp[s]date 2023-09-15 14:23:33 2025-07-22 09:33:54 | updated_at timestamp[s]date 2023-09-18 16:20:09 2025-07-22 10:44:03 | closed_at timestamp[s]date 2023-09-18 16:20:09 2025-07-19 22:45:08 ⌀ | html_url stringlengths 49 51 | pull_request dict | number int64 6.24k 7.7k | is_pull_request bool 2
classes | comments listlengths 0 24 |
|---|---|---|---|---|---|---|---|---|---|---|---|
2,343,394,378 | fix(ci): remove unnecessary permissions | closed | ### What does this PR do?
Remove unnecessary permissions granted to the actions workflow.
Sorry for the mishap. | 2024-06-10T09:28:02 | 2024-06-11T08:31:52 | 2024-06-11T08:25:47 | https://github.com/huggingface/datasets/pull/6962 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6962",
"html_url": "https://github.com/huggingface/datasets/pull/6962",
"diff_url": "https://github.com/huggingface/datasets/pull/6962.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6962.patch",
"merged_at": "2024-06-11T08:25... | 6,962 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6962). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,342,022,418 | Manual downloads should count as downloads | open | ### Feature request
I would like to request that manual downloads of data files from Hugging Face dataset repositories count as downloads of a dataset. According to the documentation for the Hugging Face Hub, that is currently not the case: https://huggingface.co/docs/hub/en/datasets-download-stats
### Motivation
Th... | 2024-06-09T04:52:06 | 2024-06-13T16:05:00 | null | https://github.com/huggingface/datasets/issues/6961 | null | 6,961 | false | [
"We're unlikely to add more features/support for datasets with python loading scripts, which include datasets with manual download. Sorry for the inconvenience"
] |
2,340,791,685 | feat(ci): add trufflehog secrets detection | closed | ### What does this PR do?
Adding a GH action to scan for leaked secrets on each commit.
| 2024-06-07T16:18:23 | 2024-06-08T14:58:27 | 2024-06-08T14:52:18 | https://github.com/huggingface/datasets/pull/6960 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6960",
"html_url": "https://github.com/huggingface/datasets/pull/6960",
"diff_url": "https://github.com/huggingface/datasets/pull/6960.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6960.patch",
"merged_at": "2024-06-08T14:52... | 6,960 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6960). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Yes!",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\... |
2,340,229,908 | Better error handling in `dataset_module_factory` | closed | cc @cakiki who reported it on [slack](https://huggingface.slack.com/archives/C039P47V1L5/p1717754405578539) (private link)
This PR updates how errors are handled in `dataset_module_factory` when the `dataset_info` cannot be accessed:
1. Use multiple `except ... as e` instead of using `isinstance(e, ...)`
2. Alway... | 2024-06-07T11:24:15 | 2024-06-10T07:33:53 | 2024-06-10T07:27:43 | https://github.com/huggingface/datasets/pull/6959 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6959",
"html_url": "https://github.com/huggingface/datasets/pull/6959",
"diff_url": "https://github.com/huggingface/datasets/pull/6959.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6959.patch",
"merged_at": "2024-06-10T07:27... | 6,959 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6959). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Test should be fixed by https://github.com/huggingface/datasets/pull/6959/commits/ef8f7... |
2,337,476,383 | My Private Dataset doesn't exist on the Hub or cannot be accessed | closed | ### Describe the bug
```
File "/root/miniconda3/envs/gino_conda/lib/python3.9/site-packages/datasets/load.py", line 1852, in dataset_module_factory
raise DatasetNotFoundError(msg + f" at revision '{revision}'" if revision else msg)
datasets.exceptions.DatasetNotFoundError: Dataset 'xxx' doesn't exist on t... | 2024-06-06T06:52:19 | 2024-07-01T11:27:46 | 2024-07-01T11:27:46 | https://github.com/huggingface/datasets/issues/6958 | null | 6,958 | false | [
"I can load public dataset, but for my private dataset it fails",
"https://huggingface.co/docs/datasets/upload_dataset",
"I have checked the API HTTP link. Repository Not Found for url: https://huggingface.co/api/datasets/xxx/xxx.\r\n\r\n. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,333,940,021 | update docs on N-dim arrays | closed | null | 2024-06-04T16:32:19 | 2024-06-04T16:46:34 | 2024-06-04T16:40:27 | https://github.com/huggingface/datasets/pull/6956 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6956",
"html_url": "https://github.com/huggingface/datasets/pull/6956",
"diff_url": "https://github.com/huggingface/datasets/pull/6956.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6956.patch",
"merged_at": "2024-06-04T16:40... | 6,956 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6956). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,333,802,815 | Fix small typo | closed | null | 2024-06-04T15:19:02 | 2024-06-05T10:18:56 | 2024-06-04T15:20:55 | https://github.com/huggingface/datasets/pull/6955 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6955",
"html_url": "https://github.com/huggingface/datasets/pull/6955",
"diff_url": "https://github.com/huggingface/datasets/pull/6955.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6955.patch",
"merged_at": "2024-06-04T15:20... | 6,955 | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
2,333,530,558 | Remove default `trust_remote_code=True` | closed | TODO:
- [x] fix tests | 2024-06-04T13:22:56 | 2024-06-17T16:32:24 | 2024-06-07T12:20:29 | https://github.com/huggingface/datasets/pull/6954 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6954",
"html_url": "https://github.com/huggingface/datasets/pull/6954",
"diff_url": "https://github.com/huggingface/datasets/pull/6954.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6954.patch",
"merged_at": "2024-06-07T12:20... | 6,954 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6954). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"yay! 🎉 ",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<detai... |
2,333,366,120 | Remove canonical datasets from docs | closed | Remove canonical datasets from docs, now that we no longer have canonical datasets. | 2024-06-04T12:09:03 | 2024-07-01T11:31:25 | 2024-07-01T11:31:25 | https://github.com/huggingface/datasets/issues/6953 | null | 6,953 | false | [
"Canonical datasets are no longer mentioned in the docs."
] |
2,333,320,411 | Move info_utils errors to exceptions module | closed | Move `info_utils` errors to `exceptions` module.
Additionally rename some of them, deprecate the former ones, and make the deprecation backward compatible (by making the new errors inherit from the former ones). | 2024-06-04T11:48:32 | 2024-06-10T14:09:59 | 2024-06-10T14:03:55 | https://github.com/huggingface/datasets/pull/6952 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6952",
"html_url": "https://github.com/huggingface/datasets/pull/6952",
"diff_url": "https://github.com/huggingface/datasets/pull/6952.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6952.patch",
"merged_at": "2024-06-10T14:03... | 6,952 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6952). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,333,231,042 | load_dataset() should load all subsets, if no specific subset is specified | closed | ### Feature request
Currently load_dataset() is forcing users to specify a subset. Example
`from datasets import load_dataset
dataset = load_dataset("m-a-p/COIG-CQIA")`
```---------------------------------------------------------------------------
ValueError Traceback (most recen... | 2024-06-04T11:02:33 | 2024-11-26T08:32:18 | 2024-07-01T11:33:10 | https://github.com/huggingface/datasets/issues/6951 | null | 6,951 | false | [
"@xianbaoqian ",
"Feel free to open a PR in `m-a-p/COIG-CQIA` to define a default subset. Currently there is no default.\r\n\r\nYou can find some documentation at https://huggingface.co/docs/hub/datasets-manual-configuration#multiple-configurations",
"@lhoestq \r\n\r\nWhilst having a default subset readily avai... |
2,333,005,974 | `Dataset.with_format` behaves inconsistently with documentation | closed | ### Describe the bug
The actual behavior of the interface `Dataset.with_format` is inconsistent with the documentation.
https://huggingface.co/docs/datasets/use_with_pytorch#n-dimensional-arrays
https://huggingface.co/docs/datasets/v2.19.0/en/use_with_tensorflow#n-dimensional-arrays
> If your dataset consists of ... | 2024-06-04T09:18:32 | 2024-06-25T08:05:49 | 2024-06-25T08:05:49 | https://github.com/huggingface/datasets/issues/6950 | null | 6,950 | false | [
"Hi ! It seems the documentation was outdated in this paragraph\r\n\r\nI fixed it here: https://github.com/huggingface/datasets/pull/6956",
"Fixed."
] |
2,332,336,573 | load_dataset error | closed | ### Describe the bug
Why does the program get stuck when I use load_dataset method, and it still gets stuck after loading for several hours? In fact, my json file is only 21m, and I can load it in one go using open('', 'r').
### Steps to reproduce the bug
1. pip install datasets==2.19.2
2. from datasets import Data... | 2024-06-04T01:24:45 | 2024-07-01T11:33:46 | 2024-07-01T11:33:46 | https://github.com/huggingface/datasets/issues/6949 | null | 6,949 | false | [
"Hi, @lion-ops.\r\n\r\nIn our Continuous Integration we have many tests on loading JSON files and all of them work properly.\r\n\r\nCould you please share your \"train.json\" file, so that we can try to reproduce the issue you have? ",
"> Hi, @lion-ops.\r\n> \r\n> In our Continuous Integration we have many tests ... |
2,331,758,300 | to_tf_dataset: Visible devices cannot be modified after being initialized | open | ### Describe the bug
When trying to use to_tf_dataset with a custom data_loader collate_fn when I use parallelism I am met with the following error as many times as number of workers there were in ``num_workers``.
File "/opt/miniconda/envs/env/lib/python3.11/site-packages/multiprocess/process.py", line 314, in _b... | 2024-06-03T18:10:57 | 2024-06-03T18:10:57 | null | https://github.com/huggingface/datasets/issues/6948 | null | 6,948 | false | [] |
2,331,114,055 | FileNotFoundError:error when loading C4 dataset | closed | ### Describe the bug
can't load c4 datasets
When I replace the datasets package to 2.12.2 I get raise datasets.utils.info_utils.ExpectedMoreSplits: {'train'}
How can I fix this?
### Steps to reproduce the bug
1.from datasets import load_dataset
2.dataset = load_dataset('allenai/c4', data_files={'validat... | 2024-06-03T13:06:33 | 2024-06-25T06:21:28 | 2024-06-25T06:21:28 | https://github.com/huggingface/datasets/issues/6947 | null | 6,947 | false | [
"same problem here",
"Hello,\r\n\r\nAre you sure you are really using datasets version 2.19.2? We just made the patch release yesterday specifically to fix this issue:\r\n- #6925\r\n\r\nI can't reproduce the error:\r\n```python\r\nIn [1]: from datasets import load_dataset\r\n\r\nIn [2]: ds = load_dataset('allenai... |
2,330,276,848 | Re-enable import sorting disabled by flake8:noqa directive when using ruff linter | closed | Re-enable import sorting that was wrongly disabled by `flake8: noqa` directive after switching to `ruff` linter in datasets-2.10.0 PR:
- #5519
Note that after the linter switch, we wrongly replaced `flake8: noqa` with `ruff: noqa` in datasets-2.17.0 PR:
- #6619
That replacement was wrong because we kept the `is... | 2024-06-03T06:24:47 | 2024-06-04T10:00:08 | 2024-06-04T09:54:23 | https://github.com/huggingface/datasets/pull/6946 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6946",
"html_url": "https://github.com/huggingface/datasets/pull/6946",
"diff_url": "https://github.com/huggingface/datasets/pull/6946.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6946.patch",
"merged_at": "2024-06-04T09:54... | 6,946 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6946). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,330,224,869 | Update yanked version of minimum requests requirement | closed | Update yanked version of minimum requests requirement.
Version 2.32.1 was yanked: https://pypi.org/project/requests/2.32.1/ | 2024-06-03T05:45:50 | 2024-06-18T07:36:15 | 2024-06-03T06:09:43 | https://github.com/huggingface/datasets/pull/6945 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6945",
"html_url": "https://github.com/huggingface/datasets/pull/6945",
"diff_url": "https://github.com/huggingface/datasets/pull/6945.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6945.patch",
"merged_at": "2024-06-03T06:09... | 6,945 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6945). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,330,207,120 | Set dev version | closed | null | 2024-06-03T05:29:59 | 2024-06-03T05:37:51 | 2024-06-03T05:31:47 | https://github.com/huggingface/datasets/pull/6944 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6944",
"html_url": "https://github.com/huggingface/datasets/pull/6944",
"diff_url": "https://github.com/huggingface/datasets/pull/6944.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6944.patch",
"merged_at": "2024-06-03T05:31... | 6,944 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6944). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,330,176,890 | Release 2.19.2 | closed | null | 2024-06-03T05:01:50 | 2024-06-03T05:17:41 | 2024-06-03T05:17:40 | https://github.com/huggingface/datasets/pull/6943 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6943",
"html_url": "https://github.com/huggingface/datasets/pull/6943",
"diff_url": "https://github.com/huggingface/datasets/pull/6943.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6943.patch",
"merged_at": "2024-06-03T05:17... | 6,943 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6943). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,329,562,382 | Import sorting is disabled by flake8 noqa directive after switching to ruff linter | closed | When we switched to `ruff` linter in PR:
- #5519
import sorting was disabled in all files containing the `# flake8: noqa` directive
- https://github.com/astral-sh/ruff/issues/11679
We should re-enable import sorting on those files. | 2024-06-02T09:43:34 | 2024-06-04T09:54:24 | 2024-06-04T09:54:24 | https://github.com/huggingface/datasets/issues/6942 | null | 6,942 | false | [] |
2,328,930,165 | Supporting FFCV: Fast Forward Computer Vision | open | ### Feature request
Supporting FFCV, https://github.com/libffcv/ffcv
### Motivation
According to the benchmark, FFCV seems to be fastest image loading method.
### Your contribution
no | 2024-06-01T05:34:52 | 2024-06-01T05:34:52 | null | https://github.com/huggingface/datasets/issues/6941 | null | 6,941 | false | [] |
2,328,637,831 | Enable Sharding to Equal Sized Shards | open | ### Feature request
Add an option when sharding a dataset to have all shards the same size. Will be good to provide both an option of duplication, and by truncation.
### Motivation
Currently the behavior of sharding is "If n % i == l, then the first l shards will have length (n // i) + 1, and the remaining sha... | 2024-05-31T21:55:50 | 2024-06-01T07:34:12 | null | https://github.com/huggingface/datasets/issues/6940 | null | 6,940 | false | [] |
2,328,059,386 | ExpectedMoreSplits error when using data_dir | closed | As reported by @regisss, an `ExpectedMoreSplits` error is raised when passing `data_dir`:
```python
from datasets import load_dataset
dataset = load_dataset(
"lvwerra/stack-exchange-paired",
split="train",
cache_dir=None,
data_dir="data/rl",
)
```
```
Traceback (most recent call last):
F... | 2024-05-31T15:08:42 | 2024-05-31T17:10:39 | 2024-05-31T17:10:39 | https://github.com/huggingface/datasets/issues/6939 | null | 6,939 | false | [] |
2,327,568,281 | Fix expected splits when passing data_files or dir | closed | reported on slack:
The following code snippet gives an error with v2.19 but not with v2.18:
from datasets import load_dataset
```
dataset = load_dataset(
"lvwerra/stack-exchange-paired",
split="train",
cache_dir=None,
data_dir="data/rl",
)
```
and the error is:
```
Traceback (most recent ... | 2024-05-31T11:04:22 | 2024-05-31T15:28:03 | 2024-05-31T15:28:02 | https://github.com/huggingface/datasets/pull/6938 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6938",
"html_url": "https://github.com/huggingface/datasets/pull/6938",
"diff_url": "https://github.com/huggingface/datasets/pull/6938.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6938.patch",
"merged_at": null
} | 6,938 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6938). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"fix is included in https://github.com/huggingface/datasets/pull/6925"
] |
2,327,212,611 | JSON loader implicitly coerces floats to integers | open | The JSON loader implicitly coerces floats to integers.
The column values `[0.0, 1.0, 2.0]` are coerced to `[0, 1, 2]`.
See CI error in dataset-viewer: https://github.com/huggingface/dataset-viewer/actions/runs/9290164936/job/25576926446
```
=================================== FAILURES ===========================... | 2024-05-31T08:09:12 | 2025-06-24T05:49:20 | null | https://github.com/huggingface/datasets/issues/6937 | null | 6,937 | false | [
"Hi @albertvillanova, I'd like to work on this issue if it's still open!\n\nFrom what I see, the float-to-int coercion happens during JSON parsing, possibly due to recent `pandas` behavior. I'll investigate the loading logic inside `json.py` and ensure float values like `[0.0, 1.0, 2.0]` retain their type throughou... |
2,326,119,853 | save_to_disk() freezes when saving on s3 bucket with multiprocessing | open | ### Describe the bug
I'm trying to save a `Dataset` using the `save_to_disk()` function with:
- `num_proc > 1`
- `dataset_path` being a s3 bucket path e.g. "s3://{bucket_name}/{dataset_folder}/"
The hf progress bar shows up but the saving does not seem to start.
When using one processor only (`num_proc=1`), e... | 2024-05-30T16:48:39 | 2025-02-06T22:12:52 | null | https://github.com/huggingface/datasets/issues/6936 | null | 6,936 | false | [
"I got the same issue. Any updates so far for this issue?",
"Same here. Any updates?",
"+1, experiencing this as well"
] |
2,325,612,022 | Support for pathlib.Path in datasets 2.19.0 | open | ### Describe the bug
After the recent update of `datasets`, Dataset.save_to_disk does not accept a pathlib.Path anymore. It was supported in 2.18.0 and previous versions. Is this intentional? Was it supported before only because of a Python dusk-typing miracle?
### Steps to reproduce the bug
```
from datasets impor... | 2024-05-30T12:53:36 | 2025-01-14T11:50:22 | null | https://github.com/huggingface/datasets/issues/6935 | null | 6,935 | false | [
"+1 I just noticed this when I tried to update `datasets` today.",
"The same issue, I also get error."
] |
2,325,341,717 | Revert ci user | closed | null | 2024-05-30T10:45:26 | 2024-05-31T10:25:08 | 2024-05-30T10:45:37 | https://github.com/huggingface/datasets/pull/6934 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6934",
"html_url": "https://github.com/huggingface/datasets/pull/6934",
"diff_url": "https://github.com/huggingface/datasets/pull/6934.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6934.patch",
"merged_at": "2024-05-30T10:45... | 6,934 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6934). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,325,300,800 | update ci user | closed | token is ok to be public since it's only for the hub-ci | 2024-05-30T10:23:02 | 2024-05-30T10:30:54 | 2024-05-30T10:23:12 | https://github.com/huggingface/datasets/pull/6933 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6933",
"html_url": "https://github.com/huggingface/datasets/pull/6933",
"diff_url": "https://github.com/huggingface/datasets/pull/6933.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6933.patch",
"merged_at": "2024-05-30T10:23... | 6,933 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6933). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,324,729,267 | Update dataset_dict.py | closed | shape returns (number of rows, number of columns) | 2024-05-30T05:22:35 | 2024-06-04T12:56:20 | 2024-06-04T12:50:13 | https://github.com/huggingface/datasets/pull/6932 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6932",
"html_url": "https://github.com/huggingface/datasets/pull/6932",
"diff_url": "https://github.com/huggingface/datasets/pull/6932.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6932.patch",
"merged_at": "2024-06-04T12:50... | 6,932 | true | [
"thanks !",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_bat... |
2,323,457,525 | [WebDataset] Support compressed files | closed | null | 2024-05-29T14:19:06 | 2024-05-29T16:33:18 | 2024-05-29T16:24:21 | https://github.com/huggingface/datasets/pull/6931 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6931",
"html_url": "https://github.com/huggingface/datasets/pull/6931",
"diff_url": "https://github.com/huggingface/datasets/pull/6931.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6931.patch",
"merged_at": "2024-05-29T16:24... | 6,931 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6931). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,323,225,922 | ValueError: Couldn't infer the same data file format for all splits. Got {'train': ('json', {}), 'validation': (None, {})} | open | ### Describe the bug
When I run the code en = load_dataset("allenai/c4", "en", streaming=True), I encounter an error: raise ValueError(f"Couldn't infer the same data file format for all splits. Got {split_modules}") ValueError: Couldn't infer the same data file format for all splits. Got {'train': ('json', {}), 'valid... | 2024-05-29T12:40:05 | 2024-07-23T06:25:24 | null | https://github.com/huggingface/datasets/issues/6930 | null | 6,930 | false | [
"How do you solve it ?\r\n",
"> How do you solve it ?\r\n\r\nPlease check your Python environment and dataset version. I have just resolved the issue, which was caused by a Python environment switching error\r\n"
] |
2,322,980,077 | Avoid downloading the whole dataset when only README.me has been touched on hub. | open | ### Feature request
`datasets.load_dataset()` triggers a new download of the **whole dataset** when the README.md file has been touched on huggingface hub, even if data files / parquet files are the exact same.
I think the current behaviour of the load_dataset function is triggered whenever a change of the hash o... | 2024-05-29T10:36:06 | 2024-05-29T20:51:56 | null | https://github.com/huggingface/datasets/issues/6929 | null | 6,929 | false | [
"you're right, we're tackling this here: https://github.com/huggingface/dataset-viewer/issues/2757",
"@severo : great !"
] |
2,322,267,727 | Update process.mdx: Code Listings Fixes | closed | null | 2024-05-29T03:17:07 | 2024-06-04T13:08:19 | 2024-06-04T12:55:00 | https://github.com/huggingface/datasets/pull/6928 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6928",
"html_url": "https://github.com/huggingface/datasets/pull/6928",
"diff_url": "https://github.com/huggingface/datasets/pull/6928.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6928.patch",
"merged_at": "2024-06-04T12:55... | 6,928 | true | [
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a... |
2,322,260,725 | Update process.mdx: Minor Code Listings Updates and Fixes | closed | null | 2024-05-29T03:09:01 | 2024-05-29T03:12:46 | 2024-05-29T03:12:46 | https://github.com/huggingface/datasets/pull/6927 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6927",
"html_url": "https://github.com/huggingface/datasets/pull/6927",
"diff_url": "https://github.com/huggingface/datasets/pull/6927.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6927.patch",
"merged_at": null
} | 6,927 | true | [] |
2,322,164,287 | Update process.mdx: Fix code listing in Shard section | closed | null | 2024-05-29T01:25:55 | 2024-05-29T03:11:20 | 2024-05-29T03:11:08 | https://github.com/huggingface/datasets/pull/6926 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6926",
"html_url": "https://github.com/huggingface/datasets/pull/6926",
"diff_url": "https://github.com/huggingface/datasets/pull/6926.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6926.patch",
"merged_at": null
} | 6,926 | true | [] |
2,321,084,967 | Fix NonMatchingSplitsSizesError/ExpectedMoreSplits when passing data_dir/data_files in no-code Hub datasets | closed | Fix `NonMatchingSplitsSizesError` or `ExpectedMoreSplits` error for no-code Hub datasets if the user passes:
- `data_dir`
- `data_files`
The proposed solution is to avoid using exported dataset info (from Parquet exports) in these cases.
Additionally, also if the user passes `revision` other than "main" (so that ... | 2024-05-28T13:33:38 | 2024-11-07T20:41:58 | 2024-05-31T17:10:37 | https://github.com/huggingface/datasets/pull/6925 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6925",
"html_url": "https://github.com/huggingface/datasets/pull/6925",
"diff_url": "https://github.com/huggingface/datasets/pull/6925.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6925.patch",
"merged_at": "2024-05-31T17:10... | 6,925 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6925). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Do you think this is worth making a patch release for?\r\nCC: @huggingface/datasets",
... |
2,320,531,015 | Caching map result of DatasetDict. | open | Hi!
I'm currenty using the map function to tokenize a somewhat large dataset, so I need to use the cache to save ~25 mins.
Changing num_proc incduces the recomputation of the map, I'm not sure why and if this is excepted behavior?
here it says, that cached files are loaded sequentially:
https://github.com/... | 2024-05-28T09:07:41 | 2024-05-28T09:07:41 | null | https://github.com/huggingface/datasets/issues/6924 | null | 6,924 | false | [] |
2,319,292,872 | Export Parquet Tablet Audio-Set is null bytes in Arrow | open | ### Describe the bug
Exporting the processed audio inside the table with the dataset.to_parquet function, the object pyarrow {bytes: null, path: "Some/Path"}
At the same time, the same dataset uploaded to the hub has bit arrays
. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,318,394,398 | Support fsspec 2024.5.0 | closed | Support fsspec 2024.5.0. | 2024-05-27T07:00:59 | 2024-05-27T08:07:16 | 2024-05-27T08:01:08 | https://github.com/huggingface/datasets/pull/6921 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6921",
"html_url": "https://github.com/huggingface/datasets/pull/6921",
"diff_url": "https://github.com/huggingface/datasets/pull/6921.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6921.patch",
"merged_at": "2024-05-27T08:01... | 6,921 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6921). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,317,648,021 | [WebDataset] Add `.pth` support for torch tensors | closed | In this PR I add support for `.pth` but with `weights_only=True` to disallow the use of pickle | 2024-05-26T11:12:07 | 2024-05-27T09:11:17 | 2024-05-27T09:04:54 | https://github.com/huggingface/datasets/pull/6920 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6920",
"html_url": "https://github.com/huggingface/datasets/pull/6920",
"diff_url": "https://github.com/huggingface/datasets/pull/6920.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6920.patch",
"merged_at": "2024-05-27T09:04... | 6,920 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6920). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,315,618,993 | Invalid YAML in README.md: unknown tag !<tag:yaml.org,2002:python/tuple> | open | ### Describe the bug
I wrote a notebook to load an existing dataset, process it, and upload as a private dataset using `dataset.push_to_hub(...)` at the end. The push to hub is failing with:
```
ValueError: Invalid metadata in README.md.
- Invalid YAML in README.md: unknown tag !<tag:yaml.org,2002:python[/tuple](... | 2024-05-24T14:59:45 | 2024-05-24T14:59:45 | null | https://github.com/huggingface/datasets/issues/6919 | null | 6,919 | false | [] |
2,315,322,738 | NonMatchingSplitsSizesError when using data_dir | closed | ### Describe the bug
Loading a dataset from with a data_dir argument generates a NonMatchingSplitsSizesError if there are multiple directories in the dataset.
This appears to happen because the expected split is calculated based on the data in all the directories whereas the recorded split is calculated based on t... | 2024-05-24T12:43:39 | 2024-05-31T17:10:38 | 2024-05-31T17:10:38 | https://github.com/huggingface/datasets/issues/6918 | null | 6,918 | false | [
"Thanks for reporting, @srehaag.\r\n\r\nWe are investigating this issue.",
"I confirm there is a bug for data-based Hub datasets when the user passes `data_dir`, which was introduced by PR:\r\n- #6714"
] |
2,314,683,663 | WinError 32 The process cannot access the file during load_dataset | open | ### Describe the bug
When I try to load the opus_book from hugging face (following the [guide on the website](https://huggingface.co/docs/transformers/main/en/tasks/translation))
```python
from datasets import load_dataset, Dataset
dataset = load_dataset("Helsinki-NLP/opus_books", "en-fr", features=["id", "tran... | 2024-05-24T07:54:51 | 2024-05-24T07:54:51 | null | https://github.com/huggingface/datasets/issues/6917 | null | 6,917 | false | [] |
2,311,675,564 | ```push_to_hub()``` - Prevent Automatic Generation of Splits | closed | ### Describe the bug
I currently have a dataset which has not been splited. When pushing the dataset to my hugging face dataset repository, it is split into a testing and training set. How can I prevent the split from happening?
### Steps to reproduce the bug
1. Have a unsplit dataset
```python
Dataset({ featur... | 2024-05-22T23:52:15 | 2024-05-23T00:07:53 | 2024-05-23T00:07:53 | https://github.com/huggingface/datasets/issues/6916 | null | 6,916 | false | [] |
2,310,564,961 | Validate config name and data_files in packaged modules | closed | Validate the config attributes `name` and `data_files` in packaged modules by making the derived classes call their parent `__post_init__` method.
Note that their parent `BuilderConfig` validates its attributes `name` and `data_files` in its `__post_init__` method: https://github.com/huggingface/datasets/blob/60d21e... | 2024-05-22T13:36:33 | 2024-06-06T09:32:10 | 2024-06-06T09:24:35 | https://github.com/huggingface/datasets/pull/6915 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6915",
"html_url": "https://github.com/huggingface/datasets/pull/6915",
"diff_url": "https://github.com/huggingface/datasets/pull/6915.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6915.patch",
"merged_at": "2024-06-06T09:24... | 6,915 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6915). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"I pushed a change that fixes 2.15 cache reloading (I fixed the packaged module hash), f... |
2,310,107,326 | Preserve JSON column order and support list of strings field | closed | Preserve column order when loading from a JSON file with a list of dict (or with a field containing a list of dicts).
Additionally, support JSON file with a list of strings field.
Fix #6913. | 2024-05-22T09:58:54 | 2024-05-29T13:18:47 | 2024-05-29T13:12:23 | https://github.com/huggingface/datasets/pull/6914 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6914",
"html_url": "https://github.com/huggingface/datasets/pull/6914",
"diff_url": "https://github.com/huggingface/datasets/pull/6914.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6914.patch",
"merged_at": "2024-05-29T13:12... | 6,914 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6914). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,309,605,889 | Column order is nondeterministic when loading from JSON | closed | As reported by @meg-huggingface, the order of the JSON object keys is not preserved while loading a dataset from a JSON file with a list of objects.
For example, when loading a JSON files with a list of objects, each with the following ordered keys:
- [ID, Language, Topic],
the resulting dataset may have column... | 2024-05-22T05:30:14 | 2024-05-29T13:12:24 | 2024-05-29T13:12:24 | https://github.com/huggingface/datasets/issues/6913 | null | 6,913 | false | [] |
2,309,365,961 | Add MedImg for streaming | open | ### Feature request
Host the MedImg dataset (similar to Imagenet but for biomedical images).
### Motivation
There is a clear need for biomedical image foundation models and large scale biomedical datasets that are easily streamable. This would be an excellent tool for the biomedical community.
### Your con... | 2024-05-22T00:55:30 | 2024-09-05T16:53:54 | null | https://github.com/huggingface/datasets/issues/6912 | null | 6,912 | false | [
"@mariosasko, @lhoestq, @albertvillanova\r\nHello! Can anyone help? or can you guys suggest who can help with this?",
"Hi ! Feel free to download the dataset and create a `Dataset` object with it.\r\n\r\nThen your'll be able to use `push_to_hub()` to upload the dataset to HF in Parquet format and make it streamab... |
2,308,152,711 | Remove dead code for non-dict data_files from packaged modules | closed | Remove dead code for non-dict data_files from packaged modules.
Since the merge of this PR:
- #2986
the builders' variable self.config.data_files is always a dict, which makes the condition on (str, list, tuple) dead code. | 2024-05-21T12:10:24 | 2024-05-23T08:05:58 | 2024-05-23T07:59:57 | https://github.com/huggingface/datasets/pull/6911 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6911",
"html_url": "https://github.com/huggingface/datasets/pull/6911",
"diff_url": "https://github.com/huggingface/datasets/pull/6911.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6911.patch",
"merged_at": "2024-05-23T07:59... | 6,911 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6911). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,307,570,084 | Fix wrong type hints in data_files | closed | Fix wrong type hints in data_files introduced in:
- #6493 | 2024-05-21T07:41:09 | 2024-05-23T06:04:05 | 2024-05-23T05:58:05 | https://github.com/huggingface/datasets/pull/6910 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6910",
"html_url": "https://github.com/huggingface/datasets/pull/6910",
"diff_url": "https://github.com/huggingface/datasets/pull/6910.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6910.patch",
"merged_at": "2024-05-23T05:58... | 6,910 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6910). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,307,508,120 | Update requests >=2.32.1 to fix vulnerability | closed | Update requests >=2.32.1 to fix vulnerability. | 2024-05-21T07:11:20 | 2024-05-21T07:45:58 | 2024-05-21T07:38:25 | https://github.com/huggingface/datasets/pull/6909 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6909",
"html_url": "https://github.com/huggingface/datasets/pull/6909",
"diff_url": "https://github.com/huggingface/datasets/pull/6909.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6909.patch",
"merged_at": "2024-05-21T07:38... | 6,909 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6909). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,304,958,116 | Fail to load "stas/c4-en-10k" dataset since 2.16 version | closed | ### Describe the bug
When update datasets library to version 2.16+ ( I test it on 2.16, 2.19.0 and 2.19.1), using the following code to load stas/c4-en-10k dataset
```python
from datasets import load_dataset, Dataset
dataset = load_dataset('stas/c4-en-10k')
```
and then it raise UnicodeDecodeError like
... | 2024-05-20T02:43:59 | 2024-05-24T10:58:09 | 2024-05-24T10:58:09 | https://github.com/huggingface/datasets/issues/6908 | null | 6,908 | false | [
"I am not able to reproduce the error with datasets 2.19.1:\r\n```python\r\nIn [1]: from datasets import load_dataset; ds = load_dataset(\"stas/c4-en-10k\", streaming=True); item = next(iter(ds[\"train\"])); item\r\nOut[1]: {'text': 'Beginners BBQ Class Taking Place in Missoula!\\nDo you want to get better at makin... |
2,303,855,833 | Support the deserialization of json lines files comprised of lists | open | ### Feature request
I manage a somewhat large and popular Hugging Face dataset known as the [Open Australian Legal Corpus](https://huggingface.co/datasets/umarbutler/open-australian-legal-corpus). I recently updated my corpus to be stored in a json lines file where each line is an array and each element represents a v... | 2024-05-18T05:07:23 | 2024-05-18T08:53:28 | null | https://github.com/huggingface/datasets/issues/6907 | null | 6,907 | false | [
"Update: I ended up deciding to go back to use lines of dictionaries instead of arrays, not because of this issue as my users would be capable of downloading my corpus without `datasets`, but the speed and storage savings are not currently worth breaking my API and harming the backwards compatibility of each new re... |
2,303,679,119 | irc_disentangle - Issue with splitting data | closed | ### Describe the bug
I am trying to access your database through python using "datasets.load_dataset("irc_disentangle")" and I am getting this error message:
ValueError: Instruction "train" corresponds to no data!
### Steps to reproduce the bug
import datasets
ds = datasets.load_dataset('irc_disentangle')
ds
#... | 2024-05-17T23:19:37 | 2024-07-16T00:21:56 | 2024-07-08T06:18:08 | https://github.com/huggingface/datasets/issues/6906 | null | 6,906 | false | [
"Thank you I will try this out!\r\n\r\nOn Tue, Jun 11, 2024 at 3:55 AM Vincent Lau ***@***.***>\r\nwrote:\r\n\r\n> I add a \"streaming=True\" after the name of the dataset, and it\r\n> works.....hope it can help you\r\n>\r\n> And if you install the version datasets==2.15.0, this bug will not happen.\r\n> I don't kn... |
2,303,098,587 | Extraction protocol for arrow files is not defined | closed | ### Describe the bug
Passing files with `.arrow` extension into data_files argument, at least when `streaming=True` is very slow.
### Steps to reproduce the bug
Basically it goes through the `_get_extraction_protocol` method located [here](https://github.com/huggingface/datasets/blob/main/src/datasets/utils/file_ut... | 2024-05-17T16:01:41 | 2025-02-06T19:50:22 | 2025-02-06T19:50:20 | https://github.com/huggingface/datasets/issues/6905 | null | 6,905 | false | [
"Fixed in https://github.com/huggingface/datasets/pull/7083"
] |
2,302,912,179 | Fix decoding multi part extension | closed | e.g. a field named `url.txt` should be a treated as text
I also included a small fix to support .npz correctly | 2024-05-17T14:32:57 | 2024-05-17T14:52:56 | 2024-05-17T14:46:54 | https://github.com/huggingface/datasets/pull/6904 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6904",
"html_url": "https://github.com/huggingface/datasets/pull/6904",
"diff_url": "https://github.com/huggingface/datasets/pull/6904.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6904.patch",
"merged_at": "2024-05-17T14:46... | 6,904 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6904). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"takign the liberty to merge this for the viewer and a new dataset being released",
"<... |
2,300,436,053 | Add the option of saving in parquet instead of arrow | open | ### Feature request
In dataset.save_to_disk('/path/to/save/dataset'),
add the option to save in parquet format
dataset.save_to_disk('/path/to/save/dataset', format="parquet"),
because arrow is not used for Production Big data.... (only parquet)
### Motivation
because arrow is not used for Production Big... | 2024-05-16T13:35:51 | 2025-05-19T12:14:14 | null | https://github.com/huggingface/datasets/issues/6903 | null | 6,903 | false | [
"I think [`Dataset.to_parquet`](https://huggingface.co/docs/datasets/v1.10.2/package_reference/main_classes.html#datasets.Dataset.to_parquet) is what you're looking for.\r\n\r\nLet me know if I'm wrong ",
"No, it does not save the metadata json.\r\n\r\nWe have to recode all meta json load/save\r\nwith another cus... |
2,300,256,241 | Make CLI convert_to_parquet not raise error if no rights to create script branch | closed | Make CLI convert_to_parquet not raise error if no rights to create "script" branch.
Not that before this PR, the error was not critical because it was raised at the end of the script, once all the rest of the steps were already performed.
Fix #6901.
Bug introduced in datasets-2.19.0 by:
- #6809 | 2024-05-16T12:21:27 | 2024-06-03T04:43:17 | 2024-05-16T12:51:05 | https://github.com/huggingface/datasets/pull/6902 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6902",
"html_url": "https://github.com/huggingface/datasets/pull/6902",
"diff_url": "https://github.com/huggingface/datasets/pull/6902.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6902.patch",
"merged_at": "2024-05-16T12:51... | 6,902 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6902). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,300,167,465 | HTTPError 403 raised by CLI convert_to_parquet when creating script branch on 3rd party repos | closed | CLI convert_to_parquet cannot create "script" branch on 3rd party repos.
It can only create it on repos where the user executing the script has write access.
Otherwise, a 403 Forbidden HTTPError is raised:
```
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/huggingface_hub/ut... | 2024-05-16T11:40:22 | 2024-05-16T12:51:06 | 2024-05-16T12:51:06 | https://github.com/huggingface/datasets/issues/6901 | null | 6,901 | false | [] |
2,298,489,733 | [WebDataset] KeyError with user-defined `Features` when a field is missing in an example | closed | reported at https://huggingface.co/datasets/ProGamerGov/synthetic-dataset-1m-dalle3-high-quality-captions/discussions/1
```
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/webdataset/webdataset.py", line 109, in _generate_examples
example[field_name] = {"path": example["_... | 2024-05-15T17:48:34 | 2024-06-28T09:30:13 | 2024-06-28T09:30:13 | https://github.com/huggingface/datasets/issues/6900 | null | 6,900 | false | [
"@lhoestq How difficult of fix is this?",
"It shouldn't be difficult, I think it's just a matter of adding the missing fields from `self.config.features` in `example` here: before it iterates on image_field_names and audio_field_names. A missing field should have a value set to None\r\n\r\nhttps://github.com/hugg... |
2,298,059,597 | List of dictionary features get standardized | open | ### Describe the bug
Hi, i’m trying to create a HF dataset from a list using Dataset.from_list.
Each sample in the list is a dict with the same keys (which will be my features). The values for each feature are a list of dictionaries, and each such dictionary has a different set of keys. However, the datasets librar... | 2024-05-15T14:11:35 | 2025-04-01T20:48:03 | null | https://github.com/huggingface/datasets/issues/6899 | null | 6,899 | false | [
"I think this may be a limitation of the arrow format",
"Dupe of #5950\n"
] |
2,294,432,108 | Fix YAML error in README files appearing on GitHub | closed | Fix YAML error in README files appearing on GitHub.
See error message:

Fix #6897. | 2024-05-14T05:21:57 | 2024-05-16T14:36:57 | 2024-05-16T14:28:16 | https://github.com/huggingface/datasets/pull/6898 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6898",
"html_url": "https://github.com/huggingface/datasets/pull/6898",
"diff_url": "https://github.com/huggingface/datasets/pull/6898.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6898.patch",
"merged_at": "2024-05-16T14:28... | 6,898 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6898). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"After this PR, the README file looks like:\r\n\r\n
2. Observe a big red error at the top
3. The rest of the ... | 2024-05-13T17:33:59 | 2024-05-16T14:28:17 | 2024-05-16T14:28:17 | https://github.com/huggingface/datasets/issues/6897 | null | 6,897 | false | [
"Hello, @bghira.\r\n\r\nThanks for reporting. Please note that the text originating the error is not supposed to be valid YAML: it contains the instructions to generate the actual YAML content, that should replace the instructions comment.\r\n\r\nOn the other hand, I agree that it is not nice to have that YAML erro... |
2,293,176,061 | Regression bug: `NonMatchingSplitsSizesError` for (possibly) overwritten dataset | open | ### Describe the bug
While trying to load the dataset `https://huggingface.co/datasets/pysentimiento/spanish-tweets-small`, I get this error:
```python
---------------------------------------------------------------------------
NonMatchingSplitsSizesError Traceback (most recent call last)
[<ipyth... | 2024-05-13T15:41:57 | 2025-03-25T01:21:06 | null | https://github.com/huggingface/datasets/issues/6896 | null | 6,896 | false | [
"Same issue here\n"
] |
2,292,993,156 | Document that to_json defaults to JSON Lines | closed | Document that `Dataset.to_json` defaults to JSON Lines, by adding explanation in the corresponding docstring.
Fix #6894. | 2024-05-13T14:22:34 | 2024-05-16T14:37:25 | 2024-05-16T14:31:26 | https://github.com/huggingface/datasets/pull/6895 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6895",
"html_url": "https://github.com/huggingface/datasets/pull/6895",
"diff_url": "https://github.com/huggingface/datasets/pull/6895.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6895.patch",
"merged_at": "2024-05-16T14:31... | 6,895 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6895). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,292,840,226 | Better document defaults of to_json | closed | Better document defaults of `to_json`: the default format is [JSON-Lines](https://jsonlines.org/).
Related to:
- #6891 | 2024-05-13T13:30:54 | 2024-05-16T14:31:27 | 2024-05-16T14:31:27 | https://github.com/huggingface/datasets/issues/6894 | null | 6,894 | false | [] |
2,292,677,439 | Close gzipped files properly | closed | close https://github.com/huggingface/datasets/issues/6877 | 2024-05-13T12:24:39 | 2024-05-13T13:53:17 | 2024-05-13T13:01:54 | https://github.com/huggingface/datasets/pull/6893 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6893",
"html_url": "https://github.com/huggingface/datasets/pull/6893",
"diff_url": "https://github.com/huggingface/datasets/pull/6893.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6893.patch",
"merged_at": "2024-05-13T13:01... | 6,893 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6893). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,291,201,347 | Add support for categorical/dictionary types | closed | Arrow has a very useful dictionary/categorical type (https://arrow.apache.org/docs/python/generated/pyarrow.dictionary.html). This data type has significant speed, memory and disk benefits over pa.string() when there are only a few unique text strings in a column.
Unfortunately, huggingface datasets currently does n... | 2024-05-12T07:15:08 | 2024-06-07T15:01:39 | 2024-06-07T12:20:42 | https://github.com/huggingface/datasets/pull/6892 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6892",
"html_url": "https://github.com/huggingface/datasets/pull/6892",
"diff_url": "https://github.com/huggingface/datasets/pull/6892.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6892.patch",
"merged_at": "2024-06-07T12:20... | 6,892 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6892). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,291,118,869 | Unable to load JSON saved using `to_json` | closed | ### Describe the bug
Datasets stored in the JSON format cannot be loaded using `json.load()`
### Steps to reproduce the bug
```
import json
from datasets import load_dataset
dataset = load_dataset("squad")
train_dataset, test_dataset = dataset["train"], dataset["validation"]
test_dataset.to_json("full_dataset... | 2024-05-12T01:02:51 | 2024-05-16T14:32:55 | 2024-05-12T07:02:02 | https://github.com/huggingface/datasets/issues/6891 | null | 6,891 | false | [
"Hi @DarshanDeshpande,\r\n\r\nPlease note that the default format of the method `Dataset.to_json` is [JSON-Lines](https://jsonlines.org/): it passes `orient=\"records\", lines=True` to `pandas.DataFrame.to_json`. This format is specially useful for large datasets, since unlike regular JSON files, it does not requir... |
2,288,699,041 | add `with_transform` and/or `set_transform` to IterableDataset | open | ### Feature request
when working with a really large dataset it would save us a lot of time (and compute resources) to use either with_transform or the set_transform from the Dataset class instead of waiting for the entire dataset to map
### Motivation
don't want to wait for a really long dataset to map, this would ... | 2024-05-10T01:00:12 | 2024-05-10T01:00:46 | null | https://github.com/huggingface/datasets/issues/6890 | null | 6,890 | false | [] |
2,287,720,539 | fix bug #6877 | closed | fix bug #6877 due to maybe f becomes invaild after yield process
the results are below:
Resolving data files: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 828/828 [00:01<00:00, 420.41it/s]
Resolving data files: 100%|████████... | 2024-05-09T13:38:40 | 2024-05-13T13:35:32 | 2024-05-13T13:35:32 | https://github.com/huggingface/datasets/pull/6889 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6889",
"html_url": "https://github.com/huggingface/datasets/pull/6889",
"diff_url": "https://github.com/huggingface/datasets/pull/6889.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6889.patch",
"merged_at": null
} | 6,889 | true | [
"@loicmagne, @KennethEnevoldsen",
"Can you give more details on why this fix works ?",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6889). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",... |
2,287,169,676 | Support WebDataset containing file basenames with dots | closed | Support WebDataset containing file basenames with dots.
Fix #6880. | 2024-05-09T08:25:30 | 2024-05-10T13:54:06 | 2024-05-10T13:54:06 | https://github.com/huggingface/datasets/pull/6888 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6888",
"html_url": "https://github.com/huggingface/datasets/pull/6888",
"diff_url": "https://github.com/huggingface/datasets/pull/6888.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6888.patch",
"merged_at": null
} | 6,888 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6888). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"I think webdataset splits the file name and extension using the first dot no ?\r\n\r\nh... |
2,286,786,396 | FAISS load to None | open | ### Describe the bug
I've use FAISS with Datasets and save to FAISS.
Then load to save FAISS then no error, then ds to None
```python
ds.load_faiss_index('embeddings', 'my_index.faiss')
```
### Steps to reproduce the bug
# 1.
```python
ds_with_embeddings = ds.map(lambda example: {'embeddings': model(transf... | 2024-05-09T02:43:50 | 2024-05-16T20:44:23 | null | https://github.com/huggingface/datasets/issues/6887 | null | 6,887 | false | [
"Hello,\r\n\r\nI'm not sure I understand. \r\nThe return value of `ds.load_faiss_index` is None as expected.\r\n\r\nI see that loading an Index on a dataset that doesn't have an `embedding` column doesn't raise an Issue. Is that the issue?\r\n\r\nSo `ds` doesn't have an `embedding` column, but we load an index that... |
2,286,328,984 | load_dataset with data_dir and cache_dir set fail with not supported | open | ### Describe the bug
with python 3.11 I execute:
```py
from transformers import Wav2Vec2Processor, Data2VecAudioModel
import torch
from torch import nn
from datasets import load_dataset, concatenate_datasets
# load demo audio and set processor
dataset_clean = load_dataset("librispeech_asr", "clean", split="... | 2024-05-08T19:52:35 | 2024-05-08T19:58:11 | null | https://github.com/huggingface/datasets/issues/6886 | null | 6,886 | false | [] |
2,285,115,400 | Support jax 0.4.27 in CI tests | closed | Support jax 0.4.27 in CI tests by using jax Array `devices` method instead of `device` (which no longer exists).
Fix #6884. | 2024-05-08T09:19:37 | 2024-05-08T09:43:19 | 2024-05-08T09:35:16 | https://github.com/huggingface/datasets/pull/6885 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6885",
"html_url": "https://github.com/huggingface/datasets/pull/6885",
"diff_url": "https://github.com/huggingface/datasets/pull/6885.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6885.patch",
"merged_at": "2024-05-08T09:35... | 6,885 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6885). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,284,839,687 | CI is broken after jax-0.4.27 release: AttributeError: 'jaxlib.xla_extension.DeviceList' object has no attribute 'device' | closed | After jax-0.4.27 release (https://github.com/google/jax/releases/tag/jax-v0.4.27), our CI is broken with the error:
```Python traceback
AttributeError: 'jaxlib.xla_extension.DeviceList' object has no attribute 'device'. Did you mean: 'devices'?
```
See: https://github.com/huggingface/datasets/actions/runs/8997488... | 2024-05-08T07:01:47 | 2024-05-08T09:35:17 | 2024-05-08T09:35:17 | https://github.com/huggingface/datasets/issues/6884 | null | 6,884 | false | [] |
2,284,808,399 | Require Pillow >= 9.4.0 to avoid AttributeError when loading image dataset | closed | Require Pillow >= 9.4.0 to avoid AttributeError when loading image dataset.
The `PIL.Image.ExifTags` that we use in our code was implemented in Pillow-9.4.0: https://github.com/python-pillow/Pillow/commit/24a5405a9f7ea22f28f9c98b3e407292ea5ee1d3
The bug #6881 was introduced in datasets-2.19.0 by this PR:
- #6739... | 2024-05-08T06:43:29 | 2024-08-28T13:13:57 | 2024-05-16T14:34:02 | https://github.com/huggingface/datasets/pull/6883 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6883",
"html_url": "https://github.com/huggingface/datasets/pull/6883",
"diff_url": "https://github.com/huggingface/datasets/pull/6883.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6883.patch",
"merged_at": "2024-05-16T14:34... | 6,883 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6883). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Do you think this is worth making a patch release for?\r\nCC: @huggingface/datasets ",
... |
2,284,803,158 | Connection Error When Using By-pass Proxies | open | ### Describe the bug
I'm currently using Clash for Windows as my proxy tunnel, after exporting HTTP_PROXY and HTTPS_PROXY to the port that clash provides🤔, it runs into a connection error saying "Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/2.19.1/metrics/seqeval/seqeval.py (ConnectionError(M... | 2024-05-08T06:40:14 | 2024-05-17T06:38:30 | null | https://github.com/huggingface/datasets/issues/6882 | null | 6,882 | false | [
"Changing the supplier of the proxy will solve this problem, or you can visit and follow the instructions in https://hf-mirror.com "
] |
2,284,794,009 | AttributeError: module 'PIL.Image' has no attribute 'ExifTags' | closed | When trying to load an image dataset in an old Python environment (with Pillow-8.4.0), an error is raised:
```Python traceback
AttributeError: module 'PIL.Image' has no attribute 'ExifTags'
```
The error traceback:
```Python traceback
~/huggingface/datasets/src/datasets/iterable_dataset.py in __iter__(self)
1... | 2024-05-08T06:33:57 | 2024-07-18T06:49:30 | 2024-05-16T14:34:03 | https://github.com/huggingface/datasets/issues/6881 | null | 6,881 | false | [
"@albertvillanova @lhoestq just ran into it and requiring newer pillow isn't a solution as it breaks Pillow-SIMD which is behind Pillow quite a few versions but necessary for training with reasonable throughput. \r\n\r\nA couple things here... \r\n\r\n1. This can be done with a method that isn't an issue for any so... |
2,283,278,337 | Webdataset: KeyError: 'png' on some datasets when streaming | open | reported at https://huggingface.co/datasets/tbone5563/tar_images/discussions/1
```python
>>> from datasets import load_dataset
>>> ds = load_dataset("tbone5563/tar_images")
Downloading data: 100%
1.41G/1.41G [00:48<00:00, 17.2MB/s]
Downloading data: 100%
619M/619M [00:11<00:00, 57.4MB/s]
Generating train sp... | 2024-05-07T13:09:02 | 2024-05-14T20:34:05 | null | https://github.com/huggingface/datasets/issues/6880 | null | 6,880 | false | [
"The error is caused by malformed basenames of the files within the TARs:\r\n- `15_Cohen_1-s2.0-S0929664620300449-gr3_lrg-b.png` becomes `15_Cohen_1-s2` as the grouping `__key__`, and `0-S0929664620300449-gr3_lrg-b.png` as the additional key to be added to the example\r\n- whereas the intended behavior was to use `... |
2,282,968,259 | Batched mapping does not raise an error if values for an existing column are empty | open | ### Describe the bug
Using `Dataset.map(fn, batched=True)` allows resizing the dataset by returning a dict of lists, all of which must be the same size. If they are not the same size, an error like `pyarrow.lib.ArrowInvalid: Column 1 named x expected length 1 but got length 0` is raised.
This is not the case if the... | 2024-05-07T11:02:40 | 2024-05-07T11:02:40 | null | https://github.com/huggingface/datasets/issues/6879 | null | 6,879 | false | [] |
2,282,879,491 | Create function to convert to parquet | closed | Analogously with `delete_from_hub`, this PR:
- creates the Python function `convert_to_parquet`
- makes the corresponding CLI command use that function.
This way, the functionality can be used both from a terminal and from a Python console.
This PR also implements a test for convert_to_parquet function. | 2024-05-07T10:27:07 | 2024-05-16T14:46:44 | 2024-05-16T14:38:23 | https://github.com/huggingface/datasets/pull/6878 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6878",
"html_url": "https://github.com/huggingface/datasets/pull/6878",
"diff_url": "https://github.com/huggingface/datasets/pull/6878.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6878.patch",
"merged_at": "2024-05-16T14:38... | 6,878 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6878). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,282,068,337 | OSError: [Errno 24] Too many open files | closed | ### Describe the bug
I am trying to load the 'default' subset of the following dataset which contains lots of files (828 per split): [https://huggingface.co/datasets/mteb/biblenlp-corpus-mmteb](https://huggingface.co/datasets/mteb/biblenlp-corpus-mmteb)
When trying to load it using the `load_dataset` function I get... | 2024-05-07T01:15:09 | 2024-06-02T14:22:23 | 2024-05-13T13:01:55 | https://github.com/huggingface/datasets/issues/6877 | null | 6,877 | false | [
"ulimit -n 8192 can solve this problem",
"> ulimit -n 8192 can solve this problem\r\n\r\nWould there be a systematic way to do this ? The data loading is part of the [MTEB](https://github.com/embeddings-benchmark/mteb) library",
"> > ulimit -n 8192 can solve this problem\r\n> \r\n> Would there be a systematic w... |
2,281,450,743 | Unpin hfh | closed | Needed to use those in dataset-viewer:
- dev version of hfh https://github.com/huggingface/dataset-viewer/pull/2781: don't span the hub with /paths-info requests
- dev version of datasets at https://github.com/huggingface/datasets/pull/6875: don't write too big logs in the viewer
close https://github.com/hugging... | 2024-05-06T18:10:49 | 2024-05-27T10:20:42 | 2024-05-27T10:14:40 | https://github.com/huggingface/datasets/pull/6876 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6876",
"html_url": "https://github.com/huggingface/datasets/pull/6876",
"diff_url": "https://github.com/huggingface/datasets/pull/6876.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6876.patch",
"merged_at": "2024-05-27T10:14... | 6,876 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6876). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"transformers 4.40.2 was release yesterday but not sure if it contains the fix",
"@lho... |
2,281,428,826 | Shorten long logs | closed | Some datasets may have unexpectedly long features/types (e.g. if the files are not formatted correctly).
In that case we should still be able to log something readable | 2024-05-06T17:57:07 | 2024-05-07T12:31:46 | 2024-05-07T12:25:45 | https://github.com/huggingface/datasets/pull/6875 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6875",
"html_url": "https://github.com/huggingface/datasets/pull/6875",
"diff_url": "https://github.com/huggingface/datasets/pull/6875.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6875.patch",
"merged_at": "2024-05-07T12:25... | 6,875 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6875). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,280,717,233 | Use pandas ujson in JSON loader to improve performance | closed | Use pandas ujson in JSON loader to improve performance.
Note that `datasets` has `pandas` as required dependency. And `pandas` includes `ujson` in `pd.io.json.ujson_loads`.
Fix #6867.
CC: @natolambert | 2024-05-06T12:01:27 | 2024-05-17T16:28:29 | 2024-05-17T16:22:27 | https://github.com/huggingface/datasets/pull/6874 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6874",
"html_url": "https://github.com/huggingface/datasets/pull/6874",
"diff_url": "https://github.com/huggingface/datasets/pull/6874.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6874.patch",
"merged_at": "2024-05-17T16:22... | 6,874 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6874). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Before pandas-2.2.0, the function `ujson_loads` was named `loads`: https://github.com/p... |
2,280,463,182 | Set dev version | closed | null | 2024-05-06T09:43:18 | 2024-05-06T10:03:19 | 2024-05-06T09:57:12 | https://github.com/huggingface/datasets/pull/6873 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6873",
"html_url": "https://github.com/huggingface/datasets/pull/6873",
"diff_url": "https://github.com/huggingface/datasets/pull/6873.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6873.patch",
"merged_at": "2024-05-06T09:57... | 6,873 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6873). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,280,438,432 | Release 2.19.1 | closed | null | 2024-05-06T09:29:15 | 2024-05-06T09:35:33 | 2024-05-06T09:35:32 | https://github.com/huggingface/datasets/pull/6872 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6872",
"html_url": "https://github.com/huggingface/datasets/pull/6872",
"diff_url": "https://github.com/huggingface/datasets/pull/6872.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6872.patch",
"merged_at": "2024-05-06T09:35... | 6,872 | true | [] |
2,280,102,869 | Fix download for dict of dicts of URLs | closed | Fix download for a dict of dicts of URLs when batched (default), introduced by:
- #6794
This PR also implements regression tests.
Fix #6869, fix #6850. | 2024-05-06T06:06:52 | 2024-05-06T09:32:03 | 2024-05-06T09:25:52 | https://github.com/huggingface/datasets/pull/6871 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6871",
"html_url": "https://github.com/huggingface/datasets/pull/6871",
"diff_url": "https://github.com/huggingface/datasets/pull/6871.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6871.patch",
"merged_at": "2024-05-06T09:25... | 6,871 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6871). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Once merged, I think a patch release is needed.",
"Once the CI is green, I am merging... |
2,280,084,008 | Update tqdm >= 4.66.3 to fix vulnerability | closed | Update tqdm >= 4.66.3 to fix vulnerability, | 2024-05-06T05:49:36 | 2024-05-06T06:08:06 | 2024-05-06T06:02:00 | https://github.com/huggingface/datasets/pull/6870 | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6870",
"html_url": "https://github.com/huggingface/datasets/pull/6870",
"diff_url": "https://github.com/huggingface/datasets/pull/6870.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/6870.patch",
"merged_at": "2024-05-06T06:02... | 6,870 | true | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6870). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>... |
2,280,048,297 | Download is broken for dict of dicts: FileNotFoundError | closed | It seems there is a bug when downloading a dict of dicts of URLs introduced by:
- #6794
## Steps to reproduce the bug:
```python
from datasets import DownloadManager
dl_manager = DownloadManager()
paths = dl_manager.download({"train": {"frr": "hf://datasets/wikimedia/wikipedia/20231101.frr/train-00000-of-0000... | 2024-05-06T05:13:36 | 2024-05-06T09:25:53 | 2024-05-06T09:25:53 | https://github.com/huggingface/datasets/issues/6869 | null | 6,869 | false | [] |
2,279,385,159 | datasets.BuilderConfig does not work. | closed | ### Describe the bug
I custom a BuilderConfig and GeneratorBasedBuilder.
Here is the code for BuilderConfig
```
class UIEConfig(datasets.BuilderConfig):
def __init__(
self,
*args,
data_dir=None,
instruction_file=None,
instruction_strategy=None,... | 2024-05-05T08:08:55 | 2024-05-05T12:15:02 | 2024-05-05T12:15:01 | https://github.com/huggingface/datasets/issues/6868 | null | 6,868 | false | [
"I guess the issue is caused by the customization of BuilderConfig that you use from the repo [https://github.com/BeyonderXX/InstructUIE](https://github.com/BeyonderXX/InstructUIE/blob/master/src/uie_dataset.py). You should report to them.\r\n\r\nI see you already opened an issue in their repo:\r\n- https://github.... |
2,279,059,787 | Improve performance of JSON loader | closed | As reported by @natolambert, loading regular JSON files with `datasets` shows poor performance.
The cause is that we use the `json` Python standard library instead of other faster libraries. See my old comment: https://github.com/huggingface/datasets/pull/2638#pullrequestreview-706983714
> There are benchmarks that... | 2024-05-04T15:04:16 | 2024-05-17T16:22:28 | 2024-05-17T16:22:28 | https://github.com/huggingface/datasets/issues/6867 | null | 6,867 | false | [
"Thanks! Feel free to ping me for examples. May not respond immediately because we're all busy but would like to help.",
"Hi @natolambert, could you please give some examples of JSON files to benchmark?\r\n\r\nPlease note that this JSON file (https://huggingface.co/datasets/allenai/reward-bench-results/blob/main/... |
2,278,736,221 | DataFilesNotFoundError for datasets in the open-llm-leaderboard | closed | ### Describe the bug
When trying to get config names or load any dataset within the open-llm-leaderboard ecosystem (`open-llm-leaderboard/details_`) I receive the DataFilesNotFoundError. For the last month or so I've been loading datasets from the leaderboard almost everyday; yesterday was the first time I started see... | 2024-05-04T04:59:00 | 2024-05-14T08:09:56 | 2024-05-14T08:09:56 | https://github.com/huggingface/datasets/issues/6866 | null | 6,866 | false | [
"Potentially related:\r\n* #6864\r\n* #6850\r\n* #6848\r\n* #6819",
"Hi @jerome-white, thnaks for reporting.\r\n\r\nHowever, I cannot reproduce your issue:\r\n```python\r\n>>> from datasets import get_dataset_config_names\r\n\r\n>>> get_dataset_config_names(\"open-llm-leaderboard/details_davidkim205__Rhea-72b-v0.... |
2,277,304,832 | Example on Semantic segmentation contains bug | open | ### Describe the bug
https://huggingface.co/docs/datasets/en/semantic_segmentation shows wrong example with torchvision transforms.
Specifically, as one can see in screenshot below, the object boundaries have weird colors.
<img width="689" alt="image" src="https://github.com/huggingface/datasets/assets/4803565/59... | 2024-05-03T09:40:12 | 2024-05-03T09:40:12 | null | https://github.com/huggingface/datasets/issues/6865 | null | 6,865 | false | [] |
2,276,986,981 | Dataset 'rewardsignal/reddit_writing_prompts' doesn't exist on the Hub | closed | ### Describe the bug
The dataset `rewardsignal/reddit_writing_prompts` is missing in Huggingface Hub.
### Steps to reproduce the bug
```
from datasets import load_dataset
prompt_response_dataset = load_dataset("rewardsignal/reddit_writing_prompts", data_files="prompt_responses_full.csv", split='train[:80%]... | 2024-05-03T06:03:30 | 2024-05-06T06:36:42 | 2024-05-06T06:36:41 | https://github.com/huggingface/datasets/issues/6864 | null | 6,864 | false | [
"Hi @vinodrajendran001, thanks for reporting.\r\n\r\nIndeed the dataset no longer exists on the Hub. The URL https://huggingface.co/datasets/rewardsignal/reddit_writing_prompts gives 404 Not Found error."
] |
2,276,977,534 | Revert temporary pin huggingface-hub < 0.23.0 | closed | Revert temporary pin huggingface-hub < 0.23.0 introduced by
- #6861
once the following issue is fixed and released:
- huggingface/transformers#30618 | 2024-05-03T05:53:55 | 2024-05-27T10:14:41 | 2024-05-27T10:14:41 | https://github.com/huggingface/datasets/issues/6863 | null | 6,863 | false | [] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.