url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
3.67B
node_id
stringlengths
18
32
number
int64
1
7.88k
title
stringlengths
1
290
user
dict
labels
listlengths
0
4
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
listlengths
0
30
created_at
timestamp[ns, tz=UTC]date
2020-04-14 10:18:02
2025-11-26 16:16:56
updated_at
timestamp[ns, tz=UTC]date
2020-04-27 16:04:17
2025-11-27 11:08:44
closed_at
timestamp[ns, tz=UTC]date
2020-04-14 12:01:40
2025-11-21 12:31:19
author_association
stringclasses
4 values
type
float64
active_lock_reason
float64
sub_issues_summary
dict
issue_dependencies_summary
dict
body
stringlengths
0
228k
closed_by
dict
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
float64
state_reason
stringclasses
4 values
draft
float64
0
1
pull_request
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/7883
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7883/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7883/comments
https://api.github.com/repos/huggingface/datasets/issues/7883/events
https://github.com/huggingface/datasets/issues/7883
3,668,182,561
I_kwDODunzps7apAYh
7,883
Data.to_csv() cannot be recognized by pylance
{ "avatar_url": "https://avatars.githubusercontent.com/u/154290630?v=4", "events_url": "https://api.github.com/users/xi4ngxin/events{/privacy}", "followers_url": "https://api.github.com/users/xi4ngxin/followers", "following_url": "https://api.github.com/users/xi4ngxin/following{/other_user}", "gists_url": "https://api.github.com/users/xi4ngxin/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/xi4ngxin", "id": 154290630, "login": "xi4ngxin", "node_id": "U_kgDOCTJJxg", "organizations_url": "https://api.github.com/users/xi4ngxin/orgs", "received_events_url": "https://api.github.com/users/xi4ngxin/received_events", "repos_url": "https://api.github.com/users/xi4ngxin/repos", "site_admin": false, "starred_url": "https://api.github.com/users/xi4ngxin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xi4ngxin/subscriptions", "type": "User", "url": "https://api.github.com/users/xi4ngxin", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-26T16:16:56Z
2025-11-26T16:16:56Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug Hi, everyone ! I am a beginner with datasets. I am testing reading multiple CSV files from a zip archive. The result of reading the dataset shows success, and it can ultimately be correctly saved to CSV. Intermediate results: ``` Generating train split: 62973 examples [00:00, 175939.01 examples/s] DatasetDict({ train: Dataset({ features: ['交易时间\t', '收支方向\t', '业务(产品)种类\t', '交易金额\t', '币种\t', '时点余额\t', '对手方名称\t', '对方机构名称\t', ' 对方钱包ID/账号\t', '交易对手名称\t', '交易对手编号\t', '交易流水号\t', '摘要\t', '附言\t', '备注\t', '用途\t', '客户流水号\t'], num_rows: 62973 }) }) ``` However, Pylance gives me the following error: ``` Cannot access attribute "to_csv" for class "DatasetDict" Attribute "to_csv" is unknownPylance[reportAttributeAccessIssue](https://github.com/microsoft/pylance-release/blob/main/docs/diagnostics/reportAttributeAccessIssue.md)``` Cannot access attribute "to_csv" for class "IterableDatasetDict" Attribute "to_csv" is unknownPylance[reportAttributeAccessIssue](https://github.com/microsoft/pylance-release/blob/main/docs/diagnostics/reportAttributeAccessIssue.md) (method) to_csv: Unknown | ((path_or_buf: datasets.utils.typing.PathLike | BinaryIO, batch_size: int | None = None, num_proc: int | None = None, storage_options: dict[Unknown, Unknown] | None = None, **to_csv_kwargs: Unknown) -> int) | ((path_or_buf: datasets.utils.typing.PathLike | BinaryIO, batch_size: int | None = None, storage_options: dict[Unknown, Unknown] | None = None, **to_csv_kwargs: Unknown) -> int) ``` I ignored the error and continued executing to get the correct result: ``` Dataset({ features: ['交易时间\t', '收支方向\t', '业务(产品)种类\t', '交易金额\t', '币种\t', '时点余额\t', '对手方名称\t', '对方机构名称\t', '对方 钱包ID/账号\t', '交易对手名称\t', '交易对手编号\t', '交易流水号\t', '摘要\t', '附言\t', '备注\t', '用途\t', '客户流水号\t'], num_rows: 62973 }) ``` Since the data volume is small, I manually merged the CSV files, and the final result is consistent with what the program saved. looks like : <img width="1264" height="150" alt="Image" src="https://github.com/user-attachments/assets/743540d7-ad8c-4531-ae7e-de71a5243a32" /> ### Steps to reproduce the bug this is my code. ``` from datasets import load_dataset def main(): url = "data/test.zip" data_files = {"train": url} dataset = load_dataset("csv", data_files=data_files,split="train", encoding="gbk", skiprows=2) # print(dataset) dataset.to_csv("data/test.csv") if __name__ == "__main__": main() ``` ### Expected behavior I want to know why this happens. Is there something wrong with my code? ### Environment info OS: Windows 11 **upgrade from** OS: Windows_NT x64 10.0.22631 Editor: VS Code Version: 1.106.2 (user setup) "datasets" version = "4.4.1"
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7883/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7883/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7882
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7882/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7882/comments
https://api.github.com/repos/huggingface/datasets/issues/7882/events
https://github.com/huggingface/datasets/issues/7882
3,667,664,527
I_kwDODunzps7anB6P
7,882
Inconsistent loading of LFS-hosted files in epfml/FineWeb-HQ dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/6270922?v=4", "events_url": "https://api.github.com/users/Oligou/events{/privacy}", "followers_url": "https://api.github.com/users/Oligou/followers", "following_url": "https://api.github.com/users/Oligou/following{/other_user}", "gists_url": "https://api.github.com/users/Oligou/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Oligou", "id": 6270922, "login": "Oligou", "node_id": "MDQ6VXNlcjYyNzA5MjI=", "organizations_url": "https://api.github.com/users/Oligou/orgs", "received_events_url": "https://api.github.com/users/Oligou/received_events", "repos_url": "https://api.github.com/users/Oligou/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Oligou/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Oligou/subscriptions", "type": "User", "url": "https://api.github.com/users/Oligou", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-26T14:06:02Z
2025-11-26T14:06:02Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug Some files in the `epfml/FineWeb-HQ` dataset fail to load via the Hugging Face `datasets` library. - xet-hosted files load fine - LFS-hosted files sometimes fail Example: - Fails: https://huggingface.co/datasets/epfml/FineWeb-HQ/blob/main/data/CC-MAIN-2024-26/000_00003.parquet - Works: https://huggingface.co/datasets/epfml/FineWeb-HQ/blob/main/data/CC-MAIN-2024-42/000_00027.parquet Discussion: https://huggingface.co/datasets/epfml/FineWeb-HQ/discussions/2 ### Steps to reproduce the bug ```python from datasets import load_dataset ds = load_dataset( "epfml/FineWeb-HQ", data_files="data/CC-MAIN-2024-26/000_00003.parquet", ) ``` Error message: ``` HfHubHTTPError: 403 Forbidden: None. Cannot access content at: https://cdn-lfs-us-1.hf.co/repos/... Make sure your token has the correct permissions. ... <Error><Code>AccessDenied</Code><Message>Access Denied</Message></Error> ``` ### Expected behavior It should load the dataset for all files. ### Environment info - python 3.10 - datasets 4.4.1
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7882/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7882/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7881
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7881/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7881/comments
https://api.github.com/repos/huggingface/datasets/issues/7881/events
https://github.com/huggingface/datasets/pull/7881
3,667,642,524
PR_kwDODunzps61qI8F
7,881
Fix spurious label column when directories match split names
{ "avatar_url": "https://avatars.githubusercontent.com/u/132138786?v=4", "events_url": "https://api.github.com/users/neha222222/events{/privacy}", "followers_url": "https://api.github.com/users/neha222222/followers", "following_url": "https://api.github.com/users/neha222222/following{/other_user}", "gists_url": "https://api.github.com/users/neha222222/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/neha222222", "id": 132138786, "login": "neha222222", "node_id": "U_kgDOB-BHIg", "organizations_url": "https://api.github.com/users/neha222222/orgs", "received_events_url": "https://api.github.com/users/neha222222/received_events", "repos_url": "https://api.github.com/users/neha222222/repos", "site_admin": false, "starred_url": "https://api.github.com/users/neha222222/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neha222222/subscriptions", "type": "User", "url": "https://api.github.com/users/neha222222", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-26T13:59:46Z
2025-11-26T13:59:46Z
null
NONE
null
null
null
null
Issue - https://github.com/huggingface/datasets/issues/7880
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7881/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7881/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7881.diff", "html_url": "https://github.com/huggingface/datasets/pull/7881", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7881.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7881" }
true
https://api.github.com/repos/huggingface/datasets/issues/7880
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7880/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7880/comments
https://api.github.com/repos/huggingface/datasets/issues/7880/events
https://github.com/huggingface/datasets/issues/7880
3,667,561,864
I_kwDODunzps7amo2I
7,880
Spurious label column created when audiofolder/imagefolder directories match split names
{ "avatar_url": "https://avatars.githubusercontent.com/u/132138786?v=4", "events_url": "https://api.github.com/users/neha222222/events{/privacy}", "followers_url": "https://api.github.com/users/neha222222/followers", "following_url": "https://api.github.com/users/neha222222/following{/other_user}", "gists_url": "https://api.github.com/users/neha222222/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/neha222222", "id": 132138786, "login": "neha222222", "node_id": "U_kgDOB-BHIg", "organizations_url": "https://api.github.com/users/neha222222/orgs", "received_events_url": "https://api.github.com/users/neha222222/received_events", "repos_url": "https://api.github.com/users/neha222222/repos", "site_admin": false, "starred_url": "https://api.github.com/users/neha222222/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neha222222/subscriptions", "type": "User", "url": "https://api.github.com/users/neha222222", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-26T13:36:24Z
2025-11-26T13:36:24Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
## Describe the bug When using `audiofolder` or `imagefolder` with directories for **splits** (train/test) rather than class labels, a spurious `label` column is incorrectly created. **Example:** https://huggingface.co/datasets/datasets-examples/doc-audio-4 ``` from datasets import load_dataset ds = load_dataset("datasets-examples/doc-audio-4") print(ds["train"].features) ``` Shows 'label' column with ClassLabel(names=['test', 'train']) - incorrect!## Root cause In `folder_based_builder.py`, the `labels` set is accumulated across ALL splits (line 77). When directories are `train/` and `test/`: - `labels = {"train", "test"}` → `len(labels) > 1` → `add_labels = True` - Spurious label column is created with split names as class labels ## Expected behavior No `label` column should be added when directory names match split names. ## Proposed fix Skip label inference when inferred labels match split names. cc @lhoestq
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7880/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7880/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7879
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7879/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7879/comments
https://api.github.com/repos/huggingface/datasets/issues/7879/events
https://github.com/huggingface/datasets/issues/7879
3,657,249,446
I_kwDODunzps7Z_TKm
7,879
python core dump when downloading dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/5960219?v=4", "events_url": "https://api.github.com/users/hansewetz/events{/privacy}", "followers_url": "https://api.github.com/users/hansewetz/followers", "following_url": "https://api.github.com/users/hansewetz/following{/other_user}", "gists_url": "https://api.github.com/users/hansewetz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hansewetz", "id": 5960219, "login": "hansewetz", "node_id": "MDQ6VXNlcjU5NjAyMTk=", "organizations_url": "https://api.github.com/users/hansewetz/orgs", "received_events_url": "https://api.github.com/users/hansewetz/received_events", "repos_url": "https://api.github.com/users/hansewetz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hansewetz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hansewetz/subscriptions", "type": "User", "url": "https://api.github.com/users/hansewetz", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Hi @hansewetz I'm curious, for me it works just fine. Are you still observing the issue?", "Yup ... still the same issue.\nHowever, after adding a ```sleep(1)``` call after the ``` for``` loop by accident during debugging, the program terminates properly (not a good solution though ... :-) ).\nAre there some threads created that handles the download that are still running when the program exits?\nHaven't had time yet to go through the code in ```iterable_dataset.py::IterableDataset```\n", "Interesting, I was able to reproduce it, on a jupyter notebook the code runs just fine, as a Python script indeed it seems to never finish running (which is probably leading to the core dumped error). I'll try and take a look at the source code as well to see if I can figure it out.", "Hi @hansewetz ,\nIf possible can I be assigned with this issue?\n\n", "```If possible can I be assigned with this issue?```\nHi, I don't know how assignments work here and who can take decisions about assignments ... ", "Hi @hansewetz and @Aymuos22, I have made some progress:\n\n1) Confirmed last working version is 3.1.0\n\n2) From 3.1.0 to 3.2.0, there was a change in how parquet files are read (see [here](https://github.com/huggingface/datasets/blob/main/src/datasets/packaged_modules/parquet/parquet.py/#168).\n\nThe issue seems to be the following code:\n\n```\nparquet_fragment.to_batches(\n batch_size=batch_size,\n columns=self.config.columns,\n filter=filter_expr,\n batch_readahead=0,\n fragment_readahead=0,\n )\n```\n\nAdding a `use_threads=False` parameter to the `to_batches` call solves the bug. However, this seems far from an optimal solution, since we'd like to be able to use multiple threads for reading the fragments. \n\nI'll keep investigating to see if there's a better solution.", "Hi @lhoestq, may I ask if the current behaviour was expected by you folks and you don't think it needs solving, or should I keep on investigating a compromise between using multithreading / avoid unexpected behaviour? Thanks in advance :) ", "Having the same issue. the code never stops executing. Using datasets 4.4.1\nTried with \"islice\" as well. When the streaming flag is True, the code doesn't end execution. On vs-code.", "The issue on pyarrow side is here: https://github.com/apache/arrow/issues/45214 and the original issue in `datasets` here: https://github.com/huggingface/datasets/issues/7357\n\nIt would be cool to have a fix on the pyarrow side", "Thank you very much @lhoestq, I'm reading the issue thread in pyarrow and realizing you've been raising awareness around this for a long time now. When I have some time I'll look at @pitrou's PR to see if I can get a better understanding of what's going on on pyarrow. " ]
2025-11-24T06:22:53Z
2025-11-25T20:45:55Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug When downloading a dataset in streamed mode and exiting the program before the download completes, the python program core dumps when exiting: ``` terminate called without an active exception Aborted (core dumped) ``` Tested with python 3.12.3, python 3.9.21 ### Steps to reproduce the bug Create python venv: ```bash python -m venv venv ./venv/bin/activate pip install datasets==4.4.1 ``` Execute the following program: ``` from datasets import load_dataset ds = load_dataset("HuggingFaceFW/fineweb-2", 'hrv_Latn', split="test", streaming=True) for sample in ds: break ``` ### Expected behavior Clean program exit ### Environment info described above **note**: the example works correctly when using ```datasets==3.1.0```
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7879/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7879/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7878
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7878/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7878/comments
https://api.github.com/repos/huggingface/datasets/issues/7878/events
https://github.com/huggingface/datasets/pull/7878
3,653,262,027
PR_kwDODunzps606R81
7,878
Replace papaya with niivue
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "@CloseChoice thanks for your work on this. As you mentioned, the prime developers for Papaya have moved on, so it is in maintenance mode, albeit it is mature and may fill all your requirements. \r\n\r\nPapaya does reflect the era of its creation, so it uses WebGL1 (which only supports 2D textures) for display and pako for decompression. In contrast, NiiVue uses WebGL2 (where 3D textures provide a native representation for volumes) and compression streams (x4 decoding speed). A major benefit of 3D textures is simple support for 3D volume rendering using ray casting. Note the Papaya README shows an isosurface rendering based on a triangulated mesh. In contrast, NiiVue can show both volume rendering (good for data with fuzzy boundaries) as well as surface rendering (good when a clean isosurface can be defined). I think the [gallery](https://niivue.com/gallery) provides a nice example of NiiVue capabilities as well as minimal recipes.\r\n\r\nI do agree that Papaya UI is more advanced: by design NiiVue is a graphic widget that can be embedded into a container that provides your preferred user interface (React, Angular, Vue, pure html, or even jupyter notebooks). \r\n\r\nI think DICOM support is a challenge for any tool for several reasons: the diversity of the implementations and compression methods (transfer syntaxes), the fact that in classic DICOM each 2D slice is saved as a separate file (though note modern enhanced DICOM can save an entire 3D volume or even 4D timeseries in a single file), and the rate that this format has evolved over time. Papaya uses [Daikon](https://github.com/rii-mango/Daikon) to handle DICOM images, and I think it is only one file at a time. In contrast, NiiVue provides plugins for complex image formats, so you can choose your desired tool. We do provide illustrate how to use [dcm2niix WASM](https://github.com/niivue/niivue-dcm2niix) as a DICOM loader, and it can extract coherent volumes from a random assortment of files or a manifest of files - see the [live demo](https://github.com/niivue/niivue-dcm2niix). Note that diakon development has halted, while dcm2niix is actively maintained, which impacts support for emerging compression methods (e.g. JPEG2000-HT). Having said that, if your primary focus is DICOM, [cornerstonejs](https://www.cornerstonejs.org/) is probably a better choice than NiiVue or Papaya.\r\n\r\nAnother feature that may or may not be worth noting is that NiiVue has a plugin model that allows you to use a lot of mature image processing tools. So you can do image conversion, image processing (itk-wasm, niimath), image registration (flirt, elastix) and edge-based AI models. [brainchop](https://brainchop.org/) illustrates edge-based AI model inference for brain segmentation, extraction and parcellation, though we provide minimal examples for ONNX, tensorflowjs and tinygrad. This would provide a convenient way for huggingface inference models to be shared. After training, the models could be converted to ONNX and deployed on a web page, allowing the user to drag-and-drop images and process them regardless of operating system or graphics card manufacturer. Since the AI model inference leverages the users own graphics card, the privacy issues and hardware scaling concerns of cloud distribution are mitigated.\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n", "@neurolabusc thanks so much for the nuanced and informative reply.\r\nI am convinced that niivue is the better option here, having 3D support is huge and Papaya's UI features are actually not necessary at all, and AFAIS we can get what we need and more with some additional configuration for niivue as well.\r\nThanks a lot for words about DICOM, though the focus of this PR is not NifTI and not DICOM, I think having one tool being able to load both (and potentially more formats) is best, I'll definitely test the live demo. My primary interest in your thoughts about DICOM is to enable visualization as a follow-up to this PR #https://github.com/huggingface/datasets/pull/7835. Even for the DICOM case NiiVue seems like a great option using the [dcm2niix](https://github.com/niivue/niivue-dcm2niix) webassembly plugin, I think the main challenge is here how we let the user organize files in an intuitive way (e.g. provide DICOM folder class, and a DICOM document class where one folder can contain multiple documents and 3d visualization is on the folder level). \r\n\r\nGiven that NiiVue is a modern neuroimaging viewer, well maintained and widely used and we have @neurolabusc attention in case of questions/problems I think we should go ahead with NiiVue.\r\n\r\n@lhoestq your thoughts are highly appreciated.", "Following the @neurolabusc 's suggestion I updated to [ipyniivue](https://github.com/niivue/ipyniivue?tab=readme-ov-file) which helps so that we don't need to bother with javascript and speeds up load times since ipyniivue comes with a bundled niivue version and therefore avoids to download. Since DICOM is out of the picture for now, I consider this ready to be reviewed.", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7878). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-21T22:19:56Z
2025-11-27T11:08:44Z
null
CONTRIBUTOR
null
null
null
null
I was contacted by Chris Rorden whose group is developing NiiVue (see https://github.com/niivue/niivue), which leverages WebGL2 (in contrast to Papaya which is WebGL1 based). He also offered support in the implementation, which might come in handy in case of any questions later on (see DICOM implemenation). I completely overlooked NiiVue when searching for frameworks. Development speed or lack thereof was already mentioned as a potential risk with Papaya. NiiVue is well and actively maintained, simply compare these two contribution charts: NiiVue: <img width="920" height="378" alt="image" src="https://github.com/user-attachments/assets/37a0a256-60aa-4758-bb07-97e421c68ae1" /> Papaya: <img width="920" height="378" alt="image" src="https://github.com/user-attachments/assets/1e1cf0c9-ec0a-4ffc-ae03-a79ea12bcb3b" /> I gave NiiVue a try and it supports all features Papaya does, though I find Papaya's UI slightly more appealing but that is just personal taste. There is also a 3D image of the scanned object included in the NiiVue UI, but that is possible for Papaya aswell (at least in some way, check the image in their github repo README.md). ```python from datasets import load_dataset # new dataset compared to papaya PR, this has more interesting images ds = load_dataset("TobiasPitters/nifti-papaya-testdata", split="train") ds[1]['nifti'] # ds[2]['nifti'] is also interesting ``` Here's a brief video how this looks with NiiVue: https://github.com/user-attachments/assets/3f2a52d4-2109-45e2-aca8-e4a4b1e46b32 NOTE: I explicitly created this as draft PR since I suspect the DICOM support to be a crucial factor to decide which of these two is better suited for our needs. DICOM is supported by Papaya, and for NiiVue as well using a plugin, but as far as I understand one DICOM file contains one 2D image, therefore support for loading a whole folder, containing all 2D layers for a complete 3D image is desired. NiiVue supports this according to their docs, I am unsure about Papaya.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7878/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7878/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7878.diff", "html_url": "https://github.com/huggingface/datasets/pull/7878", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7878.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7878" }
true
https://api.github.com/repos/huggingface/datasets/issues/7877
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7877/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7877/comments
https://api.github.com/repos/huggingface/datasets/issues/7877/events
https://github.com/huggingface/datasets/issues/7877
3,652,906,788
I_kwDODunzps7Zuu8k
7,877
work around `tempfile` silently ignoring `TMPDIR` if the dir doesn't exist
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stas00", "id": 10676103, "login": "stas00", "node_id": "MDQ6VXNlcjEwNjc2MTAz", "organizations_url": "https://api.github.com/users/stas00/orgs", "received_events_url": "https://api.github.com/users/stas00/received_events", "repos_url": "https://api.github.com/users/stas00/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "type": "User", "url": "https://api.github.com/users/stas00", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-21T19:51:48Z
2025-11-21T19:51:48Z
null
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
This should help a lot of users running into `No space left on device` while using `datasets`. Normally the issue is is that `/tmp` is too small and the user needs to use another path, which they would normally set as `export TMPDIR=/some/big/storage` However, the `tempfile` facility that `datasets` and `pyarrow` use is somewhat broken. If the path doesn't exist it'd ignore it and fall back to using `/tmp`. Watch this: ``` $ export TMPDIR='/tmp/username' $ python -c "\ import os import tempfile print(os.environ['TMPDIR']) print(tempfile.gettempdir())" /tmp/username /tmp ``` Now let's ensure the path exists: ``` $ export TMPDIR='/tmp/username' $ mkdir -p $TMPDIR $ python -c "\ import os import tempfile print(os.environ['TMPDIR']) print(tempfile.gettempdir())" /tmp/username /tmp/username ``` So I recommend `datasets` do either of the 2: 1. assert if `$TMPDIR` dir doesn't exist, telling the user to create it 2. auto-create it The reason for (1) is that I don't know why `tempdir` doesn't auto-create the dir - perhaps some security implication? I will let you guys make the decision, but the key is not to let things silently fall through and the user puzzling why no matter what they do they can't break past `No space left on device` while using `datasets` Thank you. I found this via https://stackoverflow.com/questions/37229398/python-tempfile-gettempdir-does-not-respect-tmpdir while trying to help a colleague to solve this exact issue.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7877/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7877/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7876
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7876/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7876/comments
https://api.github.com/repos/huggingface/datasets/issues/7876/events
https://github.com/huggingface/datasets/pull/7876
3,652,170,832
PR_kwDODunzps602lac
7,876
test: add verification for HuggingFaceM4/InterleavedWebDocuments
{ "avatar_url": "https://avatars.githubusercontent.com/u/122142345?v=4", "events_url": "https://api.github.com/users/venkatsai2004/events{/privacy}", "followers_url": "https://api.github.com/users/venkatsai2004/followers", "following_url": "https://api.github.com/users/venkatsai2004/following{/other_user}", "gists_url": "https://api.github.com/users/venkatsai2004/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/venkatsai2004", "id": 122142345, "login": "venkatsai2004", "node_id": "U_kgDOB0e-iQ", "organizations_url": "https://api.github.com/users/venkatsai2004/orgs", "received_events_url": "https://api.github.com/users/venkatsai2004/received_events", "repos_url": "https://api.github.com/users/venkatsai2004/repos", "site_admin": false, "starred_url": "https://api.github.com/users/venkatsai2004/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/venkatsai2004/subscriptions", "type": "User", "url": "https://api.github.com/users/venkatsai2004", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-21T15:42:09Z
2025-11-21T15:42:09Z
null
NONE
null
null
null
null
Adds an integration test for the `HuggingFaceM4/InterleavedWebDocuments` dataset. - Gracefully skips if the dataset is not yet available on the Hub - Checks basic loading and structure once it becomes available Closes #7394 First-time contributor to `datasets` — really excited about this! Happy to make any adjustments needed. 🙂
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7876/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7876/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7876.diff", "html_url": "https://github.com/huggingface/datasets/pull/7876", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7876.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7876" }
true
https://api.github.com/repos/huggingface/datasets/issues/7875
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7875/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7875/comments
https://api.github.com/repos/huggingface/datasets/issues/7875/events
https://github.com/huggingface/datasets/pull/7875
3,649,326,175
PR_kwDODunzps60s9my
7,875
Add quickstart example to datasets README
{ "avatar_url": "https://avatars.githubusercontent.com/u/101023542?v=4", "events_url": "https://api.github.com/users/hajermabrouk/events{/privacy}", "followers_url": "https://api.github.com/users/hajermabrouk/followers", "following_url": "https://api.github.com/users/hajermabrouk/following{/other_user}", "gists_url": "https://api.github.com/users/hajermabrouk/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hajermabrouk", "id": 101023542, "login": "hajermabrouk", "node_id": "U_kgDOBgV_Ng", "organizations_url": "https://api.github.com/users/hajermabrouk/orgs", "received_events_url": "https://api.github.com/users/hajermabrouk/received_events", "repos_url": "https://api.github.com/users/hajermabrouk/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hajermabrouk/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hajermabrouk/subscriptions", "type": "User", "url": "https://api.github.com/users/hajermabrouk", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-20T22:13:52Z
2025-11-20T22:13:52Z
null
NONE
null
null
null
null
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7875/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7875/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7875.diff", "html_url": "https://github.com/huggingface/datasets/pull/7875", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7875.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7875" }
true
https://api.github.com/repos/huggingface/datasets/issues/7874
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7874/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7874/comments
https://api.github.com/repos/huggingface/datasets/issues/7874/events
https://github.com/huggingface/datasets/pull/7874
3,644,558,046
PR_kwDODunzps60c4sg
7,874
Nifti visualization support
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7874). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "I tested in Colab and it works perfectly :) now I want to add `_repr_html_` everywhere xD\r\n\r\nRe: testing, I think it's fine to test manually such features" ]
2025-11-19T21:56:56Z
2025-11-21T12:41:43Z
2025-11-21T12:31:18Z
CONTRIBUTOR
null
null
null
null
closes #7870 leverage Papaya to visualize nifti images. For this I created a Wrapper class for `nibabel.nifti1.Nifti1Image` that provides the same interface but exposes an additional `_repr_html_` method, which is needed to visualize the image in jupyter (didn't test in colab, but that should work equivalently). Code to test (execute in a notebook): ```python from datasets import load_dataset ds = load_dataset("TobiasPitters/nifti-nitest-extracted", split="train") image = ds[1] image ``` Here a small video, not the most exciting scan though: https://github.com/user-attachments/assets/1cca5f01-6fd2-48ef-a4d7-a92c1259c224 Am open to good ways to test this. EDIT: papaya also supports dicom, didn't test it yet though
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7874/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7874/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7874.diff", "html_url": "https://github.com/huggingface/datasets/pull/7874", "merged_at": "2025-11-21T12:31:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/7874.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7874" }
true
https://api.github.com/repos/huggingface/datasets/issues/7873
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7873/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7873/comments
https://api.github.com/repos/huggingface/datasets/issues/7873/events
https://github.com/huggingface/datasets/pull/7873
3,643,993,705
PR_kwDODunzps60a_IZ
7,873
Fix chunk casting and schema unification in dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "\r\n@lhoestq would like to hear from you!\r\n" ]
2025-11-19T18:43:47Z
2025-11-22T19:51:30Z
null
CONTRIBUTOR
null
null
null
null
Updated chunk handling to cast to expected schema when features are provided or to unify schemas when not. This ensures proper schema alignment for the yielded batches. fixes #7872 This PR fixes a bug where `IterableDataset` created from a generator with explicit `features` parameter would fail during arrow operations (like `.to_pandas()`) when the data contains missing or null values. ## Problem When an `IterableDataset` is created with explicit features but the generator yields data with missing values (e.g., empty lists), PyArrow would infer different schemas for different batches based on the actual data rather than using the provided schema. This caused `ArrowInvalid` errors when trying to concatenate batches with mismatched schemas. ### Example error: ```python pyarrow.lib.ArrowInvalid: Schema at index 1 was different: a: int64 b: list vs a: int64 b: list> ``` ## Solution Modified `RebatchedArrowExamplesIterable._iter_arrow()` to: 1. Cast chunks to the expected schema when explicit features are provided 2. Unify schemas across chunks when no explicit features are set 3. Gracefully handle cast failures by falling back to the original chunk This ensures that the user-provided schema is respected throughout the iteration process. ## Testing Verified the fix with the following test case: ```python import datasets from datasets import features def test_to_pandas_works_with_explicit_schema(): common_features = features.Features( { "a": features.Value("int64"), "b": features.List({"c": features.Value("int64")}), } ) def row_generator(): data = [{"a": 1, "b": []}, {"a": 1, "b": [{"c": 1}]}] for row in data: yield row d = datasets.IterableDataset.from_generator(row_generator, features=common_features) print("Iterating…") for _ in d.to_pandas(): pass test_to_pandas_works_with_explicit_schema() ``` Before Patch - ``` @ArjunJagdale ➜ /workspaces/datasets (main) $ python test_arjun.py Iterating… Traceback (most recent call last): File "/workspaces/datasets/test_arjun.py", line 24, in <module> test_to_pandas_works_with_explicit_schema() File "/workspaces/datasets/test_arjun.py", line 21, in test_to_pandas_works_with_explicit_schema for _ in d.to_pandas(): File "/workspaces/datasets/src/datasets/iterable_dataset.py", line 3736, in to_pandas table = pa.concat_tables(list(self.with_format("arrow").iter(batch_size=1000))) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspaces/datasets/src/datasets/iterable_dataset.py", line 2596, in iter for key, pa_table in iterator: File "/workspaces/datasets/src/datasets/iterable_dataset.py", line 2111, in _iter_arrow for key, pa_table in self.ex_iterable._iter_arrow(): File "/workspaces/datasets/src/datasets/iterable_dataset.py", line 632, in _iter_arrow yield new_key, pa.Table.from_batches(chunks_buffer) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "pyarrow/table.pxi", line 5039, in pyarrow.lib.Table.from_batches File "pyarrow/error.pxi", line 155, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: Schema at index 1 was different: a: int64 b: list<item: null> vs a: int64 b: list<item: struct<c: int64>> ``` After Patch - ``` @ArjunJagdale ➜ /workspaces/datasets (main) $ python test_arjun.py Iterating… @ArjunJagdale ➜ /workspaces/datasets (main) $ ```
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7873/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7873/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7873.diff", "html_url": "https://github.com/huggingface/datasets/pull/7873", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7873.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7873" }
true
https://api.github.com/repos/huggingface/datasets/issues/7872
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7872/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7872/comments
https://api.github.com/repos/huggingface/datasets/issues/7872/events
https://github.com/huggingface/datasets/issues/7872
3,643,681,893
I_kwDODunzps7ZLixl
7,872
IterableDataset does not use features information in to_pandas
{ "avatar_url": "https://avatars.githubusercontent.com/u/790640?v=4", "events_url": "https://api.github.com/users/bonext/events{/privacy}", "followers_url": "https://api.github.com/users/bonext/followers", "following_url": "https://api.github.com/users/bonext/following{/other_user}", "gists_url": "https://api.github.com/users/bonext/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bonext", "id": 790640, "login": "bonext", "node_id": "MDQ6VXNlcjc5MDY0MA==", "organizations_url": "https://api.github.com/users/bonext/orgs", "received_events_url": "https://api.github.com/users/bonext/received_events", "repos_url": "https://api.github.com/users/bonext/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bonext/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bonext/subscriptions", "type": "User", "url": "https://api.github.com/users/bonext", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Created A PR!", "Another test script that can be used to test the behavior - \n\n```\nimport datasets\nfrom datasets import features\n\ndef test_crash():\n common_features = features.Features({\n \"a\": features.Value(\"int64\"),\n \"b\": features.List({\"c\": features.Value(\"int64\")}),\n })\n\n def row_generator():\n yield {\"a\": 1, \"b\": []}\n yield {\"a\": 1, \"b\": [{\"c\": 1}]}\n\n d = datasets.IterableDataset.from_generator(row_generator, features=common_features)\n\n list(d.to_pandas()) # <-- this triggers the crash\n\n```" ]
2025-11-19T17:12:59Z
2025-11-19T18:52:14Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug `IterableDataset` created from generator with explicit `features=` parameter seems to ignore provided features description for certain operations, e.g. `.to_pandas(...)` when data coming from the generator has missing values. ### Steps to reproduce the bug ```python import datasets from datasets import features def test_to_pandas_works_with_explicit_schema(): common_features = features.Features( { "a": features.Value("int64"), "b": features.List({"c": features.Value("int64")}), } ) def row_generator(): data = [{"a": 1, "b": []}, {"a": 1, "b": [{"c": 1}]}] for row in data: yield row d = datasets.IterableDataset.from_generator(row_generator, features=common_features) for _ in d.to_pandas(): pass # _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ # .venv/lib/python3.13/site-packages/datasets/iterable_dataset.py:3703: in to_pandas # table = pa.concat_tables(list(self.with_format("arrow").iter(batch_size=1000))) # ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ # .venv/lib/python3.13/site-packages/datasets/iterable_dataset.py:2563: in iter # for key, pa_table in iterator: # ^^^^^^^^ # .venv/lib/python3.13/site-packages/datasets/iterable_dataset.py:2078: in _iter_arrow # for key, pa_table in self.ex_iterable._iter_arrow(): # ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ # .venv/lib/python3.13/site-packages/datasets/iterable_dataset.py:599: in _iter_arrow # yield new_key, pa.Table.from_batches(chunks_buffer) # ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ # pyarrow/table.pxi:5039: in pyarrow.lib.Table.from_batches # ??? # pyarrow/error.pxi:155: in pyarrow.lib.pyarrow_internal_check_status # ??? # _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ # > ??? # E pyarrow.lib.ArrowInvalid: Schema at index 1 was different: # E a: int64 # E b: list<item: null> # E vs # E a: int64 # E b: list<item: struct<c: int64>> # pyarrow/error.pxi:92: ArrowInvalid ``` ### Expected behavior arrow operations use schema provided through `features=` and not the one inferred from the data ### Environment info - datasets version: 4.4.1 - Platform: macOS-15.7.1-arm64-arm-64bit-Mach-O - Python version: 3.13.1 - huggingface_hub version: 1.1.4 - PyArrow version: 22.0.0 - Pandas version: 2.3.3 - fsspec version: 2025.10.0
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7872/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7872/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7871
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7871/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7871/comments
https://api.github.com/repos/huggingface/datasets/issues/7871/events
https://github.com/huggingface/datasets/issues/7871
3,643,607,371
I_kwDODunzps7ZLQlL
7,871
Reqwest Error: HTTP status client error (429 Too Many Requests)
{ "avatar_url": "https://avatars.githubusercontent.com/u/26405281?v=4", "events_url": "https://api.github.com/users/yanan1116/events{/privacy}", "followers_url": "https://api.github.com/users/yanan1116/followers", "following_url": "https://api.github.com/users/yanan1116/following{/other_user}", "gists_url": "https://api.github.com/users/yanan1116/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yanan1116", "id": 26405281, "login": "yanan1116", "node_id": "MDQ6VXNlcjI2NDA1Mjgx", "organizations_url": "https://api.github.com/users/yanan1116/orgs", "received_events_url": "https://api.github.com/users/yanan1116/received_events", "repos_url": "https://api.github.com/users/yanan1116/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yanan1116/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yanan1116/subscriptions", "type": "User", "url": "https://api.github.com/users/yanan1116", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "the dataset repo: `https://huggingface.co/datasets/nvidia/PhysicalAI-Robotics-GR00T-X-Embodiment-Sim`" ]
2025-11-19T16:52:24Z
2025-11-19T16:53:07Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug full error message: ``` Traceback (most recent call last): File "/home/yanan/miniconda3/bin/hf", line 7, in <module> sys.exit(main()) ~~~~^^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/cli/hf.py", line 56, in main app() ~~~^^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/typer/main.py", line 327, in __call__ raise e File "/home/yanan/miniconda3/lib/python3.13/site-packages/typer/main.py", line 310, in __call__ return get_command(self)(*args, **kwargs) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/click/core.py", line 1161, in __call__ return self.main(*args, **kwargs) ~~~~~~~~~^^^^^^^^^^^^^^^^^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/typer/core.py", line 803, in main return _main( self, ...<6 lines>... **extra, ) File "/home/yanan/miniconda3/lib/python3.13/site-packages/typer/core.py", line 192, in _main rv = self.invoke(ctx) File "/home/yanan/miniconda3/lib/python3.13/site-packages/click/core.py", line 1697, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) ~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/click/core.py", line 1443, in invoke return ctx.invoke(self.callback, **ctx.params) ~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/click/core.py", line 788, in invoke return __callback(*args, **kwargs) File "/home/yanan/miniconda3/lib/python3.13/site-packages/typer/main.py", line 691, in wrapper return callback(**use_params) File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/cli/download.py", line 188, in download _print_result(run_download()) ~~~~~~~~~~~~^^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/cli/download.py", line 149, in run_download return snapshot_download( repo_id=repo_id, ...<10 lines>... dry_run=dry_run, ) File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn return fn(*args, **kwargs) File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/_snapshot_download.py", line 451, in snapshot_download thread_map( ~~~~~~~~~~^ _inner_hf_hub_download, ^^^^^^^^^^^^^^^^^^^^^^^ ...<3 lines>... tqdm_class=tqdm_class, ^^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/tqdm/contrib/concurrent.py", line 69, in thread_map return _executor_map(ThreadPoolExecutor, fn, *iterables, **tqdm_kwargs) File "/home/yanan/miniconda3/lib/python3.13/site-packages/tqdm/contrib/concurrent.py", line 51, in _executor_map return list(tqdm_class(ex.map(fn, *iterables, chunksize=chunksize), **kwargs)) File "/home/yanan/miniconda3/lib/python3.13/site-packages/tqdm/std.py", line 1181, in __iter__ for obj in iterable: ^^^^^^^^ File "/home/yanan/miniconda3/lib/python3.13/concurrent/futures/_base.py", line 619, in result_iterator yield _result_or_cancel(fs.pop()) ~~~~~~~~~~~~~~~~~^^^^^^^^^^ File "/home/yanan/miniconda3/lib/python3.13/concurrent/futures/_base.py", line 317, in _result_or_cancel return fut.result(timeout) ~~~~~~~~~~^^^^^^^^^ File "/home/yanan/miniconda3/lib/python3.13/concurrent/futures/_base.py", line 449, in result return self.__get_result() ~~~~~~~~~~~~~~~~~^^ File "/home/yanan/miniconda3/lib/python3.13/concurrent/futures/_base.py", line 401, in __get_result raise self._exception File "/home/yanan/miniconda3/lib/python3.13/concurrent/futures/thread.py", line 59, in run result = self.fn(*self.args, **self.kwargs) File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/_snapshot_download.py", line 431, in _inner_hf_hub_download hf_hub_download( # type: ignore ~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^ repo_id, ^^^^^^^^ ...<14 lines>... dry_run=dry_run, ^^^^^^^^^^^^^^^^ ) ^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn return fn(*args, **kwargs) File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 986, in hf_hub_download return _hf_hub_download_to_local_dir( # Destination ...<16 lines>... dry_run=dry_run, ) File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1390, in _hf_hub_download_to_local_dir _download_to_tmp_and_move( ~~~~~~~~~~~~~~~~~~~~~~~~~^ incomplete_path=paths.incomplete_path(etag), ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ...<8 lines>... tqdm_class=tqdm_class, ^^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1791, in _download_to_tmp_and_move xet_get( ~~~~~~~^ incomplete_path=incomplete_path, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ...<4 lines>... tqdm_class=tqdm_class, ^^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/home/yanan/miniconda3/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 571, in xet_get download_files( ~~~~~~~~~~~~~~^ xet_download_info, ^^^^^^^^^^^^^^^^^^ ...<3 lines>... progress_updater=[progress_updater], ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ) ^ RuntimeError: Data processing error: CAS service error : Reqwest Error: HTTP status client error (429 Too Many Requests), domain: https://cas-server.xethub.hf.co/reconstructions/04b8a4667b84b3b874a6a2f070cec88920f6289e71185d69fa87e3cf29834710 ``` ### Steps to reproduce the bug my command ```bash hf download nvidia/PhysicalAI-Robotics-GR00T-X-Embodiment-Sim --repo-type dataset --include "single_panda_gripper.CoffeePressButton/**" --local-dir /home/yanan/robotics/Isaac-GR00T/gr00t_dataset_official/ ``` ### Expected behavior expect the data can be downloaded without any issue ### Environment info huggingface_hub 1.1.4
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7871/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7871/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7870
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7870/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7870/comments
https://api.github.com/repos/huggingface/datasets/issues/7870/events
https://github.com/huggingface/datasets/issues/7870
3,642,209,953
I_kwDODunzps7ZF7ah
7,870
Visualization for Medical Imaging Datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "It would be amazing to be able to show the Papaya UI in google colab / jupyter notebook. IIRC both allow serving javascript via nbextensions that we can surely use in HTML() objects.\n\nAlternatively we could also start with a simple approach and dump the medical image data as a video file that goes through the slices, so we don't need javascript." ]
2025-11-19T11:05:39Z
2025-11-21T12:31:19Z
2025-11-21T12:31:19Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
This is a followup to: https://github.com/huggingface/datasets/pull/7815. I checked the possibilities to visualize the nifti (and potentially dicom), and here's what I found: - https://github.com/aces/brainbrowser, AGPL3 license, last commit 3 months ago, latest (github) release from 2017. It's available on jsdelivr: https://www.jsdelivr.com/package/npm/brainbrowser (but that is from 2015!) - https://github.com/rii-mango/Papaya, custom but BSD-style license that would require datasets to list the conditions in their readme somewhere, last commit June 2024. I looked into this library and it looks mature and good enough for our use case, but just working on it for a short time I wasn't able to get this to work, but am sure we could get this working, would probably require some JS on datasets' end. Available on jsdelivr as well: https://www.jsdelivr.com/package/npm/papaya-viewer. Seems like it's frequently loaded. - https://github.com/hanayik/niivue, BSD3 license, last commit May 26, 2021. Archived. Doesn't look like an option. I think the only real option for us Papaya, but there is also the risk that we'll end up with an unmaintained package after a while, since development seems to be slow or even halted. I think conceptually we would need to figure out how we can build a good solution for visualizing Medical Image data. On shap, we have a separate javascript folder in which we render visualizations, this could be a blueprint but will require a bundler, etc. Alternatively one could go with a naive approach to just write some html code in a python string and load the package via jsdelivr. @lhoestq thoughts?
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7870/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7870/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7869
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7869/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7869/comments
https://api.github.com/repos/huggingface/datasets/issues/7869/events
https://github.com/huggingface/datasets/issues/7869
3,636,808,734
I_kwDODunzps7YxUwe
7,869
Why does dataset merge fail when tools have different parameters?
{ "avatar_url": "https://avatars.githubusercontent.com/u/116297296?v=4", "events_url": "https://api.github.com/users/hitszxs/events{/privacy}", "followers_url": "https://api.github.com/users/hitszxs/followers", "following_url": "https://api.github.com/users/hitszxs/following{/other_user}", "gists_url": "https://api.github.com/users/hitszxs/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hitszxs", "id": 116297296, "login": "hitszxs", "node_id": "U_kgDOBu6OUA", "organizations_url": "https://api.github.com/users/hitszxs/orgs", "received_events_url": "https://api.github.com/users/hitszxs/received_events", "repos_url": "https://api.github.com/users/hitszxs/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hitszxs/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hitszxs/subscriptions", "type": "User", "url": "https://api.github.com/users/hitszxs", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-18T08:33:04Z
2025-11-18T08:33:04Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Hi, I have a question about SFT (Supervised Fine-tuning) for an agent model. Suppose I want to fine-tune an agent model that may receive two different tools: tool1 and tool2. These tools have different parameters and types in their schema definitions. When I try to merge datasets containing different tool definitions, I get the following error: TypeError: Couldn't cast array of type struct<refundFee: struct<description: string, type: string>, ... , servicerId: struct<description: string, type: string>> to { 'refundFee': {'description': Value(dtype='string'), 'type': Value(dtype='string')}, ... 'templateId': {'description': Value(dtype='string'), 'type': Value(dtype='string')} } From my understanding, the merge fails because the tools column's nested structure is different across datasets — e.g., one struct contains an extra field servicerId while the other does not. This causes HuggingFace Datasets (and its underlying Apache Arrow schema) to reject the merge. My question is: why is it designed this way? Is this strict schema matching a hard requirement of the library? Is there a recommended way to merge datasets with different tool schemas (different parameters and types)? For an agent model supporting multiple tools, what's the best practice for preparing/merging training data without losing flexibility? Any guidance or design rationale would be greatly appreciated. Thanks!
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7869/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7869/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7868
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7868/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7868/comments
https://api.github.com/repos/huggingface/datasets/issues/7868/events
https://github.com/huggingface/datasets/issues/7868
3,632,429,308
I_kwDODunzps7Ygnj8
7,868
Data duplication with `split_dataset_by_node` and `interleaved_dataset`
{ "avatar_url": "https://avatars.githubusercontent.com/u/42485228?v=4", "events_url": "https://api.github.com/users/ValMystletainn/events{/privacy}", "followers_url": "https://api.github.com/users/ValMystletainn/followers", "following_url": "https://api.github.com/users/ValMystletainn/following{/other_user}", "gists_url": "https://api.github.com/users/ValMystletainn/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ValMystletainn", "id": 42485228, "login": "ValMystletainn", "node_id": "MDQ6VXNlcjQyNDg1MjI4", "organizations_url": "https://api.github.com/users/ValMystletainn/orgs", "received_events_url": "https://api.github.com/users/ValMystletainn/received_events", "repos_url": "https://api.github.com/users/ValMystletainn/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ValMystletainn/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ValMystletainn/subscriptions", "type": "User", "url": "https://api.github.com/users/ValMystletainn", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Hi @ValMystletainn ,\nCan I be assigned this issue?" ]
2025-11-17T09:15:24Z
2025-11-25T04:27:05Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug Data duplication in different rank, when process a iterabledataset with first `split_dataset_by_node` and then `interleaved_dataset` ### Steps to reproduce the bug I have provide a minimum scripts ```python import os from datasets import interleave_datasets, load_dataset from datasets.distributed import split_dataset_by_node path = "/mnt/wwx/datasets/fineweb/data/CC-MAIN-2013-20/" files = [os.path.join(path, fn) for fn in os.listdir(path)] dataset = load_dataset("parquet", split="train", data_files=files, streaming=True) print(f"{dataset.n_shards=}") dataset_rank0 = split_dataset_by_node(dataset, 0, 4) dataset_rank1 = split_dataset_by_node(dataset, 1, 4) dataset_rank0_interleaved = interleave_datasets([dataset_rank0], seed=42, probabilities=[1.0]) dataset_rank1_interleaved = interleave_datasets([dataset_rank1], seed=42, probabilities=[1.0]) print("print the first sample id from all datasets") print("dataset", next(iter(dataset))['id']) print("dataset_rank0", next(iter(dataset_rank0))['id']) print("dataset_rank1", next(iter(dataset_rank1))['id']) print("dataset_rank0_interleaved", next(iter(dataset_rank0_interleaved))['id']) print("dataset_rank1_interleaved", next(iter(dataset_rank1_interleaved))['id']) dataset_rank0_shard = dataset.shard(4, 0) dataset_rank1_shard = dataset.shard(4, 1) dataset_rank0_shard_interleaved = interleave_datasets([dataset_rank0_shard], seed=42, probabilities=[1.0]) dataset_rank1_shard_interleaved = interleave_datasets([dataset_rank1_shard], seed=42, probabilities=[1.0]) print("dataset_rank0_shard", next(iter(dataset_rank0_shard))['id']) print("dataset_rank1_shard", next(iter(dataset_rank1_shard))['id']) print("dataset_rank0_shard_interleaved", next(iter(dataset_rank0_shard_interleaved))['id']) print("dataset_rank1_shard_interleaved", next(iter(dataset_rank1_shard_interleaved))['id']) ``` I just use a subfold of C4 with 14 paruets to do the quick run and get ``` dataset.n_shards=14 print the first sample id from all datasets dataset <urn:uuid:c84a7f00-f3e8-4b67-baa4-df5adaf23bae> dataset_rank0 <urn:uuid:c84a7f00-f3e8-4b67-baa4-df5adaf23bae> dataset_rank1 <urn:uuid:6b7da64f-c26e-4086-aef5-4b6f01106223> dataset_rank0_interleaved <urn:uuid:c84a7f00-f3e8-4b67-baa4-df5adaf23bae> dataset_rank1_interleaved <urn:uuid:c84a7f00-f3e8-4b67-baa4-df5adaf23bae> dataset_rank0_shard <urn:uuid:c84a7f00-f3e8-4b67-baa4-df5adaf23bae> dataset_rank1_shard <urn:uuid:67cf7216-dd05-4f55-a28a-1a1c96989c51> dataset_rank0_shard_interleaved <urn:uuid:c84a7f00-f3e8-4b67-baa4-df5adaf23bae> dataset_rank1_shard_interleaved <urn:uuid:67cf7216-dd05-4f55-a28a-1a1c96989c51> ``` ### Expected behavior the first sample of `dataset_rank0_interleaved` and `dataset_rank1_interleaved` should be different, as other `rank0` `rank1` couples. I have dive into the function and try to find how it work in `split -> interleaved` process. the `split_dataset_by_node` of iterable dataset does't not change `._ex_iterable` attribute of the dataset. it just set the distributed config in dataset, and the distributed dataset is used in actually `__iter__` call, to handle with shard split or sample skipping. however, in `interleaved_dataset` of iterable dataset. it copy out all of the `._ex_iterable` of provided datasets, and consist a new `_ex_iterable`, so the missing copy of `distributed config` caused the data duplication in different dp rank. So I may first ask, is it an unexpected using order of those function, which means: - always do `split_dataset_by_node` at final rather than in middle way. - or use `dataset.shard(dp_size, dp_rank)` rather than `split_dataset_by_node` in case similar of mine. if the using order is permiited, I think it is a bug, and I can do a PR to fix it (I meet this bug in real training, related issue is https://github.com/ByteDance-Seed/VeOmni/issues/200 if it helps. ### Environment info datasets 4.4.1 ubuntu 20.04 python 3.11.4
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7868/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7868/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7867
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7867/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7867/comments
https://api.github.com/repos/huggingface/datasets/issues/7867/events
https://github.com/huggingface/datasets/issues/7867
3,620,931,722
I_kwDODunzps7X0wiK
7,867
NonMatchingSplitsSizesError when loading partial dataset files
{ "avatar_url": "https://avatars.githubusercontent.com/u/13678719?v=4", "events_url": "https://api.github.com/users/QingGo/events{/privacy}", "followers_url": "https://api.github.com/users/QingGo/followers", "following_url": "https://api.github.com/users/QingGo/following{/other_user}", "gists_url": "https://api.github.com/users/QingGo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/QingGo", "id": 13678719, "login": "QingGo", "node_id": "MDQ6VXNlcjEzNjc4NzE5", "organizations_url": "https://api.github.com/users/QingGo/orgs", "received_events_url": "https://api.github.com/users/QingGo/received_events", "repos_url": "https://api.github.com/users/QingGo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/QingGo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/QingGo/subscriptions", "type": "User", "url": "https://api.github.com/users/QingGo", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "While using verification_mode='no_checks' parameter in load_dataset() can bypass this validation, this solution is not intuitive or convenient for most users, especially those who are not familiar with all the parameters of the load_dataset() function.\n\n```python\nbook_corpus_ds = load_dataset(\n \"SaylorTwift/the_pile_books3_minus_gutenberg\",\n name=\"default\",\n data_files=\"data/train-00000-of-00213-312fd8d7a3c58a63.parquet\",\n split=\"train\",\n cache_dir=\"./data\",\n verification_mode='no_checks'\n)\n```", "Thanks for the report and reproduction steps @QingGo \n@lhoestq which one of the following looks like a nicer way to handle this?\n\n1] Skip split-size validation entirely for partial loads\nIf the user passes data_files manually and it represents only a subset, then verify_splits() should simply not run, or skip validation only for that split.\n\n2] Replace the error with a warning\n\n3] Automatically detect partial-load cases(i mean we can try this out!)\n\nAssume this, \nIf data_files is provided AND\nthe number of provided files ≠ number of expected files in metadata,\nthen treat it as a partial load and disable strict verification.\n" ]
2025-11-13T12:03:23Z
2025-11-16T15:39:23Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug When loading only a subset of dataset files while the dataset's README.md contains split metadata, the system throws a NonMatchingSplitsSizesError . This prevents users from loading partial datasets for quick validation in cases of poor network conditions or very large datasets. ### Steps to reproduce the bug 1. Use the Hugging Face `datasets` library to load a dataset with only specific files specified 2. Ensure the dataset repository has split metadata defined in README.md 3. Observe the error when attempting to load a subset of files ```python # Example code that triggers the error from datasets import load_dataset book_corpus_ds = load_dataset( "SaylorTwift/the_pile_books3_minus_gutenberg", name="default", data_files="data/train-00000-of-00213-312fd8d7a3c58a63.parquet", split="train", cache_dir="./data" ) ``` ### Error Message ``` Traceback (most recent call last): File "/Users/QingGo/code/llm_learn/src/data/clean_cc_bc.py", line 13, in <module> book_corpus_ds = load_dataset( "SaylorTwift/the_pile_books3_minus_gutenberg", ... File "/Users/QingGo/code/llm_learn/.venv/lib/python3.13/site-packages/datasets/utils/info_utils.py", line 77, in verify_splits raise NonMatchingSplitsSizesError(str(bad_splits)) datasets.exceptions.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=106199627990.47722, num_examples=192661, shard_lengths=None, dataset_name=None), 'recorded': SplitInfo(name='train', num_bytes=454897326, num_examples=905, shard_lengths=None, dataset_name='the_pile_books3_minus_gutenberg')}] ``` ### Expected behavior When loading partial dataset files, the system should: 1. Skip the `NonMatchingSplitsSizesError` validation, OR 2. Only log a warning message instead of raising an error ### Environment info - `datasets` version: 4.3.0 - Platform: macOS-15.7.1-arm64-arm-64bit-Mach-O - Python version: 3.13.2 - `huggingface_hub` version: 0.36.0 - PyArrow version: 22.0.0 - Pandas version: 2.3.3 - `fsspec` version: 2025.9.0
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7867/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7867/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7866
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7866/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7866/comments
https://api.github.com/repos/huggingface/datasets/issues/7866/events
https://github.com/huggingface/datasets/pull/7866
3,620,436,248
PR_kwDODunzps6zL7Sz
7,866
docs: add Python version requirement note to installation section
{ "avatar_url": "https://avatars.githubusercontent.com/u/222381706?v=4", "events_url": "https://api.github.com/users/ananthasai-2006/events{/privacy}", "followers_url": "https://api.github.com/users/ananthasai-2006/followers", "following_url": "https://api.github.com/users/ananthasai-2006/following{/other_user}", "gists_url": "https://api.github.com/users/ananthasai-2006/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ananthasai-2006", "id": 222381706, "login": "ananthasai-2006", "node_id": "U_kgDODUFGig", "organizations_url": "https://api.github.com/users/ananthasai-2006/orgs", "received_events_url": "https://api.github.com/users/ananthasai-2006/received_events", "repos_url": "https://api.github.com/users/ananthasai-2006/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ananthasai-2006/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ananthasai-2006/subscriptions", "type": "User", "url": "https://api.github.com/users/ananthasai-2006", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-13T09:54:35Z
2025-11-13T09:54:35Z
null
NONE
null
null
null
null
Added note about Python version requirement for conda installation.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7866/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7866/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7866.diff", "html_url": "https://github.com/huggingface/datasets/pull/7866", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7866.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7866" }
true
https://api.github.com/repos/huggingface/datasets/issues/7865
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7865/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7865/comments
https://api.github.com/repos/huggingface/datasets/issues/7865/events
https://github.com/huggingface/datasets/pull/7865
3,620,116,195
PR_kwDODunzps6zK2H_
7,865
[FEAT] MIDI feature support
{ "avatar_url": "https://avatars.githubusercontent.com/u/2518789?v=4", "events_url": "https://api.github.com/users/frascuchon/events{/privacy}", "followers_url": "https://api.github.com/users/frascuchon/followers", "following_url": "https://api.github.com/users/frascuchon/following{/other_user}", "gists_url": "https://api.github.com/users/frascuchon/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/frascuchon", "id": 2518789, "login": "frascuchon", "node_id": "MDQ6VXNlcjI1MTg3ODk=", "organizations_url": "https://api.github.com/users/frascuchon/orgs", "received_events_url": "https://api.github.com/users/frascuchon/received_events", "repos_url": "https://api.github.com/users/frascuchon/repos", "site_admin": false, "starred_url": "https://api.github.com/users/frascuchon/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/frascuchon/subscriptions", "type": "User", "url": "https://api.github.com/users/frascuchon", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7865). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-13T08:31:51Z
2025-11-14T13:58:52Z
null
NONE
null
null
null
null
This PR adds a new `Midi` feature for reading and importing MIDI files into the datasets.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7865/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7865/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7865.diff", "html_url": "https://github.com/huggingface/datasets/pull/7865", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7865.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7865" }
true
https://api.github.com/repos/huggingface/datasets/issues/7864
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7864/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7864/comments
https://api.github.com/repos/huggingface/datasets/issues/7864/events
https://github.com/huggingface/datasets/issues/7864
3,619,137,823
I_kwDODunzps7Xt6kf
7,864
add_column and add_item erroneously(?) require new_fingerprint parameter
{ "avatar_url": "https://avatars.githubusercontent.com/u/17151810?v=4", "events_url": "https://api.github.com/users/echthesia/events{/privacy}", "followers_url": "https://api.github.com/users/echthesia/followers", "following_url": "https://api.github.com/users/echthesia/following{/other_user}", "gists_url": "https://api.github.com/users/echthesia/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/echthesia", "id": 17151810, "login": "echthesia", "node_id": "MDQ6VXNlcjE3MTUxODEw", "organizations_url": "https://api.github.com/users/echthesia/orgs", "received_events_url": "https://api.github.com/users/echthesia/received_events", "repos_url": "https://api.github.com/users/echthesia/repos", "site_admin": false, "starred_url": "https://api.github.com/users/echthesia/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/echthesia/subscriptions", "type": "User", "url": "https://api.github.com/users/echthesia", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Take this with a grain of salt, this is just my personal understanding:\nWhile you technically can overwrite the new_fingerprint with a string, e.g.\n```python\nt = d.add_column(\"new_column\", col_value, new_fingerprint=\"dummy_fp\")\nassert t._fingerprint == \"dummy_fp\" # this is true and will pass\n```\nthis is not desired since the fingerprint should be calculated based on the operations (and their arguments) to be unique. This is handled by the [fingerprint_transform](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L6077) function which needs a \"new_fingerprint\" keyword argument and creates a unique hash if its value is not set, see [here](https://github.com/huggingface/datasets/blob/main/src/datasets/fingerprint.py#L432). So it is probably safer to not document this keyword, since one doesn't want the user to actually use it and it's only a feature in very limited cases for people really knowing what they are doing. The thing that might be bugging people who read the code is that `new_fingerprint` seems to be required for `add_item` and `add_column` but it is actually set by the decorator (in which's definition it is optional), so maybe changing the signature of `add_item` and `add_column` to `new_fingerprint: Optional[str] = None` would make sense, since this is also how it's handled in the other cases (created by claude):\n\n - [flatten](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L2034)\n - [cast_column](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L2165)\n - [remove_columns](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L2209)\n - [rename_column](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L2263)\n - [rename_columns](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L2329)\n - [select_columns](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L2397)\n - [batch](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L3760)\n - [filter](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L3813)\n - [flatten_indices](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L3959)\n - [select](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L4038)\n - [_select_contiguous](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L4128)\n - [sort](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L4376)\n - [shuffle](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L4506)\n - [train_test_split](https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L4641)\nSo as you mentioned, I believe the methods erronously require the `new_fingerprint` parameter and making them optional is a little consistency win." ]
2025-11-13T02:56:49Z
2025-11-24T20:33:59Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug Contradicting their documentation (which doesn't mention the parameter at all), both Dataset.add_column and Dataset.add_item require a new_fingerprint string. This parameter is passed directly to the dataset constructor, which has the fingerprint parameter listed as optional; is there any reason it shouldn't be optional in these methods as well? ### Steps to reproduce the bug Reproduction steps: 1. Look at the function signature for add_column: https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L6078 2. Repeat for add_item: https://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/arrow_dataset.py#L6336 ### Expected behavior add_column and add_item should either set the fingerprint parameter to optional or include it in their docstrings ### Environment info Not environment-dependent
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7864/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7864/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7863
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7863/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7863/comments
https://api.github.com/repos/huggingface/datasets/issues/7863/events
https://github.com/huggingface/datasets/issues/7863
3,618,836,821
I_kwDODunzps7XsxFV
7,863
Support hosting lance / vortex / iceberg / zarr datasets on huggingface hub
{ "avatar_url": "https://avatars.githubusercontent.com/u/3664715?v=4", "events_url": "https://api.github.com/users/pavanramkumar/events{/privacy}", "followers_url": "https://api.github.com/users/pavanramkumar/followers", "following_url": "https://api.github.com/users/pavanramkumar/following{/other_user}", "gists_url": "https://api.github.com/users/pavanramkumar/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/pavanramkumar", "id": 3664715, "login": "pavanramkumar", "node_id": "MDQ6VXNlcjM2NjQ3MTU=", "organizations_url": "https://api.github.com/users/pavanramkumar/orgs", "received_events_url": "https://api.github.com/users/pavanramkumar/received_events", "repos_url": "https://api.github.com/users/pavanramkumar/repos", "site_admin": false, "starred_url": "https://api.github.com/users/pavanramkumar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pavanramkumar/subscriptions", "type": "User", "url": "https://api.github.com/users/pavanramkumar", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[ "Kudos!", "So cool ! Would love to see support for lance :)", "@lhoestq thanks for your support! Any suggestions across `datasets` or `huggingface_hub` projects to make this happen?\n\nI just noticed this blog post: https://huggingface.co/blog/streaming-datasets\n\nDo you know if `hfFileSystem` from `huggingface_hub` is flexible enough to accommodate lance? I don't want to `open` and scan a file, I want to create generators with the `lance.dataset.to_batches()` from each fragment (partition) that I can iterate over in a distributed dataloader.\n\nIdeally, something like this should just work:\n\n```\nimport lance\nlance_ds_path = f\"hf://datasets/{dataset_id}/{path_in_repo}.lance\"\nds = lance.dataset(lance_ds_path)\nfragments = ds.get_fragments()\nfragment_generators = []\nfor fragment in fragments:\n fragment_generators = fragment.to_batches()\n```\n\nLooking at the huggingface blog post, I think we might need a PR into `pyarrow` to create a `LanceFragmentScanOptions` class that subclasses [pyarrow.dataset.FragmentScanOptions](https://arrow.apache.org/docs/python/generated/pyarrow.dataset.FragmentScanOptions.html#pyarrow.dataset.FragmentScanOptions) cc @prrao87, @changhiskhan", "> Do you know if HfFileSystem from huggingface_hub is flexible enough to accommodate lance?\n\nit provides file-like objects for files on HF, and works using range requests. PyArrow uses HfFileSystem for HF files already\n\nThough in the Parquet / PyArrow case the data is read generally row group per row group (using range requests with a minimum size `range_size_limit ` to optimize I/O in case of small row groups)\n\nPS: there is an equivalent to HfFileSystem in rust in OpenDAL, but it only supports read from HF, not write (yet ?)\n\n> I don't want to open and scan a file, I want to create generators with the lance.dataset.to_batches() from each fragment (partition) that I can iterate over in a distributed dataloader.\n\nWe do something very similar for Parquet here: \n\nhttps://github.com/huggingface/datasets/blob/17f40a318a1f8c7d33c2a4dd17934f81d14a7f57/src/datasets/packaged_modules/parquet/parquet.py#L168-L169", "Hi, I work on the Lance project. We'd be happy to see the format supported on huggingface hub.\n\nIt's not clear to me from this thread what is required for that. Could we clarify that? Are there examples we can point to?\n\n> I think we might need a PR into `pyarrow` to create a `LanceFragmentScanOptions` class that subclasses [pyarrow.dataset.FragmentScanOptions](https://arrow.apache.org/docs/python/generated/pyarrow.dataset.FragmentScanOptions.html#pyarrow.dataset.FragmentScanOptions)\n\nCould you elaborate why a `FragmentScanOptions` subclass is required? Also, if it is, we could just define that as a subclass within the `pylance` module, unless I'm missing something.\n\nLance supports OpenDAL storage, so I think we could add support for huggingface's filesystem through that and make sure it's exposed in pylance. Could also help implement some write operations. Perhaps that's the main blocker? ", "> PS: there is an equivalent to HfFileSystem in rust in OpenDAL, but it only supports read from HF, not write (yet ?)\n\nHi, I’m willing to add full-fledged support for the HF file system. This shouldn’t be considered a blocker. 🤟 ", "Exposing the existing HF filesystem from OpenDAL in pylance would be great ! and a good first step\n\nExcited for write operations too", "Thanks @lhoestq @wjones127 @Xuanwo ! I think we have all the necessary people on this thread now to make it happen :)\n\n> Could you elaborate why a FragmentScanOptions subclass is required? Also, if it is, we could just define that as a subclass within the pylance module, unless I'm missing something.\n\n@wjones127 I'm not actually sure this is needed but I'm guessing based on [this blog post](https://huggingface.co/blog/streaming-datasets) from a couple of weeks ago. Specifically, this section which allows creation of a dataset object with configurable prefetching:\n\n```\nimport pyarrow\nimport pyarrow.dataset\n\nfragment_scan_options = pyarrow.dataset.ParquetFragmentScanOptions(\n cache_options=pyarrow.CacheOptions(\n prefetch_limit=1,\n range_size_limit=128 << 20\n ),\n)\nds = load_dataset(parquet_dataset_id, streaming=True, fragment_scan_options=fragment_scan_options)\n```\n\nI might be completely wrong that we do need an equivalent `LanceFragmentScanOptions` PR into `pyarrow` and the `OpenDAL` path might be sufficient.\n\nI really just want something like this to work out of the box:\n\n```\nimport lance\nlance_ds_path = f\"hf://datasets/{dataset_id}/{path_in_repo}.lance\"\nds = lance.dataset(lance_ds_path)\nfragments = ds.get_fragments()\nfragment_generators = []\nfor fragment in fragments:\n fragment_generators = fragment.to_batches()\n```\n\nIn the ideal case, I'd like to be able to control prefetch configuration via arguments to `to_batches()` like the ones that already exist for a lance dataset on any S3-compatible object store.\n\nWould a useful approach be to create a toy lance dataset on huggingface and see if this \"just works\"; then work backwards from there?\n\nAs for writing, I'm looking to migrate datasets from my own private S3-compatible object store bucket (Tigris Data) to huggingface datasets but ~~I'm 100% sure~~ I'm _not_ 100% sure whether we even need `hfFileSystem` compatible write capability\n\n\n", "Here's a public dataset which could be a working example to work backwards from:\n\nhttps://huggingface.co/datasets/pavan-ramkumar/test-slaf\n\npylance currently looks for default object store backends and returns this `ValueError`\n\n```\n>>> import lance\n>>> hf_path = \"hf://datasets/pavan-ramkumar/test-slaf/tree/main/synthetic_50k_processed_v21.slaf/expression.lance\"\n>>> ds = lance.dataset(hf_path)\nTraceback (most recent call last):\n File \"<stdin>\", line 1, in <module>\n File \"/Users/pavan/slaf-project/slaf/.venv/lib/python3.12/site-packages/lance/__init__.py\", line 145, in dataset\n ds = LanceDataset(\n ^^^^^^^^^^^^^\n File \"/Users/pavan/slaf-project/slaf/.venv/lib/python3.12/site-packages/lance/dataset.py\", line 425, in __init__\n self._ds = _Dataset(\n ^^^^^^^^^\nValueError: Invalid user input: No object store provider found for scheme: 'hf'\nValid schemes: gs, memory, s3, az, file-object-store, file, oss, s3+ddb, /Users/runner/work/lance/lance/rust/lance-io/src/object_store/providers.rs:161:54\n```", "@Xuanwo @wjones127 just checking in to see if you had a chance to add a huggingface provider via opendal to pylance. I'm assuming we need a new `huggingface.rs` provider [here](https://github.com/lance-format/lance/tree/4d9c1a4d459ea486556de0ee90828a442d0425b0/rust/lance-io/src/object_store/providers).\n\nDo let me know if I can do anything to help, really excited to help stream lance datasets from huggingface hub", "> @Xuanwo @wjones127 just checking in to see if you had a chance to add a huggingface provider via opendal to pylance. I'm assuming we need a new `huggingface.rs` provider [here](https://github.com/lance-format/lance/tree/4d9c1a4d459ea486556de0ee90828a442d0425b0/rust/lance-io/src/object_store/providers).\n> \n> Do let me know if I can do anything to help, really excited to help stream lance datasets from huggingface hub\n\nI'm willing to work on this! Would you like to create an issue on lance side and ping me there?", " > I'm willing to work on this! Would you like to create an issue on lance side and ping me there?\n\nDone! [Link](https://github.com/lance-format/lance/issues/5346)\n", "@pavanramkumar pls check this out once it's merged! https://github.com/lance-format/lance/pull/5353" ]
2025-11-13T00:51:07Z
2025-11-26T14:10:29Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Feature request Huggingface datasets has great support for large tabular datasets in parquet with large partitions. I would love to see two things in the future: - equivalent support for `lance`, `vortex`, `iceberg`, `zarr` (in that order) in a way that I can stream them using the datasets library - more fine-grained control of streaming, so that I can stream at the partition / shard level ### Motivation I work with very large `lance` datasets on S3 and often require random access for AI/ML applications like multi-node training. I was able to achieve high throughput dataloading on a lance dataset with ~150B rows by building distributed dataloaders that can be scaled both vertically (until i/o and CPU are saturated), and then horizontally (to workaround network bottlenecks). Using this strategy I was able to achieve 10-20x the throughput of the streaming data loader from the `huggingface/datasets` library. I realized that these would be great features for huggingface to support natively ### Your contribution I'm not ready yet to make a PR but open to it with the right pointers!
null
{ "+1": 4, "-1": 0, "confused": 0, "eyes": 2, "heart": 5, "hooray": 2, "laugh": 2, "rocket": 8, "total_count": 23, "url": "https://api.github.com/repos/huggingface/datasets/issues/7863/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7863/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7862
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7862/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7862/comments
https://api.github.com/repos/huggingface/datasets/issues/7862/events
https://github.com/huggingface/datasets/pull/7862
3,617,947,090
PR_kwDODunzps6zDjEj
7,862
Add flatten_indices option to save_to_disk method
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "as said by @KCKawalkar used below script to test - \r\n\r\nBEFORE PATCH - \r\nTEST.PY:\r\n```\r\nfrom datasets import Dataset\r\nimport time\r\n\r\ndataset = Dataset.from_dict({'text': [f'sample {i}' for i in range(100000)]})\r\n\r\n# Baseline save (no indices)\r\nstart = time.time()\r\ndataset.save_to_disk('baseline')\r\nbaseline_time = time.time() - start\r\n\r\n# Filtered save (creates indices)\r\nfiltered = dataset.filter(lambda x: True)\r\nstart = time.time()\r\nfiltered.save_to_disk('filtered')\r\nfiltered_time = time.time() - start\r\n\r\nprint(f\"Baseline: {baseline_time:.3f}s\")\r\nprint(f\"Filtered: {filtered_time:.3f}s\")\r\nprint(f\"Slowdown: {(filtered_time/baseline_time-1)*100:.1f}%\")\r\n```\r\nRESULTS:\r\n```\r\n@ArjunJagdale ➜ /workspaces/datasets (main) $ python test_arjun.py\r\nSaving the dataset (1/1 shards): 100%|█████████████████████████████████████████████████████████████████████████████████████████████| 100000/100000 [00:00<00:00, 3030654.07 examples/s]\r\nFilter: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 100000/100000 [00:00<00:00, 576296.61 examples/s]\r\nSaving the dataset (1/1 shards): 100%|██████████████████████████████████████████████████████████████████████████████████████████████| 100000/100000 [00:00<00:00, 310565.19 examples/s]\r\nBaseline: 0.035s\r\nFiltered: 0.323s\r\nSlowdown: 813.4%\r\n```\r\n\r\nAFTER PATCH - \r\nTEST.PY:\r\n```\r\nfrom datasets import Dataset\r\nimport time\r\n\r\n# Create dataset\r\ndataset = Dataset.from_dict({'text': [f'sample {i}' for i in range(100000)]})\r\n\r\n# Baseline save (no indices)\r\nstart = time.time()\r\ndataset.save_to_disk('baseline')\r\nbaseline_time = time.time() - start\r\n\r\n# Filtered save (creates indices)\r\nfiltered = dataset.filter(lambda x: True)\r\nstart = time.time()\r\nfiltered.save_to_disk('filtered', flatten_indices=False)\r\nfiltered_time = time.time() - start\r\n\r\nprint(f\"Baseline: {baseline_time:.3f}s\")\r\nprint(f\"Filtered: {filtered_time:.3f}s\") \r\nprint(f\"Slowdown: {(filtered_time/baseline_time-1)*100:.1f}%\")\r\n```\r\n\r\nREESULT:\r\n```\r\n@ArjunJagdale ➜ /workspaces/datasets (main) $ python test_arjun.py\r\nSaving the dataset (1/1 shards): 100%|█████████████████████████████████████████████████████████████████████████████████████████████| 100000/100000 [00:00<00:00, 3027482.12 examples/s]\r\nFilter: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 100000/100000 [00:00<00:00, 468901.89 examples/s]\r\nSaving the dataset (1/1 shards): 100%|██████████████████████████████████████████████████████████████████████████████████████████████| 100000/100000 [00:00<00:00, 324036.36 examples/s]\r\nBaseline: 0.036s\r\nFiltered: 0.310s\r\nSlowdown: 771.1%\r\n\r\n```" ]
2025-11-12T19:38:51Z
2025-11-12T19:50:20Z
null
CONTRIBUTOR
null
null
null
null
Added flatten_indices parameter to control index flattening during dataset saving. Solves #7861 This PR introduces a new optional argument, flatten_indices, to the save_to_disk methods in both Dataset and DatasetDict. The change allows users to skip the expensive index-flattening step when saving datasets that already use index mappings (e.g., after filter() or shuffle()), resulting in significant speed improvements for large datasets while maintaining backward compatibility. While not a huge absolute difference at 100K rows, the improvement scales significantly with larger datasets (millions of rows). This patch gives users control — they can disable flattening when they don’t need it, avoiding unnecessary rewrites. @lhoestq WDYT?
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7862/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7862/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7862.diff", "html_url": "https://github.com/huggingface/datasets/pull/7862", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7862.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7862" }
true
https://api.github.com/repos/huggingface/datasets/issues/7861
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7861/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7861/comments
https://api.github.com/repos/huggingface/datasets/issues/7861/events
https://github.com/huggingface/datasets/issues/7861
3,611,821,713
I_kwDODunzps7XSAaR
7,861
Performance Issue: save_to_disk() 200-1200% slower due to unconditional flatten_indices()
{ "avatar_url": "https://avatars.githubusercontent.com/u/222552287?v=4", "events_url": "https://api.github.com/users/KCKawalkar/events{/privacy}", "followers_url": "https://api.github.com/users/KCKawalkar/followers", "following_url": "https://api.github.com/users/KCKawalkar/following{/other_user}", "gists_url": "https://api.github.com/users/KCKawalkar/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/KCKawalkar", "id": 222552287, "login": "KCKawalkar", "node_id": "U_kgDODUPg3w", "organizations_url": "https://api.github.com/users/KCKawalkar/orgs", "received_events_url": "https://api.github.com/users/KCKawalkar/received_events", "repos_url": "https://api.github.com/users/KCKawalkar/repos", "site_admin": false, "starred_url": "https://api.github.com/users/KCKawalkar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/KCKawalkar/subscriptions", "type": "User", "url": "https://api.github.com/users/KCKawalkar", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-11T11:05:38Z
2025-11-11T11:05:38Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
## 🐛 Bug Description The `save_to_disk()` method unconditionally calls `flatten_indices()` when `_indices` is not None, causing severe performance degradation for datasets processed with filtering, shuffling, or multiprocessed mapping operations. **Root cause**: This line rebuilds the entire dataset unnecessarily: ```python dataset = self.flatten_indices() if self._indices is not None else self ``` ## 📊 Performance Impact | Dataset Size | Operation | Save Time | Slowdown | |-------------|-----------|-----------|----------| | 100K | Baseline (no indices) | 0.027s | - | | 100K | Filtered (with indices) | 0.146s | **+431%** | | 100K | Shuffled (with indices) | 0.332s | **+1107%** | | 250K | Shuffled (with indices) | 0.849s | **+1202%** | ## 🔄 Reproduction ```python from datasets import Dataset import time # Create dataset dataset = Dataset.from_dict({'text': [f'sample {i}' for i in range(100000)]}) # Baseline save (no indices) start = time.time() dataset.save_to_disk('baseline') baseline_time = time.time() - start # Filtered save (creates indices) filtered = dataset.filter(lambda x: True) start = time.time() filtered.save_to_disk('filtered') filtered_time = time.time() - start print(f"Baseline: {baseline_time:.3f}s") print(f"Filtered: {filtered_time:.3f}s") print(f"Slowdown: {(filtered_time/baseline_time-1)*100:.1f}%") ``` **Expected output**: Filtered dataset is 400-1000% slower than baseline ## 💡 Proposed Solution Add optional parameter to control flattening: ```python def save_to_disk(self, dataset_path, flatten_indices=True): dataset = self.flatten_indices() if (self._indices is not None and flatten_indices) else self # ... rest of save logic ``` **Benefits**: - ✅ Immediate performance improvement for users who don't need flattening - ✅ Backwards compatible (default behavior unchanged) - ✅ Simple implementation ## 🌍 Environment - **datasets version**: 2.x - **Python**: 3.10+ - **OS**: Linux/macOS/Windows ## 📈 Impact This affects **most ML preprocessing workflows** that filter/shuffle datasets before saving. Performance degradation scales exponentially with dataset size, making it a critical bottleneck for production systems. ## 🔗 Additional Resources We have comprehensive test scripts demonstrating this across multiple scenarios if needed for further investigation.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7861/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7861/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7860
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7860/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7860/comments
https://api.github.com/repos/huggingface/datasets/issues/7860/events
https://github.com/huggingface/datasets/pull/7860
3,610,706,034
PR_kwDODunzps6yrHQN
7,860
Support loading local arrow datasets via load_dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/16986130?v=4", "events_url": "https://api.github.com/users/gstrat88/events{/privacy}", "followers_url": "https://api.github.com/users/gstrat88/followers", "following_url": "https://api.github.com/users/gstrat88/following{/other_user}", "gists_url": "https://api.github.com/users/gstrat88/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/gstrat88", "id": 16986130, "login": "gstrat88", "node_id": "MDQ6VXNlcjE2OTg2MTMw", "organizations_url": "https://api.github.com/users/gstrat88/orgs", "received_events_url": "https://api.github.com/users/gstrat88/received_events", "repos_url": "https://api.github.com/users/gstrat88/repos", "site_admin": false, "starred_url": "https://api.github.com/users/gstrat88/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gstrat88/subscriptions", "type": "User", "url": "https://api.github.com/users/gstrat88", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-11-11T04:58:33Z
2025-11-11T20:58:46Z
null
NONE
null
null
null
null
Load_dataset will handle local saved datasets that way #7018
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7860/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7860/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7860.diff", "html_url": "https://github.com/huggingface/datasets/pull/7860", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7860.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7860" }
true
https://api.github.com/repos/huggingface/datasets/issues/7859
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7859/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7859/comments
https://api.github.com/repos/huggingface/datasets/issues/7859/events
https://github.com/huggingface/datasets/pull/7859
3,608,586,063
PR_kwDODunzps6yj-aZ
7,859
fix some broken links
{ "avatar_url": "https://avatars.githubusercontent.com/u/326577?v=4", "events_url": "https://api.github.com/users/julien-c/events{/privacy}", "followers_url": "https://api.github.com/users/julien-c/followers", "following_url": "https://api.github.com/users/julien-c/following{/other_user}", "gists_url": "https://api.github.com/users/julien-c/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/julien-c", "id": 326577, "login": "julien-c", "node_id": "MDQ6VXNlcjMyNjU3Nw==", "organizations_url": "https://api.github.com/users/julien-c/orgs", "received_events_url": "https://api.github.com/users/julien-c/received_events", "repos_url": "https://api.github.com/users/julien-c/repos", "site_admin": false, "starred_url": "https://api.github.com/users/julien-c/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/julien-c/subscriptions", "type": "User", "url": "https://api.github.com/users/julien-c", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7859). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-10T15:34:46Z
2025-11-10T17:11:07Z
2025-11-10T17:11:05Z
MEMBER
null
null
null
null
would be cool to automate finding those broken links as i think there might be many of them @lhoestq @albertvillanova
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7859/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7859/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7859.diff", "html_url": "https://github.com/huggingface/datasets/pull/7859", "merged_at": "2025-11-10T17:11:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/7859.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7859" }
true
https://api.github.com/repos/huggingface/datasets/issues/7858
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7858/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7858/comments
https://api.github.com/repos/huggingface/datasets/issues/7858/events
https://github.com/huggingface/datasets/pull/7858
3,605,471,548
PR_kwDODunzps6yZq4r
7,858
Support downloading specific splits in `load_dataset`
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "@CloseChoice This looks great! You're absolutely right about the missing comparison - that's a critical bug I missed. " ]
2025-11-09T20:44:00Z
2025-11-11T08:04:14Z
null
CONTRIBUTOR
null
null
null
null
This is PR builds on top of #7706 to revive the unfinished #6832 but isn't just cleaning up things, here are some important changes: - `download_mode="FORCE_REDOWNLOAD"` is interpreted as always creating a clean slate, that means that even if we already did: ```python load_dataset("<name>") load_dataset("<name>", split="train", download_mode="force_redownload") ``` This makes sure that only the train dataset is available after executing both. This was different in the original PR, which proposed that train and test would be available. - `download_mode="REUSE_DATASET_IF_EXISTS"` is interpreted as only ever adding new data, never redownloading OR deleting other splits. This was different in the original PR, where ```python load_dataset("<name>", split="test") load_dataset("<name>", split="train") ``` resulted in only the train data being available, which I deem very unintuitive and probably not what users want. Also I argue that this is just the first step to a more user friendly partial loading when specifying percentages (or maybe even single instances) via the ReadInstructions, and then doing ```python load_dataset("<name>", split="test[:10%]") load_dataset("<name>", split="test[10%:]") ``` should result IMO in the whole dataset being cached locally without redownloads. Furthermore this PR fixes a couple issues with the previous PR, e.g. a [missing comparison](https://github.com/huggingface/datasets/pull/7706/files#diff-f933ce41f71c6c0d1ce658e27de62cbe0b45d777e9e68056dd012ac3eb9324f7R877) and adding tests for the proposed changes in behaviour, which would both fail on @ArjunJagdale's original PR. Todo: - [ ] update docs? Future outlook (just my opinions and up for debate): As mentioned before, I would see this as just a step towards the feature of partial percentage loading (though how the API should behave in that case is not entirely clear for me now) and maybe we could introduce another `download_mode="FORCE_REDOWNLOAD_SPLIT"`, which makes sure that even if a split is specified, only the referenced split is redownloaded and everything else left unchanged, this would then allow users more granular control over what they want to redownload. @lhoestq very curious to get your opinion on this.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 2, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/7858/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7858/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7858.diff", "html_url": "https://github.com/huggingface/datasets/pull/7858", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7858.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7858" }
true
https://api.github.com/repos/huggingface/datasets/issues/7856
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7856/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7856/comments
https://api.github.com/repos/huggingface/datasets/issues/7856/events
https://github.com/huggingface/datasets/issues/7856
3,603,729,142
I_kwDODunzps7WzIr2
7,856
Missing transcript column when loading a local dataset with "audiofolder"
{ "avatar_url": "https://avatars.githubusercontent.com/u/10166907?v=4", "events_url": "https://api.github.com/users/gweltou/events{/privacy}", "followers_url": "https://api.github.com/users/gweltou/followers", "following_url": "https://api.github.com/users/gweltou/following{/other_user}", "gists_url": "https://api.github.com/users/gweltou/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/gweltou", "id": 10166907, "login": "gweltou", "node_id": "MDQ6VXNlcjEwMTY2OTA3", "organizations_url": "https://api.github.com/users/gweltou/orgs", "received_events_url": "https://api.github.com/users/gweltou/received_events", "repos_url": "https://api.github.com/users/gweltou/repos", "site_admin": false, "starred_url": "https://api.github.com/users/gweltou/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gweltou/subscriptions", "type": "User", "url": "https://api.github.com/users/gweltou", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "First bad commit 5c8869f8c36dbc8c8d423030b7b7c4fd64f8c729\n\nEDIT: This is not a bug or a regression. It was a breaking change introduced in the commit I mentioned and was also documented in there. The docs state how to handle this now, see https://huggingface.co/docs/datasets/main/en/audio_load#audiofolder-with-metadata\n\nor simply, move your metadata into the splits folder and update the paths, in your case this would look like this:\n```bash\nmy_dataset/\n - data/\n - test/\n - 54db8760de3cfbff3c8a36a36b4d0f77_00390.0_04583.0.mp3\n - 54db8760de3cfbff3c8a36a36b4d0f77_04583.0_05730.0.mp3\n - metadata.jsonl\n```\n\nand the pahts in the jsonl should be relative to the metadata.json:\n```bash\n{\"file_name\": \"54db8760de3cfbff3c8a36a36b4d0f77_00390.0_04583.0.mp3\", \"transcript\": \"Ata tudoù penaos e tro ar bed ?\"}\n{\"file_name\": \"54db8760de3cfbff3c8a36a36b4d0f77_04583.0_05730.0.mp3\", \"transcript\": \"Ur gwir blijadur eo adkavout ac'hanoc'h hiziv.\"}\n...\n```\n\nSo I think this can be closed.", "Thank you for your quick answer !\nI'm sorry I missed that in the documentation.\nEverything works fine again after following your recommendations.\nI'm closing the issue." ]
2025-11-08T16:27:58Z
2025-11-09T12:13:38Z
2025-11-09T12:13:38Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug My local dataset is not properly loaded when using `load_dataset("audiofolder", data_dir="my_dataset")` with a `jsonl` metadata file. Only the `audio` column is read while the `transcript` column is not. The last tested `datasets` version where the behavior was still correct is 2.18.0. ### Steps to reproduce the bug Dataset directory structure: ``` my_dataset/ - data/ - test/ - 54db8760de3cfbff3c8a36a36b4d0f77_00390.0_04583.0.mp3 - 54db8760de3cfbff3c8a36a36b4d0f77_04583.0_05730.0.mp3 - ... - metadata.jsonl ``` `metadata.jsonl` file content: ``` {"file_name": "data/test/54db8760de3cfbff3c8a36a36b4d0f77_00390.0_04583.0.mp3", "transcript": "Ata tudoù penaos e tro ar bed ?"} {"file_name": "data/test/54db8760de3cfbff3c8a36a36b4d0f77_04583.0_05730.0.mp3", "transcript": "Ur gwir blijadur eo adkavout ac'hanoc'h hiziv."} ... ``` ```python3 my_dataset = load_dataset("audiofolder", data_dir="my_dataset") print(my_dataset) ''' DatasetDict({ test: Dataset({ features: ['audio'], num_rows: 347 }) }) ''' print(my_dataset['test'][0]) ''' {'audio': <datasets.features._torchcodec.AudioDecoder object at 0x75ffcd172510>} ''' ``` ### Expected behavior Being able to access the `transcript` column in the loaded dataset. ### Environment info - `datasets` version: 4.4.1 - Platform: Linux-6.5.0-45-generic-x86_64-with-glibc2.39 - Python version: 3.13.9 - `huggingface_hub` version: 1.1.2 - PyArrow version: 22.0.0 - Pandas version: 2.3.3 - `fsspec` version: 2025.10.0 Note: same issue with `datasets` v3.6.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/10166907?v=4", "events_url": "https://api.github.com/users/gweltou/events{/privacy}", "followers_url": "https://api.github.com/users/gweltou/followers", "following_url": "https://api.github.com/users/gweltou/following{/other_user}", "gists_url": "https://api.github.com/users/gweltou/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/gweltou", "id": 10166907, "login": "gweltou", "node_id": "MDQ6VXNlcjEwMTY2OTA3", "organizations_url": "https://api.github.com/users/gweltou/orgs", "received_events_url": "https://api.github.com/users/gweltou/received_events", "repos_url": "https://api.github.com/users/gweltou/repos", "site_admin": false, "starred_url": "https://api.github.com/users/gweltou/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gweltou/subscriptions", "type": "User", "url": "https://api.github.com/users/gweltou", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7856/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7856/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7855
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7855/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7855/comments
https://api.github.com/repos/huggingface/datasets/issues/7855/events
https://github.com/huggingface/datasets/pull/7855
3,602,216,153
PR_kwDODunzps6yPIRy
7,855
ArXiv -> HF Papers
{ "avatar_url": "https://avatars.githubusercontent.com/u/45557362?v=4", "events_url": "https://api.github.com/users/qgallouedec/events{/privacy}", "followers_url": "https://api.github.com/users/qgallouedec/followers", "following_url": "https://api.github.com/users/qgallouedec/following{/other_user}", "gists_url": "https://api.github.com/users/qgallouedec/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/qgallouedec", "id": 45557362, "login": "qgallouedec", "node_id": "MDQ6VXNlcjQ1NTU3MzYy", "organizations_url": "https://api.github.com/users/qgallouedec/orgs", "received_events_url": "https://api.github.com/users/qgallouedec/received_events", "repos_url": "https://api.github.com/users/qgallouedec/repos", "site_admin": false, "starred_url": "https://api.github.com/users/qgallouedec/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/qgallouedec/subscriptions", "type": "User", "url": "https://api.github.com/users/qgallouedec", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7855). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-07T22:16:36Z
2025-11-10T15:01:13Z
2025-11-10T15:01:13Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7855/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7855/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7855.diff", "html_url": "https://github.com/huggingface/datasets/pull/7855", "merged_at": "2025-11-10T15:01:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/7855.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7855" }
true
https://api.github.com/repos/huggingface/datasets/issues/7854
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7854/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7854/comments
https://api.github.com/repos/huggingface/datasets/issues/7854/events
https://github.com/huggingface/datasets/pull/7854
3,596,750,849
PR_kwDODunzps6x8yiy
7,854
[Distributed] split_dataset_by_node() gives the same number of examples for each node
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7854). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "Making this work with multiple workers could create a lot of communication for not a lot of benefits, considering you can simply use `Join()` to let nodes shutdown when they run out of data while the other nodes continue training: https://docs.pytorch.org/docs/stable/distributed.algorithms.join.html" ]
2025-11-06T17:14:18Z
2025-11-10T14:57:44Z
null
MEMBER
null
null
null
null
this works: ```python import torch.distributed as dist from datasets import IterableDataset from datasets.distributed import split_dataset_by_node from collections import Counter def g(shards): for shard in shards: # shards don't have the same length num_examples = 3 + shard for i in range(num_examples): yield {"shard": f"{shard=}", "i": i} if __name__ == "__main__": dist.init_process_group(backend="gloo") rank, world_size = dist.get_rank(), dist.get_world_size() num_shards = 6 ds = IterableDataset.from_generator(g, gen_kwargs={"shards": list(range(num_shards))}) ds = split_dataset_by_node(ds, rank=rank, world_size=world_size) # Check that each rank has the same number of examples # and show the number of examples per shard and per rank counter = Counter(ds["shard"]) print(f"# {rank=}\ttotal={counter.total()}\t{counter}", flush=True) # torchrun --nproc_per_node 2 script.py # rank=0 total=16 Counter({'shard=4': 7, 'shard=2': 5, 'shard=0': 4}) # rank=1 total=16 Counter({'shard=3': 6, 'shard=5': 6, 'shard=1': 4}) ``` TODO: make it work with DataLoader (communicate with main process to know when the node runs out of data ?)
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7854/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7854/timeline
null
null
1
{ "diff_url": "https://github.com/huggingface/datasets/pull/7854.diff", "html_url": "https://github.com/huggingface/datasets/pull/7854", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7854.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7854" }
true
https://api.github.com/repos/huggingface/datasets/issues/7853
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7853/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7853/comments
https://api.github.com/repos/huggingface/datasets/issues/7853/events
https://github.com/huggingface/datasets/pull/7853
3,596,232,275
PR_kwDODunzps6x7ARa
7,853
Fix embed storage nifti
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7853). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-06T15:07:58Z
2025-11-06T17:04:57Z
2025-11-06T16:20:36Z
CONTRIBUTOR
null
null
null
null
Fixes #7852 Adds `embed_storage` function and allows gzipped files to be loaded correctly from local storage.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7853/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7853/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7853.diff", "html_url": "https://github.com/huggingface/datasets/pull/7853", "merged_at": "2025-11-06T16:20:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/7853.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7853" }
true
https://api.github.com/repos/huggingface/datasets/issues/7852
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7852/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7852/comments
https://api.github.com/repos/huggingface/datasets/issues/7852/events
https://github.com/huggingface/datasets/issues/7852
3,595,450,602
I_kwDODunzps7WTjjq
7,852
Problems with NifTI
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "> 2. when uploading via the niftifolder feature, the resulting parquet only contains relative paths to the nifti files:\n\nwhat did you use to upload the dataset ? iirc push_to_hub() does upload the bytes as well, but to_parquet() doesn't", "> > 2. when uploading via the niftifolder feature, the resulting parquet only contains relative paths to the nifti files:\n> \n> what did you use to upload the dataset ? iirc push_to_hub() does upload the bytes as well, but to_parquet() doesn't\n\nI used `push_to_hub` but the problem is that the nifti feature does not have an `embed_storage` function" ]
2025-11-06T11:46:33Z
2025-11-06T16:20:38Z
2025-11-06T16:20:38Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug There are currently 2 problems with the new NifTI feature: 1. dealing with zipped files, this is mentioned and explained [here](https://github.com/huggingface/datasets/pull/7815#issuecomment-3496199503) 2. when uploading via the `niftifolder` feature, the resulting parquet only contains relative paths to the nifti files: ```bash table['nifti'] <pyarrow.lib.ChunkedArray object at 0x798245d37d60> [ -- is_valid: all not null -- child 0 type: binary [ null, null, null, null, null, null ] -- child 1 type: string [ "/home/tobias/programming/github/datasets/nifti_extracted/T1.nii", "/home/tobias/programming/github/datasets/nifti_extracted/T2-interleaved.nii", "/home/tobias/programming/github/datasets/nifti_extracted/T2.nii", "/home/tobias/programming/github/datasets/nifti_extracted/T2_-interleaved.nii", "/home/tobias/programming/github/datasets/nifti_extracted/T2_.nii", "/home/tobias/programming/github/datasets/nifti_extracted/fieldmap.nii" ] ] ``` instead of containing bytes. The code is copy pasted from PDF, so I wonder what is going wrong here. ### Steps to reproduce the bug see the linked comment ### Expected behavior downloading should work as smoothly as for pdf ### Environment info - `datasets` version: 4.4.2.dev0 - Platform: Linux-6.14.0-33-generic-x86_64-with-glibc2.39 - Python version: 3.12.3 - `huggingface_hub` version: 0.35.3 - PyArrow version: 21.0.0 - Pandas version: 2.3.3 - `fsspec` version: 2025.9.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7852/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7852/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7851
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7851/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7851/comments
https://api.github.com/repos/huggingface/datasets/issues/7851/events
https://github.com/huggingface/datasets/pull/7851
3,592,252,116
PR_kwDODunzps6xtvVj
7,851
Add fasta support
{ "avatar_url": "https://avatars.githubusercontent.com/u/209551168?v=4", "events_url": "https://api.github.com/users/georgia-hf/events{/privacy}", "followers_url": "https://api.github.com/users/georgia-hf/followers", "following_url": "https://api.github.com/users/georgia-hf/following{/other_user}", "gists_url": "https://api.github.com/users/georgia-hf/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/georgia-hf", "id": 209551168, "login": "georgia-hf", "node_id": "U_kgDODH1_QA", "organizations_url": "https://api.github.com/users/georgia-hf/orgs", "received_events_url": "https://api.github.com/users/georgia-hf/received_events", "repos_url": "https://api.github.com/users/georgia-hf/repos", "site_admin": false, "starred_url": "https://api.github.com/users/georgia-hf/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/georgia-hf/subscriptions", "type": "User", "url": "https://api.github.com/users/georgia-hf", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7851). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "A few comments:\r\n\r\n- Have you tried using this with longer sequences? @UriNeri developed something similar internally and used it with viral genomes. He got some Parquet errors due to genomes not fitting in a `utf8` column. This was fixed by using `large_utf8`.\r\n- If you're only using it to read FASTA files, I think that having BioPython as a dependency is overkill. The library is very large and the FASTA parser isn't particularly fast. I have an example of a fast parser with no external references [here](https://gist.github.com/apcamargo/d039aa04a2cbbcbb14e2d34a0963b862) (this is actually based on [`readfq.py`](https://github.com/lh3/readfq/blob/master/readfq.py), with a couple of extra functions that might not be useful in the context of this PR)", "> * If you're only using it to read FASTA files, I think that having BioPython as a dependency is overkill. The library is very large and the FASTA parser isn't particularly fast. I have an example of a fast parser with no external references [here](https://gist.github.com/apcamargo/d039aa04a2cbbcbb14e2d34a0963b862) (this is actually based on [`readfq.py`](https://github.com/lh3/readfq/blob/master/readfq.py), with a couple of extra functions that might not be useful in the context of this PR)\r\n\r\nWhat @apcamargo said, plus FWIW in **our approach** (so might not be relevant here) we use polars (with custom fasta io parser) or polars-bio (that has a `scan_fasta` function) and we foudn out that the page size sometimes need to be adjusted:\r\n```\r\nenvs/default/lib/python3.9/site-packages/polars/lazyframe/frame.py:2422, in LazyFrame.collect(self, type_coercion, predicate_pushdown, projection_pushdown, simplify_expression, slice_pushdown, comm_subplan_elim, comm_subexpr_elim, cluster_with_columns, collapse_joins, no_optimization, engine, background, optimizations, **_kwargs)\r\n 2420 # Only for testing purposes\r\n 2421 callback = _kwargs.get(\"post_opt_callback\", callback)\r\n-> 2422 return wrap_df(ldf.collect(engine, callback))\r\nComputeError: parquet: File out of specification: A page can only contain i32::MAX uncompressed bytes. This one contains 4544943557\r\n```\r\n\r\nWhich in polars can be solved with:\r\n```\r\ndf.write_parquet(\r\n \"test1.patquet\",\r\n compression=\"zstd\",\r\n row_group_size=10_000, # smaller row groups\r\n data_page_size=1024*1024 # 1MB page size\r\n)\r\n```\r\n" ]
2025-11-05T18:11:12Z
2025-11-15T00:51:53Z
null
NONE
null
null
null
null
This PR adds support for FASTA files conversion to Parquet.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7851/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7851/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7851.diff", "html_url": "https://github.com/huggingface/datasets/pull/7851", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7851.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7851" }
true
https://api.github.com/repos/huggingface/datasets/issues/7850
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7850/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7850/comments
https://api.github.com/repos/huggingface/datasets/issues/7850/events
https://github.com/huggingface/datasets/pull/7850
3,591,758,675
PR_kwDODunzps6xsGi_
7,850
dev version
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7850). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-05T16:02:23Z
2025-11-05T16:05:40Z
2025-11-05T16:02:32Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7850/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7850/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7850.diff", "html_url": "https://github.com/huggingface/datasets/pull/7850", "merged_at": "2025-11-05T16:02:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/7850.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7850" }
true
https://api.github.com/repos/huggingface/datasets/issues/7849
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7849/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7849/comments
https://api.github.com/repos/huggingface/datasets/issues/7849/events
https://github.com/huggingface/datasets/pull/7849
3,591,749,675
PR_kwDODunzps6xsEm0
7,849
release: 4.4.1
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7849). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-05T16:00:05Z
2025-11-05T16:03:06Z
2025-11-05T16:00:46Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7849/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7849/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7849.diff", "html_url": "https://github.com/huggingface/datasets/pull/7849", "merged_at": "2025-11-05T16:00:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/7849.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7849" }
true
https://api.github.com/repos/huggingface/datasets/issues/7848
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7848/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7848/comments
https://api.github.com/repos/huggingface/datasets/issues/7848/events
https://github.com/huggingface/datasets/pull/7848
3,590,024,849
PR_kwDODunzps6xmPYZ
7,848
DOC: remove mode parameter in docstring of pdf and video feature
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
2025-11-05T09:11:46Z
2025-11-05T14:42:59Z
2025-11-05T14:04:03Z
CONTRIBUTOR
null
null
null
null
closes #7841 As mentioned in the issue `mode` has been copy-pasted but isn't used in these files.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7848/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7848/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7848.diff", "html_url": "https://github.com/huggingface/datasets/pull/7848", "merged_at": "2025-11-05T14:04:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/7848.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7848" }
true
https://api.github.com/repos/huggingface/datasets/issues/7847
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7847/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7847/comments
https://api.github.com/repos/huggingface/datasets/issues/7847/events
https://github.com/huggingface/datasets/pull/7847
3,586,135,727
PR_kwDODunzps6xZZb9
7,847
Better streaming retries (504 and 429)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7847). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-04T11:23:58Z
2025-11-04T13:52:25Z
2025-11-04T13:52:22Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7847/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7847/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7847.diff", "html_url": "https://github.com/huggingface/datasets/pull/7847", "merged_at": "2025-11-04T13:52:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/7847.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7847" }
true
https://api.github.com/repos/huggingface/datasets/issues/7846
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7846/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7846/comments
https://api.github.com/repos/huggingface/datasets/issues/7846/events
https://github.com/huggingface/datasets/pull/7846
3,585,966,335
PR_kwDODunzps6xYzny
7,846
set dev version
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7846). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-04T10:44:27Z
2025-11-04T10:49:24Z
2025-11-04T10:44:37Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7846/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7846/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7846.diff", "html_url": "https://github.com/huggingface/datasets/pull/7846", "merged_at": "2025-11-04T10:44:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/7846.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7846" }
true
https://api.github.com/repos/huggingface/datasets/issues/7845
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7845/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7845/comments
https://api.github.com/repos/huggingface/datasets/issues/7845/events
https://github.com/huggingface/datasets/pull/7845
3,585,926,647
PR_kwDODunzps6xYq2r
7,845
Release: 4.4.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7845). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-04T10:35:33Z
2025-11-04T10:39:47Z
2025-11-04T10:36:37Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7845/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7845/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7845.diff", "html_url": "https://github.com/huggingface/datasets/pull/7845", "merged_at": "2025-11-04T10:36:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/7845.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7845" }
true
https://api.github.com/repos/huggingface/datasets/issues/7844
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7844/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7844/comments
https://api.github.com/repos/huggingface/datasets/issues/7844/events
https://github.com/huggingface/datasets/pull/7844
3,582,354,507
PR_kwDODunzps6xM9hd
7,844
support fsspec 2025.10.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7844). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-03T14:34:29Z
2025-11-03T14:51:33Z
2025-11-03T14:51:32Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7844/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7844/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7844.diff", "html_url": "https://github.com/huggingface/datasets/pull/7844", "merged_at": "2025-11-03T14:51:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/7844.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7844" }
true
https://api.github.com/repos/huggingface/datasets/issues/7843
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7843/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7843/comments
https://api.github.com/repos/huggingface/datasets/issues/7843/events
https://github.com/huggingface/datasets/pull/7843
3,582,311,403
PR_kwDODunzps6xM0sq
7,843
fix column with transform
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7843). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-03T14:23:01Z
2025-11-03T14:34:15Z
2025-11-03T14:34:12Z
MEMBER
null
null
null
null
fix https://github.com/huggingface/datasets/issues/7842
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7843/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7843/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7843.diff", "html_url": "https://github.com/huggingface/datasets/pull/7843", "merged_at": "2025-11-03T14:34:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/7843.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7843" }
true
https://api.github.com/repos/huggingface/datasets/issues/7842
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7842/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7842/comments
https://api.github.com/repos/huggingface/datasets/issues/7842/events
https://github.com/huggingface/datasets/issues/7842
3,582,182,995
I_kwDODunzps7Vg8ZT
7,842
Transform with columns parameter triggers on non-specified column access
{ "avatar_url": "https://avatars.githubusercontent.com/u/18426892?v=4", "events_url": "https://api.github.com/users/mr-brobot/events{/privacy}", "followers_url": "https://api.github.com/users/mr-brobot/followers", "following_url": "https://api.github.com/users/mr-brobot/following{/other_user}", "gists_url": "https://api.github.com/users/mr-brobot/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mr-brobot", "id": 18426892, "login": "mr-brobot", "node_id": "MDQ6VXNlcjE4NDI2ODky", "organizations_url": "https://api.github.com/users/mr-brobot/orgs", "received_events_url": "https://api.github.com/users/mr-brobot/received_events", "repos_url": "https://api.github.com/users/mr-brobot/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mr-brobot/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mr-brobot/subscriptions", "type": "User", "url": "https://api.github.com/users/mr-brobot", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
2025-11-03T13:55:27Z
2025-11-03T14:34:13Z
2025-11-03T14:34:13Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug Iterating over a [`Column`](https://github.com/huggingface/datasets/blob/8b1bd4ec1cc9e9ce022f749abb6485ef984ae7c0/src/datasets/arrow_dataset.py#L633-L692) iterates through the parent [`Dataset`](https://github.com/huggingface/datasets/blob/8b1bd4ec1cc9e9ce022f749abb6485ef984ae7c0/src/datasets/arrow_dataset.py#L695) and applies all formatting/transforms on each row, regardless of which column is being accessed. This causes an error when transforms depend on columns not present in the projection. ### Steps to reproduce the bug ### Load a dataset with multiple columns ```python ds = load_dataset("mrbrobot/isic-2024", split="train") ``` ### Define a transform that specifies an input column ```python def image_transform(batch): batch["image"] = batch["image"] # KeyError when batch doesn't contain "image" return batch # apply transform only to image column ds = ds.with_format("torch") ds = ds.with_transform(image_transform, columns=["image"], output_all_columns=True) ``` ### Iterate over non-specified column ```python # iterate over a different column, triggers the transform on each row, but batch doesn't contain "image" for t in ds["target"]: # KeyError: 'image' print(t) ``` ### Expected behavior If a user iterates over `ds["target"]` and the transform specifies `columns=["image"]`, the transform should be skipped. ### Environment info `datasets`: 4.2.0 Python: 3.12.12 Linux: Debian 11.11
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7842/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7842/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7841
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7841/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7841/comments
https://api.github.com/repos/huggingface/datasets/issues/7841/events
https://github.com/huggingface/datasets/issues/7841
3,579,506,747
I_kwDODunzps7VWvA7
7,841
DOC: `mode` parameter on pdf and video features unused
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "They seem to be artefacts from a copy-paste of the Image feature ^^' we should remove them" ]
2025-11-02T12:37:47Z
2025-11-05T14:04:04Z
2025-11-05T14:04:04Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Following up on https://github.com/huggingface/datasets/pull/7840 I asked claude code to check for undocumented parameters for other features and it found: - mode parameter on video is documented but unused: https://github.com/huggingface/datasets/blob/main/src/datasets/features/video.py#L48-L49 - the same goes for the mode parameter on the pdf feature: https://github.com/huggingface/datasets/blob/main/src/datasets/features/pdf.py#L47-L48 I assume checking if these modes can be supported and otherwise removing them is the way to go here.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7841/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7841/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7840
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7840/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7840/comments
https://api.github.com/repos/huggingface/datasets/issues/7840/events
https://github.com/huggingface/datasets/pull/7840
3,579,486,348
PR_kwDODunzps6xDsbG
7,840
Add num channels to audio
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7840). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-11-02T12:10:24Z
2025-11-03T17:37:48Z
2025-11-03T14:24:11Z
CONTRIBUTOR
null
null
null
null
Fixes #7837 We currently have the [mono attribute for Audio documented](https://github.com/huggingface/datasets/blob/41c05299348a499807432ab476e1cdc4143c8772/src/datasets/features/audio.py#L52C1-L54C22) but not used anywhere resulting in confusion for users. Since torchcodec does not know this attribute I suggest using `num_channels` (currently supported `None` (leave unchanged), mono: `1`, stereo: `2`). I could also add a mono attribute but found that to be more confusing for developers and would restrict us if at any point in the future more than 2 channels are supported.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7840/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7840/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7840.diff", "html_url": "https://github.com/huggingface/datasets/pull/7840", "merged_at": "2025-11-03T14:24:11Z", "patch_url": "https://github.com/huggingface/datasets/pull/7840.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7840" }
true
https://api.github.com/repos/huggingface/datasets/issues/7839
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7839/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7839/comments
https://api.github.com/repos/huggingface/datasets/issues/7839/events
https://github.com/huggingface/datasets/issues/7839
3,579,121,843
I_kwDODunzps7VVRCz
7,839
datasets doesn't work with python 3.14
{ "avatar_url": "https://avatars.githubusercontent.com/u/4789087?v=4", "events_url": "https://api.github.com/users/zachmoshe/events{/privacy}", "followers_url": "https://api.github.com/users/zachmoshe/followers", "following_url": "https://api.github.com/users/zachmoshe/following{/other_user}", "gists_url": "https://api.github.com/users/zachmoshe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zachmoshe", "id": 4789087, "login": "zachmoshe", "node_id": "MDQ6VXNlcjQ3ODkwODc=", "organizations_url": "https://api.github.com/users/zachmoshe/orgs", "received_events_url": "https://api.github.com/users/zachmoshe/received_events", "repos_url": "https://api.github.com/users/zachmoshe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zachmoshe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zachmoshe/subscriptions", "type": "User", "url": "https://api.github.com/users/zachmoshe", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "Thanks for the report.\nHave you tried on main? This should work, there was recently a PR merged to address this problem, see #7817", "Works on main 👍 \nWhat's the release schedule for `datasets`? Seems like a cadence of ~2weeks so I assume a real version is due pretty soon?", "let's say we do a new release later today ? :)", "Premium service! \n😂 👑 \nJust checked 4.4.0 - works as expected!" ]
2025-11-02T09:09:06Z
2025-11-04T14:02:25Z
2025-11-04T14:02:25Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug Seems that `dataset` doesn't work with python==3.14. The root cause seems to be something with a `deel` API that was changed. ``` TypeError: Pickler._batch_setitems() takes 2 positional arguments but 3 were given ``` ### Steps to reproduce the bug (on a new folder) uv init uv python pin 3.14 uv add datasets uv run python (in REPL) import datasets datasets.load_dataset("cais/mmlu", "all") # will fail on any dataset ``` >>> datasets.load_dataset("cais/mmlu", "all") Traceback (most recent call last): File "<python-input-2>", line 1, in <module> datasets.load_dataset("cais/mmlu", "all") ~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^ File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/datasets/load.py", line 1397, in load_dataset builder_instance = load_dataset_builder( path=path, ...<10 lines>... **config_kwargs, ) File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/datasets/load.py", line 1185, in load_dataset_builder builder_instance._use_legacy_cache_dir_if_possible(dataset_module) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^ File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/datasets/builder.py", line 615, in _use_legacy_cache_dir_if_possible self._check_legacy_cache2(dataset_module) or self._check_legacy_cache() or None ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^ File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/datasets/builder.py", line 487, in _check_legacy_cache2 config_id = self.config.name + "-" + Hasher.hash({"data_files": self.config.data_files}) ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/datasets/fingerprint.py", line 188, in hash return cls.hash_bytes(dumps(value)) ~~~~~^^^^^^^ File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/datasets/utils/_dill.py", line 120, in dumps dump(obj, file) ~~~~^^^^^^^^^^^ File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/datasets/utils/_dill.py", line 114, in dump Pickler(file, recurse=True).dump(obj) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^ File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/dill/_dill.py", line 428, in dump StockPickler.dump(self, obj) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^ File "/Users/zmoshe/.local/uv/python/cpython-3.14.0rc2-macos-aarch64-none/lib/python3.14/pickle.py", line 498, in dump self.save(obj) ~~~~~~~~~^^^^^ File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/datasets/utils/_dill.py", line 70, in save dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/dill/_dill.py", line 422, in save StockPickler.save(self, obj, save_persistent_id) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/zmoshe/.local/uv/python/cpython-3.14.0rc2-macos-aarch64-none/lib/python3.14/pickle.py", line 572, in save f(self, obj) # Call unbound method with explicit self ~^^^^^^^^^^^ File "/Users/zmoshe/temp/test_datasets_py3.14/.venv/lib/python3.14/site-packages/dill/_dill.py", line 1262, in save_module_dict StockPickler.save_dict(pickler, obj) ~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^ File "/Users/zmoshe/.local/uv/python/cpython-3.14.0rc2-macos-aarch64-none/lib/python3.14/pickle.py", line 1064, in save_dict self._batch_setitems(obj.items(), obj) ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^ TypeError: Pickler._batch_setitems() takes 2 positional arguments but 3 were given ``` ### Expected behavior should work. ### Environment info datasets==v4.3.0 python==3.14
{ "avatar_url": "https://avatars.githubusercontent.com/u/4789087?v=4", "events_url": "https://api.github.com/users/zachmoshe/events{/privacy}", "followers_url": "https://api.github.com/users/zachmoshe/followers", "following_url": "https://api.github.com/users/zachmoshe/following{/other_user}", "gists_url": "https://api.github.com/users/zachmoshe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zachmoshe", "id": 4789087, "login": "zachmoshe", "node_id": "MDQ6VXNlcjQ3ODkwODc=", "organizations_url": "https://api.github.com/users/zachmoshe/orgs", "received_events_url": "https://api.github.com/users/zachmoshe/received_events", "repos_url": "https://api.github.com/users/zachmoshe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zachmoshe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zachmoshe/subscriptions", "type": "User", "url": "https://api.github.com/users/zachmoshe", "user_view_type": "public" }
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7839/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7839/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7837
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7837/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7837/comments
https://api.github.com/repos/huggingface/datasets/issues/7837/events
https://github.com/huggingface/datasets/issues/7837
3,575,454,726
I_kwDODunzps7VHRwG
7,837
mono parameter to the Audio feature is missing
{ "avatar_url": "https://avatars.githubusercontent.com/u/1250234?v=4", "events_url": "https://api.github.com/users/ernestum/events{/privacy}", "followers_url": "https://api.github.com/users/ernestum/followers", "following_url": "https://api.github.com/users/ernestum/following{/other_user}", "gists_url": "https://api.github.com/users/ernestum/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ernestum", "id": 1250234, "login": "ernestum", "node_id": "MDQ6VXNlcjEyNTAyMzQ=", "organizations_url": "https://api.github.com/users/ernestum/orgs", "received_events_url": "https://api.github.com/users/ernestum/received_events", "repos_url": "https://api.github.com/users/ernestum/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ernestum/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ernestum/subscriptions", "type": "User", "url": "https://api.github.com/users/ernestum", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "Hey, we removed the misleading passage in the docstring and enabled support for `num_channels` as torchcodec does", "thanks!" ]
2025-10-31T15:41:39Z
2025-11-03T15:59:18Z
2025-11-03T14:24:12Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
According to the docs, there is a "mono" parameter to the Audio feature, which turns any stereo into mono. In practice the signal is not touched and the mono parameter, even though documented, does not exist. https://github.com/huggingface/datasets/blob/41c05299348a499807432ab476e1cdc4143c8772/src/datasets/features/audio.py#L52C1-L54C22
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7837/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7837/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7836
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7836/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7836/comments
https://api.github.com/repos/huggingface/datasets/issues/7836/events
https://github.com/huggingface/datasets/pull/7836
3,562,316,362
PR_kwDODunzps6wLuh9
7,836
Python 3.14
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7836). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-28T16:11:13Z
2025-10-31T17:27:17Z
2025-10-31T17:27:15Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7836/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7836/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7836.diff", "html_url": "https://github.com/huggingface/datasets/pull/7836", "merged_at": "2025-10-31T17:27:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/7836.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7836" }
true
https://api.github.com/repos/huggingface/datasets/issues/7835
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7835/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7835/comments
https://api.github.com/repos/huggingface/datasets/issues/7835/events
https://github.com/huggingface/datasets/pull/7835
3,560,909,796
PR_kwDODunzps6wHK9e
7,835
Add DICOM support
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Awesome ! For the docs should we rename https://huggingface.co/docs/datasets/nifti_dataset to medical_imaging_dataset and have both DICOM and NIfTI together or have separate pages in you opinion ?", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7835). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "> Awesome ! For the docs should we rename https://huggingface.co/docs/datasets/nifti_dataset to medical_imaging_dataset and have both DICOM and NIfTI together or have separate pages in you opinion ?\r\n\r\nMakes sense, is more intuitive for the user and the pages as proposed in this branch have a lot of overlap. I would then structure it in such a way to write some brief things about medical imaging, then introduce the formats (so basically concatenating the two pages together and removing duplicates).", "Pls don't merge currently, since we'll need an `embed_storage` function in here as well. See\r\nhttps://github.com/huggingface/datasets/pull/7815#issuecomment-3494094692 and the following conversation", "@lhoestq, this is ready for a first round of review." ]
2025-10-28T10:41:05Z
2025-11-26T15:25:03Z
null
CONTRIBUTOR
null
null
null
null
supports #7804 Add support for the dicom file format. This PR follows PR #7815 and PR #7325 closely. Remarkable differences: I made sure that we can load all of pydicom's test data, and encountered the `force=True` parameter that we explicitly support here. This allows to trying to load corrupted dicom files, we explicitly test this! There is one dataset with all of dicom's test data on huggingface which can be loaded using this branch with the following script: ```python from datasets import load_dataset from datasets import Features, ClassLabel from datasets.features import Dicom features = Features({ "dicom": Dicom(force=True), # necessary to be able to load one corrupted file "label": ClassLabel(num_classes=2) }) ds = load_dataset("TobiasPitters/dicom-sample-dataset", features=features) error_count = 0 for idx, item in enumerate(ds["test"]): dicom = item["dicom"] try: print(f"Type: {type(dicom)}") if hasattr(dicom, 'PatientID'): print(f"PatientID: {dicom.PatientID}") if hasattr(dicom, 'StudyInstanceUID'): print(f"StudyInstanceUID: {dicom.StudyInstanceUID}") if hasattr(dicom, 'Modality'): print(f"Modality: {dicom.Modality}") except Exception as e: error_count += 1 print(e) print(f"Finished processing with {error_count} errors.") ``` todo: - [x] add docs (will do so soon)
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 2, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/7835/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7835/timeline
null
null
1
{ "diff_url": "https://github.com/huggingface/datasets/pull/7835.diff", "html_url": "https://github.com/huggingface/datasets/pull/7835", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7835.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7835" }
true
https://api.github.com/repos/huggingface/datasets/issues/7834
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7834/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7834/comments
https://api.github.com/repos/huggingface/datasets/issues/7834/events
https://github.com/huggingface/datasets/issues/7834
3,558,802,959
I_kwDODunzps7UHwYP
7,834
Audio.cast_column() or Audio.decode_example() causes Colab kernel crash (std::bad_alloc)
{ "avatar_url": "https://avatars.githubusercontent.com/u/2559570?v=4", "events_url": "https://api.github.com/users/rachidio/events{/privacy}", "followers_url": "https://api.github.com/users/rachidio/followers", "following_url": "https://api.github.com/users/rachidio/following{/other_user}", "gists_url": "https://api.github.com/users/rachidio/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/rachidio", "id": 2559570, "login": "rachidio", "node_id": "MDQ6VXNlcjI1NTk1NzA=", "organizations_url": "https://api.github.com/users/rachidio/orgs", "received_events_url": "https://api.github.com/users/rachidio/received_events", "repos_url": "https://api.github.com/users/rachidio/repos", "site_admin": false, "starred_url": "https://api.github.com/users/rachidio/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rachidio/subscriptions", "type": "User", "url": "https://api.github.com/users/rachidio", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Hi ! `datasets` v4 uses `torchcodec` for audio decoding (previous versions were using `soundfile`). What is your `torchcodec` version ? Can you try other versions of `torchcodec` and see if it works ?", "When I install `datasets` with `pip install datasets[audio]` it install this version of `torchcodec`:\n```\nName: torchcodec\nVersion: 0.8.1\n```\nCan you please point to a working version of `torchcodec`?\n\nThanks for your help", "I believe you simply need to make sure the torchcodec and torch versions work together. Here is how to fix it:\n\n```python\n!pip install -U torchcodec torch\n```", "I am also encountering this same issue when i run `print(ug_court[\"train\"][0])` to view the features of the first row of my audio data", "the problem still goes on to when i force training with seeing these features", "Thank you @lhoestq I've reinstalled the packages an the error is gone.\nMy new versions are:\n```\nName: torch\nVersion: 2.8.0\n---\nName: torchaudio\nVersion: 2.8.0\n---\nName: torchcodec\nVersion: 0.8.1\n```\n\nRegards", "mine too has worked ", "Hi,\n\nI encounter the same problem when trying to inspect the first element in the dataset. My environment is:\n```\nroot@3ac6f9f8c6c4:/workspace# pip3 list | grep torch\npytorch-lightning 2.5.6\npytorch-metric-learning 2.9.0\ntorch 2.8.0+cu126\ntorch-audiomentations 0.12.0\ntorch_pitch_shift 1.2.5\ntorchaudio 2.8.0+cu126\ntorchcodec 0.8.1\ntorchelastic 0.2.2\ntorchmetrics 1.8.2\ntorchvision 0.23.0+cu126\n```\nthe same as @rachidio 's new version that works.\n\nI am in a Docker container environment, and here is the code I am working with:\n\n<img width=\"1350\" height=\"388\" alt=\"Image\" src=\"https://github.com/user-attachments/assets/4cf0400f-9ee7-47c7-ba57-c4ef3c1e7fd6\" />" ]
2025-10-27T22:02:00Z
2025-11-15T16:28:04Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug When using the huggingface datasets.Audio feature to decode a local or remote (public HF dataset) audio file inside Google Colab, the notebook kernel crashes with std::bad_alloc (C++ memory allocation failure). The crash happens even with a minimal code example and valid .wav file that can be read successfully using soundfile. Here is a sample Collab notebook to reproduce the problem. https://colab.research.google.com/drive/1nnb-GC5748Tux3xcYRussCGp2x-zM9Id?usp=sharing code sample: ``` ... audio_dataset = audio_dataset.cast_column("audio", Audio(sampling_rate=16000)) # Accessing the first element crashes the Colab kernel print(audio_dataset[0]["audio"]) ``` Error log ``` WARNING what(): std::bad_alloc terminate called after throwing an instance of 'std::bad_alloc' ``` Environment Platform: Google Colab (Python 3.12.12) datasets Version: 4.3.0 soundfile Version: 0.13.1 torchaudio Version: 2.8.0+cu126 Thanks in advance to help me on this error I get approx two weeks now after it was working before. Regards ### Steps to reproduce the bug https://colab.research.google.com/drive/1nnb-GC5748Tux3xcYRussCGp2x-zM9Id?usp=sharing ### Expected behavior Loading the audio and decode it. It should safely return: { "path": "path/filaname.wav", "array": np.ndarray([...]), "sampling_rate": 16000 } ### Environment info Environment Platform: Google Colab (Python 3.12.12) datasets Version: 4.3.0 soundfile Version: 0.13.1 torchaudio Version: 2.8.0+cu126
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 1, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7834/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7834/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7833
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7833/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7833/comments
https://api.github.com/repos/huggingface/datasets/issues/7833/events
https://github.com/huggingface/datasets/pull/7833
3,556,014,911
PR_kwDODunzps6v2gAI
7,833
Fix: Properly render [!TIP] block in stream.shuffle documentation
{ "avatar_url": "https://avatars.githubusercontent.com/u/110672812?v=4", "events_url": "https://api.github.com/users/art-test-stack/events{/privacy}", "followers_url": "https://api.github.com/users/art-test-stack/followers", "following_url": "https://api.github.com/users/art-test-stack/following{/other_user}", "gists_url": "https://api.github.com/users/art-test-stack/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/art-test-stack", "id": 110672812, "login": "art-test-stack", "node_id": "U_kgDOBpi7rA", "organizations_url": "https://api.github.com/users/art-test-stack/orgs", "received_events_url": "https://api.github.com/users/art-test-stack/received_events", "repos_url": "https://api.github.com/users/art-test-stack/repos", "site_admin": false, "starred_url": "https://api.github.com/users/art-test-stack/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/art-test-stack/subscriptions", "type": "User", "url": "https://api.github.com/users/art-test-stack", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
2025-10-27T10:09:31Z
2025-10-28T15:57:33Z
2025-10-28T15:57:33Z
CONTRIBUTOR
null
null
null
null
The second line starting with the bracket doesn't properly render on huggingface/docs. Added "> " to address it. In the client documentation, the markdown 'TIP' paragraph in docs/stream#shuffle is not well executed, not as the other in the same page / while markdown is correctly considering it. Documentation: https://huggingface.co/docs/datasets/v4.3.0/en/stream#shuffle:~:text=%5B!TIP%5D%5BIterableDataset.shuffle()%5D(/docs/datasets/v4.3.0/en/package_reference/main_classes%23datasets.IterableDataset.shuffle)%20will%20also%20shuffle%20the%20order%20of%20the%20shards%20if%20the%20dataset%20is%20sharded%20into%20multiple%20files. Github source: https://github.com/huggingface/datasets/blob/main/docs/source/stream.mdx#:~:text=Casting%20only%20works%20if%20the%20original%20feature%20type%20and%20new%20feature%20type%20are%20compatible.%20For%20example%2C%20you%20can%20cast%20a%20column%20with%20the%20feature%20type%20Value(%27int32%27)%20to%20Value(%27bool%27)%20if%20the%20original%20column%20only%20contains%20ones%20and%20zeros.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7833/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7833/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7833.diff", "html_url": "https://github.com/huggingface/datasets/pull/7833", "merged_at": "2025-10-28T15:57:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/7833.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7833" }
true
https://api.github.com/repos/huggingface/datasets/issues/7832
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7832/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7832/comments
https://api.github.com/repos/huggingface/datasets/issues/7832/events
https://github.com/huggingface/datasets/issues/7832
3,555,991,552
I_kwDODunzps7T9CAA
7,832
[DOCS][minor] TIPS paragraph not compiled in docs/stream
{ "avatar_url": "https://avatars.githubusercontent.com/u/110672812?v=4", "events_url": "https://api.github.com/users/art-test-stack/events{/privacy}", "followers_url": "https://api.github.com/users/art-test-stack/followers", "following_url": "https://api.github.com/users/art-test-stack/following{/other_user}", "gists_url": "https://api.github.com/users/art-test-stack/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/art-test-stack", "id": 110672812, "login": "art-test-stack", "node_id": "U_kgDOBpi7rA", "organizations_url": "https://api.github.com/users/art-test-stack/orgs", "received_events_url": "https://api.github.com/users/art-test-stack/received_events", "repos_url": "https://api.github.com/users/art-test-stack/repos", "site_admin": false, "starred_url": "https://api.github.com/users/art-test-stack/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/art-test-stack/subscriptions", "type": "User", "url": "https://api.github.com/users/art-test-stack", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
2025-10-27T10:03:22Z
2025-10-27T10:10:54Z
2025-10-27T10:10:54Z
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
In the client documentation, the markdown 'TIP' paragraph for paragraph in docs/stream#shuffle is not well executed — not as the other in the same page / while markdown is correctly considering it. Documentation: https://huggingface.co/docs/datasets/v4.3.0/en/stream#shuffle:~:text=%5B!TIP%5D%5BIterableDataset.shuffle()%5D(/docs/datasets/v4.3.0/en/package_reference/main_classes%23datasets.IterableDataset.shuffle)%20will%20also%20shuffle%20the%20order%20of%20the%20shards%20if%20the%20dataset%20is%20sharded%20into%20multiple%20files. Github source: https://github.com/huggingface/datasets/blob/main/docs/source/stream.mdx#:~:text=Casting%20only%20works%20if%20the%20original%20feature%20type%20and%20new%20feature%20type%20are%20compatible.%20For%20example%2C%20you%20can%20cast%20a%20column%20with%20the%20feature%20type%20Value(%27int32%27)%20to%20Value(%27bool%27)%20if%20the%20original%20column%20only%20contains%20ones%20and%20zeros.
{ "avatar_url": "https://avatars.githubusercontent.com/u/110672812?v=4", "events_url": "https://api.github.com/users/art-test-stack/events{/privacy}", "followers_url": "https://api.github.com/users/art-test-stack/followers", "following_url": "https://api.github.com/users/art-test-stack/following{/other_user}", "gists_url": "https://api.github.com/users/art-test-stack/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/art-test-stack", "id": 110672812, "login": "art-test-stack", "node_id": "U_kgDOBpi7rA", "organizations_url": "https://api.github.com/users/art-test-stack/orgs", "received_events_url": "https://api.github.com/users/art-test-stack/received_events", "repos_url": "https://api.github.com/users/art-test-stack/repos", "site_admin": false, "starred_url": "https://api.github.com/users/art-test-stack/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/art-test-stack/subscriptions", "type": "User", "url": "https://api.github.com/users/art-test-stack", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7832/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7832/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7831
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7831/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7831/comments
https://api.github.com/repos/huggingface/datasets/issues/7831/events
https://github.com/huggingface/datasets/pull/7831
3,552,081,727
PR_kwDODunzps6vp0pz
7,831
resolves the ValueError: Unable to avoid copy while creating an array
{ "avatar_url": "https://avatars.githubusercontent.com/u/142811259?v=4", "events_url": "https://api.github.com/users/ArjunJagdale/events{/privacy}", "followers_url": "https://api.github.com/users/ArjunJagdale/followers", "following_url": "https://api.github.com/users/ArjunJagdale/following{/other_user}", "gists_url": "https://api.github.com/users/ArjunJagdale/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArjunJagdale", "id": 142811259, "login": "ArjunJagdale", "node_id": "U_kgDOCIMgew", "organizations_url": "https://api.github.com/users/ArjunJagdale/orgs", "received_events_url": "https://api.github.com/users/ArjunJagdale/received_events", "repos_url": "https://api.github.com/users/ArjunJagdale/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArjunJagdale/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArjunJagdale/subscriptions", "type": "User", "url": "https://api.github.com/users/ArjunJagdale", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "## Also i have done some tests on real dataset\r\n\r\n### Tested with real datasets:\r\n- ✅ IMDB dataset with NumPy 1.26.4 and 2.3.4\r\n- ✅ Rotten Tomatoes dataset with NumPy 1.26.4 and 2.3.4\r\n- ✅ Artificial datasets with ClassLabel features\r\n\r\n### Results:\r\n- Stratified splits work correctly in both NumPy versions\r\n- Class balance is perfectly maintained (e.g., Rotten Tomatoes: 426:426 train, 107:107 test)\r\n- Non-stratified splits continue to work as expected\r\n- Backward compatibility with NumPy 1.x confirmed\r\n\r\nBelow are the RAW logs of testing - \r\n\r\n```\r\n(venv) F:\\Python\\Machine learning\\datasets>pip install \"numpy<2.0\"\r\nCollecting numpy<2.0\r\n Using cached numpy-1.26.4-cp311-cp311-win_amd64.whl.metadata (61 kB)\r\nUsing cached numpy-1.26.4-cp311-cp311-win_amd64.whl (15.8 MB)\r\nInstalling collected packages: numpy\r\n Attempting uninstall: numpy\r\n Found existing installation: numpy 2.3.4\r\n Uninstalling numpy-2.3.4:\r\n Successfully uninstalled numpy-2.3.4\r\nSuccessfully installed numpy-1.26.4\r\n\r\n[notice] A new release of pip is available: 24.0 -> 25.3\r\n[notice] To update, run: python.exe -m pip install --upgrade pip\r\n\r\n(venv) F:\\Python\\Machine learning\\datasets>python test_fix.py\r\nNumPy version: 1.26.4\r\n============================================================\r\n\r\n[Test 1] Testing with IMDB dataset...\r\nREADME.md: 7.81kB [00:00, 7.78MB/s]\r\nF:\\Python\\Machine learning\\datasets\\venv\\Lib\\site-packages\\huggingface_hub\\file_download.py:143: UserWarning: `huggingface_hub` cache-system uses symlinks by default to efficiently store duplicated files but your machine does not support them in C:\\Users\\arjun\\.cache\\huggingface\\hub\\datasets--imdb. Caching files will still work but in a degraded version that might require more space on your disk. This warning can be disabled by setting the `HF_HUB_DISABLE_SYMLINKS_WARNING` environment variable. For more details, see https://huggingface.co/docs/huggingface_hub/how-to-cache#limitations.\r\nTo support symlinks on Windows, you either need to activate Developer Mode or to run Python as an administrator. In order to activate developer mode, see this article: https://docs.microsoft.com/en-us/windows/apps/get-started/enable-your-device-for-development\r\n warnings.warn(message)\r\nXet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`\r\ntrain-00000-of-00001.parquet: 100%|███████████████████████████████████████████████████████████████████████████████████| 21.0M/21.0M [00:06<00:00, 3.08MB/s]\r\nXet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`\r\ntest-00000-of-00001.parquet: 100%|████████████████████████████████████████████████████████████████████████████████████| 20.5M/20.5M [00:07<00:00, 2.56MB/s]\r\nXet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`\r\nunsupervised-00000-of-00001.parquet: 100%|████████████████████████████████████████████████████████████████████████████| 42.0M/42.0M [00:14<00:00, 2.84MB/s]\r\nGenerating train split: 100%|█████████████████████████████████████████████████████████████████████████████| 25000/25000 [00:00<00:00, 267643.40 examples/s]\r\nGenerating test split: 100%|██████████████████████████████████████████████████████████████████████████████| 25000/25000 [00:00<00:00, 324697.85 examples/s]\r\nGenerating unsupervised split: 100%|██████████████████████████████████████████████████████████████████████| 50000/50000 [00:00<00:00, 289202.11 examples/s]\r\nLoaded 1000 samples\r\n✅ IMDB SUCCESS!\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 800\r\n })\r\n test: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 200\r\n })\r\n})\r\nTrain class distribution: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0]\r\nTest class distribution: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0]\r\n\r\n============================================================\r\n\r\n[Test 2] Testing with Rotten Tomatoes dataset...\r\nREADME.md: 7.46kB [00:00, ?B/s]\r\nF:\\Python\\Machine learning\\datasets\\venv\\Lib\\site-packages\\huggingface_hub\\file_download.py:143: UserWarning: `huggingface_hub` cache-system uses symlinks by default to efficiently store duplicated files but your machine does not support them in C:\\Users\\arjun\\.cache\\huggingface\\hub\\datasets--rotten_tomatoes. Caching files will still work but in a degraded version that might require more space on your disk. This warning can be disabled by setting the `HF_HUB_DISABLE_SYMLINKS_WARNING` environment variable. For more details, see https://huggingface.co/docs/huggingface_hub/how-to-cache#limitations.\r\nTo support symlinks on Windows, you either need to activate Developer Mode or to run Python as an administrator. In order to activate developer mode, see this article: https://docs.microsoft.com/en-us/windows/apps/get-started/enable-your-device-for-development\r\n warnings.warn(message)\r\nXet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`\r\ntrain.parquet: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████| 699k/699k [00:00<00:00, 3.46MB/s]\r\nXet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`\r\nvalidation.parquet: 100%|█████████████████████████████████████████████████████████████████████████████████████████████| 90.0k/90.0k [00:00<00:00, 6.80MB/s]\r\nXet Storage is enabled for this repo, but the 'hf_xet' package is not installed. Falling back to regular HTTP download. For better performance, install the package with: `pip install huggingface_hub[hf_xet]` or `pip install hf_xet`\r\ntest.parquet: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████| 92.2k/92.2k [00:00<00:00, 5.85MB/s]\r\nGenerating train split: 100%|███████████████████████████████████████████████████████████████████████████████| 8530/8530 [00:00<00:00, 856082.82 examples/s]\r\nGenerating validation split: 100%|██████████████████████████████████████████████████████████████████████████| 1066/1066 [00:00<00:00, 531075.91 examples/s]\r\nGenerating test split: 100%|████████████████████████████████████████████████████████████████████████████████████████████| 1066/1066 [00:00<?, ? examples/s]\r\nLoaded 1066 samples\r\n✅ Rotten Tomatoes SUCCESS!\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 852\r\n })\r\n test: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 214\r\n })\r\n})\r\n\r\nTrain: class_0=426, class_1=426\r\nTest: class_0=107, class_1=107\r\n\r\n============================================================\r\n\r\n[Test 3] Testing without stratification (sanity check)...\r\n✅ Non-stratified split SUCCESS!\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 80\r\n })\r\n test: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 20\r\n })\r\n})\r\n\r\n============================================================\r\nAll tests completed!\r\n\r\n```\r\n\r\n### Upgrading numpy for >= 2\r\n\r\n```\r\n(venv) F:\\Python\\Machine learning\\datasets>pip install \"numpy>=2.0\"\r\nCollecting numpy>=2.0\r\n Using cached numpy-2.3.4-cp311-cp311-win_amd64.whl.metadata (60 kB)\r\nUsing cached numpy-2.3.4-cp311-cp311-win_amd64.whl (13.1 MB)\r\nInstalling collected packages: numpy\r\n Attempting uninstall: numpy\r\n Found existing installation: numpy 1.26.4\r\n Uninstalling numpy-1.26.4:\r\n Successfully uninstalled numpy-1.26.4\r\nSuccessfully installed numpy-2.3.4\r\n\r\n[notice] A new release of pip is available: 24.0 -> 25.3\r\n[notice] To update, run: python.exe -m pip install --upgrade pip\r\n\r\n(venv) F:\\Python\\Machine learning\\datasets>python test_fix.py\r\nNumPy version: 2.3.4\r\n============================================================\r\n\r\n[Test 1] Testing with IMDB dataset...\r\nLoaded 1000 samples\r\n✅ IMDB SUCCESS!\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 800\r\n })\r\n test: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 200\r\n })\r\n})\r\nTrain class distribution: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0]\r\nTest class distribution: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0]\r\n\r\n============================================================\r\n\r\n[Test 2] Testing with Rotten Tomatoes dataset...\r\nLoaded 1066 samples\r\n✅ Rotten Tomatoes SUCCESS!\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 852\r\n })\r\n test: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 214\r\n })\r\n})\r\n\r\nTrain: class_0=426, class_1=426\r\nTest: class_0=107, class_1=107\r\n\r\n============================================================\r\n\r\n[Test 3] Testing without stratification (sanity check)...\r\n✅ Non-stratified split SUCCESS!\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 80\r\n })\r\n test: Dataset({\r\n features: ['text', 'label'],\r\n num_rows: 20\r\n })\r\n})\r\n\r\n============================================================\r\nAll tests completed!\r\n", "[test_fix.py](https://github.com/user-attachments/files/23140682/test_fix.py)\r\n\r\nhere is the file I used for testing @lhoestq ", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7831). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-25T08:13:54Z
2025-10-29T06:35:32Z
2025-10-28T16:10:43Z
CONTRIBUTOR
null
null
null
null
## Summary Fixes #7818 This PR resolves the `ValueError: Unable to avoid copy while creating an array` error that occurs when using `train_test_split` with `stratify_by_column` parameter in NumPy 2.0+. ## Changes - Wrapped the stratify column array access with `np.asarray()` in `arrow_dataset.py` - This allows NumPy 2.0 to make a copy when the Arrow array is non-contiguous in memory ## Testing - ✅ Tested with NumPy 2.3.4 - stratified splits work correctly - ✅ Tested with NumPy 1.26.4 - backward compatibility maintained - ✅ Verified class balance is preserved in stratified splits - ✅ Non-stratified splits continue to work as expected
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7831/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7831/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7831.diff", "html_url": "https://github.com/huggingface/datasets/pull/7831", "merged_at": "2025-10-28T16:10:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/7831.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7831" }
true
https://api.github.com/repos/huggingface/datasets/issues/7830
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7830/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7830/comments
https://api.github.com/repos/huggingface/datasets/issues/7830/events
https://github.com/huggingface/datasets/pull/7830
3,549,769,999
PR_kwDODunzps6viIQf
7,830
fix ci compressionfs
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7830). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-24T14:29:00Z
2025-10-24T14:46:34Z
2025-10-24T14:46:29Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7830/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7830/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7830.diff", "html_url": "https://github.com/huggingface/datasets/pull/7830", "merged_at": "2025-10-24T14:46:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/7830.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7830" }
true
https://api.github.com/repos/huggingface/datasets/issues/7829
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7829/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7829/comments
https://api.github.com/repos/huggingface/datasets/issues/7829/events
https://github.com/huggingface/datasets/issues/7829
3,548,584,085
I_kwDODunzps7TgxiV
7,829
Memory leak / Large memory usage with num_workers = 0 and numerous dataset within DatasetDict
{ "avatar_url": "https://avatars.githubusercontent.com/u/24591024?v=4", "events_url": "https://api.github.com/users/raphaelsty/events{/privacy}", "followers_url": "https://api.github.com/users/raphaelsty/followers", "following_url": "https://api.github.com/users/raphaelsty/following{/other_user}", "gists_url": "https://api.github.com/users/raphaelsty/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/raphaelsty", "id": 24591024, "login": "raphaelsty", "node_id": "MDQ6VXNlcjI0NTkxMDI0", "organizations_url": "https://api.github.com/users/raphaelsty/orgs", "received_events_url": "https://api.github.com/users/raphaelsty/received_events", "repos_url": "https://api.github.com/users/raphaelsty/repos", "site_admin": false, "starred_url": "https://api.github.com/users/raphaelsty/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/raphaelsty/subscriptions", "type": "User", "url": "https://api.github.com/users/raphaelsty", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Thanks for the report, this is possibly related #7722 and #7694.\n\nCould you pls provide steps to reproduce this?", "To overcome this issue right now I did simply reduce the size of the dataset and ended up running a for loop (my training has now a constant learning rate schedule). From what I understood, and I don't know if it's possible, the solution would be to tell the backend of `datasets` to leave x% of the memory free (including memory mapping). Can't release the data right now but I will and then allow to reproduce this issue. But it will involve to have some free TB of disk", "@raphaelsty thanks for coming back to this. I assume you are running in streaming mode? That should prevent these errors but it looks like more people than just you have this problem, so a clearly reproducing example (including data + code) is highly appreciated.", "This could be related to this issue: https://github.com/huggingface/datasets/issues/4883 in which we discussed how RSS and memory mapping works and depends on the OS and disk type." ]
2025-10-24T09:51:38Z
2025-11-06T13:31:26Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug Hi team, first off, I love the datasets library! 🥰 I'm encountering a potential memory leak / increasing memory usage when training a model on a very large DatasetDict. Setup: I have a DatasetDict containing 362 distinct datasets, which sum up to ~2.8 billion rows. Training Task: I'm performing contrastive learning with SentenceTransformer and Accelerate on a single node with 4 H100, which requires me to sample from only one dataset at a time. Training Loop: At each training step, I sample ~16,000 examples from a single dataset, and then switch to a different dataset for the next step. I iterate through all 362 datasets this way. Problem: The process's memory usage continuously increases over time, eventually causing a stale status where GPUs would stop working. It seems memory from previously sampled datasets isn't being released. I've set num_workers=0 for all experiments. Chart 1: Standard DatasetDict The memory usage grows steadily until it make the training stale (RSS memory) <img width="773" height="719" alt="Image" src="https://github.com/user-attachments/assets/6606bef5-1153-4f2d-bf08-82da249d6e8d" /> Chart 2: IterableDatasetDict I also tried to use IterableDatasetDict and IterableDataset. The memory curve is "smoother," but the result is the same: it grows indefinitely and the training become stale. <img width="339" height="705" alt="Image" src="https://github.com/user-attachments/assets/ee90c1a1-6c3b-4135-9edc-90955cb1695a" /> Any feedback or guidance on how to manage this memory would be greatly appreciated! ### Steps to reproduce the bug WIP, I'll add some code that manage to reproduce this error, but not straightforward. ### Expected behavior The memory usage should remain relatively constant or plateau after a few steps. Memory used for sampling one dataset should be released before or during the sampling of the next dataset. ### Environment info Python: 3.12 Datasets: 4.3.0 SentenceTransformers: 5.1.1
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 1, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7829/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7829/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7828
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7828/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7828/comments
https://api.github.com/repos/huggingface/datasets/issues/7828/events
https://github.com/huggingface/datasets/pull/7828
3,545,562,802
PR_kwDODunzps6vT3rI
7,828
set dev version
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7828). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-23T16:34:44Z
2025-10-23T16:37:32Z
2025-10-23T16:34:50Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7828/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7828/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7828.diff", "html_url": "https://github.com/huggingface/datasets/pull/7828", "merged_at": "2025-10-23T16:34:50Z", "patch_url": "https://github.com/huggingface/datasets/pull/7828.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7828" }
true
https://api.github.com/repos/huggingface/datasets/issues/7827
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7827/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7827/comments
https://api.github.com/repos/huggingface/datasets/issues/7827/events
https://github.com/huggingface/datasets/pull/7827
3,545,543,117
PR_kwDODunzps6vTz7T
7,827
release: 4.3.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7827). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-23T16:30:48Z
2025-10-23T16:33:38Z
2025-10-23T16:31:57Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7827/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7827/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7827.diff", "html_url": "https://github.com/huggingface/datasets/pull/7827", "merged_at": "2025-10-23T16:31:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/7827.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7827" }
true
https://api.github.com/repos/huggingface/datasets/issues/7826
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7826/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7826/comments
https://api.github.com/repos/huggingface/datasets/issues/7826/events
https://github.com/huggingface/datasets/pull/7826
3,545,534,210
PR_kwDODunzps6vTyKR
7,826
picklable batch_fn
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7826). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-23T16:28:52Z
2025-10-23T16:31:59Z
2025-10-23T16:29:06Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7826/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7826/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7826.diff", "html_url": "https://github.com/huggingface/datasets/pull/7826", "merged_at": "2025-10-23T16:29:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/7826.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7826" }
true
https://api.github.com/repos/huggingface/datasets/issues/7825
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7825/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7825/comments
https://api.github.com/repos/huggingface/datasets/issues/7825/events
https://github.com/huggingface/datasets/pull/7825
3,541,176,759
PR_kwDODunzps6vFc3H
7,825
Fix ArrowInvalid for large variables
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "@lhoestq : from what I see in the CI logs ( tests/test_arrow_dataset.py::test_map_int32_overflow \r\n/opt/hostedtoolcache/Python/3.9.23/x64/lib/python3.9/site-packages/multiprocess/resource_tracker.py:219: UserWarning: resource_tracker: There appear to be 24 leaked shared_memory objects to clean up at shutdown\r\n warnings.warn('resource_tracker: There appear to be %d '), see [here](https://github.com/huggingface/datasets/actions/runs/18720027796/job/53591652242#step:13:6530), the test crashes probably since it consumes too much memory, it works locally for me though. \r\n \r\n Shall we keep the test, remove ti, or skip it, or is there a specific pytest flag you have for slow/memory-consuming tests I can add, so that it gets skipped automatically on the CI but might be run locally?\r\n \r\n EDIT: I added the `high_memory` marker (see [here](https://github.com/huggingface/datasets/pull/7825/files#diff-c480dc36aca482db963c43ee66ba21705111f62f826352c96558ef5ed1619791R4789)), so this will not be run in the CI, which is not ideal but I guess the way if we want to keep the test" ]
2025-10-22T14:42:47Z
2025-11-05T11:02:14Z
null
CONTRIBUTOR
null
null
null
null
Fixes #7821 In addition to the solution proposed in the issue, I encountered we also need to support 64bit types when writing using the `RecordBatchStreamWriter`. Not sure if we want to create such large objects in the CI, but this is the only way to test that the issue is fixed, therefore I added the unit test.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7825/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7825/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7825.diff", "html_url": "https://github.com/huggingface/datasets/pull/7825", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7825.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7825" }
true
https://api.github.com/repos/huggingface/datasets/issues/7824
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7824/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7824/comments
https://api.github.com/repos/huggingface/datasets/issues/7824/events
https://github.com/huggingface/datasets/pull/7824
3,531,240,254
PR_kwDODunzps6ukXe9
7,824
Fix batch_size default description in to_polars docstrings
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7824). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-20T06:38:19Z
2025-10-20T13:49:25Z
2025-10-20T13:49:24Z
MEMBER
null
null
null
null
Fix batch_size default description in `to_polars` docstrings.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7824/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7824/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7824.diff", "html_url": "https://github.com/huggingface/datasets/pull/7824", "merged_at": "2025-10-20T13:49:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/7824.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7824" }
true
https://api.github.com/repos/huggingface/datasets/issues/7823
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7823/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7823/comments
https://api.github.com/repos/huggingface/datasets/issues/7823/events
https://github.com/huggingface/datasets/pull/7823
3,525,440,347
PR_kwDODunzps6uRkGa
7,823
Fix random seed on shuffle and interleave_datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7823). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "Cool ! To avoid unwanted side effects it could be implemented for every class instead of using `hasattr`, and also return a new instance instead of modifying it in place. Some other functions work like that already like `shuffle_data_sources` and `shard_data_sources`.\r\n\r\nReturning a new object is actually quite important otherwise iterating on the dataset multiple times would shift the RNGs every time", "> Cool ! To avoid unwanted side effects it could be implemented for every class instead of using `hasattr`, and also return a new instance instead of modifying it in place. Some other functions work like that already like `shuffle_data_sources` and `shard_data_sources`.\r\n> \r\n> Returning a new object is actually quite important otherwise iterating on the dataset multiple times would shift the RNGs every time\r\n\r\nThanks for the review. \r\n\r\nI managed to return instances of `_BaseExamplesIterable` instead of modifying inplace and this shifts the seed correctly even in the case of multiple iterations (see additional test). But I kept the `hasattr` for the reason that we have chained ex_iterables (so an ex_iterable can have other ex_iterables) and for the child ex_iterable it is not clear if they do have a generator attribute (e.g. `ArrowExamplesIterable`, `RebatchedArrowExamplesIterable`, `SelectColumnsIterable`, `StepExamplesIterable`, `SkipExamplesIterable`, `RepeatExamplesIterable` all don't have that). We could implement a generic method `shift_rngs` on the baseclass and then just return the ex_iterable and get rid of `hasattr` but I'd prefer the `hasattr` solution here. Let me now what you think about that.\r\n " ]
2025-10-17T10:21:47Z
2025-10-24T14:21:22Z
2025-10-24T14:04:36Z
CONTRIBUTOR
null
null
null
null
closes #7567 Add `shift_rngs` method to `ExamplesIterable` that is called directly after sharding. If a generator is available (not the case for all subclasses) we update the seed of the generator by shifting by the worker_id. ~This is just the fix for `shuffle`, in the corresponding issue `interleave_datasets` is mentioned as well, which won't be fixed with this approach.~ EDIT: This is a fix for `shuffle` and `interleave_datasets`. Adding recursivity to `shift_rngs` solved `interleave_datasets` as well. Not sure though if this is completely safe or if we could destroy something with that. I don't think so but could be wrong and appreciate some guidance from the maintainers. I also checked, on a single_worker we are always handing over `index=0` so that case preserves the seed the user specified.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7823/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7823/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7823.diff", "html_url": "https://github.com/huggingface/datasets/pull/7823", "merged_at": "2025-10-24T14:04:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/7823.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7823" }
true
https://api.github.com/repos/huggingface/datasets/issues/7822
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7822/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7822/comments
https://api.github.com/repos/huggingface/datasets/issues/7822/events
https://github.com/huggingface/datasets/pull/7822
3,525,309,651
PR_kwDODunzps6uRKIJ
7,822
Retry open hf file
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7822). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-17T09:48:51Z
2025-10-17T09:52:05Z
2025-10-17T09:51:35Z
MEMBER
null
null
null
null
Fix this error ```python File "/workdir/.venv/lib/python3.13/site-packages/datasets/utils/file_utils.py", line 934, in xopen file_obj = fsspec.open(file, mode=mode, *args, **kwargs).open() File "/workdir/.venv/lib/python3.13/site-packages/fsspec/core.py", line 147, in open return self.__enter__() ~~~~~~~~~~~~~~^^ File "/workdir/.venv/lib/python3.13/site-packages/fsspec/core.py", line 105, in __enter__ f = self.fs.open(self.path, mode=mode) File "/workdir/.venv/lib/python3.13/site-packages/fsspec/spec.py", line 1338, in open f = self._open( path, ...<4 lines>... **kwargs, ) File "/workdir/.venv/lib/python3.13/site-packages/huggingface_hub/hf_file_system.py", line 275, in _open return HfFileSystemFile(self, path, mode=mode, revision=revision, block_size=block_size, **kwargs) File "/workdir/.venv/lib/python3.13/site-packages/huggingface_hub/hf_file_system.py", line 950, in __init__ self.resolved_path = fs.resolve_path(path, revision=revision) ~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workdir/.venv/lib/python3.13/site-packages/huggingface_hub/hf_file_system.py", line 198, in resolve_path repo_and_revision_exist, err = self._repo_and_revision_exist(repo_type, repo_id, revision) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workdir/.venv/lib/python3.13/site-packages/huggingface_hub/hf_file_system.py", line 125, in _repo_and_revision_exist self._api.repo_info( ~~~~~~~~~~~~~~~~~~~^ repo_id, revision=revision, repo_type=repo_type, timeout=constants.HF_HUB_ETAG_TIMEOUT ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/workdir/.venv/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "/workdir/.venv/lib/python3.13/site-packages/huggingface_hub/hf_api.py", line 2864, in repo_info return method( repo_id, ...<4 lines>... files_metadata=files_metadata, ) File "/workdir/.venv/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "/workdir/.venv/lib/python3.13/site-packages/huggingface_hub/hf_api.py", line 2721, in dataset_info r = get_session().get(path, headers=headers, timeout=timeout, params=params) File "/workdir/.venv/lib/python3.13/site-packages/requests/sessions.py", line 602, in get return self.request("GET", url, **kwargs) ~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^ File "/workdir/.venv/lib/python3.13/site-packages/requests/sessions.py", line 589, in request resp = self.send(prep, **send_kwargs) File "/workdir/.venv/lib/python3.13/site-packages/requests/sessions.py", line 703, in send r = adapter.send(request, **kwargs) File "/workdir/.venv/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 95, in send return super().send(request, *args, **kwargs) ~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workdir/.venv/lib/python3.13/site-packages/requests/adapters.py", line 690, in send raise ReadTimeout(e, request=request) requests.exceptions.ReadTimeout: (ReadTimeoutError("HTTPSConnectionPool(host='huggingface.co', port=443): Read timed out. (read timeout=10)"), '(Request ID: e7e1ae72-54a0-4ce4-b011-144fb7a3fb06)') ``` which could also be related to ```python File "/workdir/.venv/lib/python3.13/site-packages/datasets/utils/file_utils.py", line 1364, in _iter_from_urlpaths raise FileNotFoundError(urlpath) FileNotFoundError: hf://datasets/.../train-00013-of-00031.parquet ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7822/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7822/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7822.diff", "html_url": "https://github.com/huggingface/datasets/pull/7822", "merged_at": "2025-10-17T09:51:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/7822.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7822" }
true
https://api.github.com/repos/huggingface/datasets/issues/7821
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7821/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7821/comments
https://api.github.com/repos/huggingface/datasets/issues/7821/events
https://github.com/huggingface/datasets/issues/7821
3,520,913,195
I_kwDODunzps7R3N8r
7,821
Building a dataset with large variable size arrays results in error ArrowInvalid: Value X too large to fit in C integer type
{ "avatar_url": "https://avatars.githubusercontent.com/u/51880718?v=4", "events_url": "https://api.github.com/users/kkoutini/events{/privacy}", "followers_url": "https://api.github.com/users/kkoutini/followers", "following_url": "https://api.github.com/users/kkoutini/following{/other_user}", "gists_url": "https://api.github.com/users/kkoutini/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kkoutini", "id": 51880718, "login": "kkoutini", "node_id": "MDQ6VXNlcjUxODgwNzE4", "organizations_url": "https://api.github.com/users/kkoutini/orgs", "received_events_url": "https://api.github.com/users/kkoutini/received_events", "repos_url": "https://api.github.com/users/kkoutini/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kkoutini/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kkoutini/subscriptions", "type": "User", "url": "https://api.github.com/users/kkoutini", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Thanks for reporting ! You can fix this by specifying the output type explicitly and use `LargeList` which uses int64 for offsets:\n\n```python\nfeatures = Features({\"audio\": LargeList(Value(\"uint16\"))})\nds = ds.map(..., features=features)\n```\n\nIt would be cool to improve `list_of_pa_arrays_to_pyarrow_listarray()` to automatically use `LargeList` if the lists are longer than the int32 limit though. Contributions are welcome if you'd like to improve it" ]
2025-10-16T08:45:17Z
2025-10-20T13:42:05Z
null
CONTRIBUTOR
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug I used map to store raw audio waveforms of variable lengths in a column of a dataset the `map` call fails with ArrowInvalid: Value X too large to fit in C integer type. ``` Traceback (most recent call last): Traceback (most recent call last): File "...lib/python3.12/site-packages/multiprocess/pool.py", line 125, in worker result = (True, func(*args, **kwds)) ^^^^^^^^^^^^^^^^^^^ File "...lib/python3.12/site-packages/datasets/utils/py_utils.py", line 678, in _write_generator_to_queue for i, result in enumerate(func(**kwargs)): ^^^^^^^^^^^^^^^^^^^^^^^^^ File "...lib/python3.12/site-packages/datasets/arrow_dataset.py", line 3526, in _map_single writer.write_batch(batch) File "...lib/python3.12/site-packages/datasets/arrow_writer.py", line 605, in write_batch arrays.append(pa.array(typed_sequence)) ^^^^^^^^^^^^^^^^^^^^^^^^ File "pyarrow/array.pxi", line 252, in pyarrow.lib.array File "pyarrow/array.pxi", line 114, in pyarrow.lib._handle_arrow_array_protocol File "...lib/python3.12/site-packages/datasets/arrow_writer.py", line 225, in __arrow_array__ out = list_of_np_array_to_pyarrow_listarray(data) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "...lib/python3.12/site-packages/datasets/features/features.py", line 1538, in list_of_np_array_to_pyarrow_listarray return list_of_pa_arrays_to_pyarrow_listarray( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "...lib/python3.12/site-packages/datasets/features/features.py", line 1530, in list_of_pa_arrays_to_pyarrow_listarray offsets = pa.array(offsets, type=pa.int32()) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "pyarrow/array.pxi", line 362, in pyarrow.lib.array File "pyarrow/array.pxi", line 87, in pyarrow.lib._ndarray_to_array File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: Value 2148479376 too large to fit in C integer type ``` ### Steps to reproduce the bug Calling map on a dataset that returns a column with long 1d numpy arrays of variable length. Example: ```python # %% import logging import datasets import pandas as pd import numpy as np # %% def process_batch(batch, rank): res = [] for _ in batch["id"]: res.append(np.zeros((2**30)).astype(np.uint16)) return {"audio": res} if __name__ == "__main__": df = pd.DataFrame( { "id": list(range(400)), } ) ds = datasets.Dataset.from_pandas(df) try: from multiprocess import set_start_method set_start_method("spawn") except RuntimeError: print("Spawn method already set, continuing...") mapped_ds = ds.map( process_batch, batched=True, batch_size=2, with_rank=True, num_proc=2, cache_file_name="path_to_cache/tmp.arrow", writer_batch_size=200, remove_columns=ds.column_names, # disable_nullable=True, ) ``` ### Expected behavior I think the offsets should be pa.int64() if needed and not forced to be `pa.int32()` in https://github.com/huggingface/datasets/blob/3e13d30823f8ec498d56adbc18c6880a5463b313/src/datasets/features/features.py#L1535 ### Environment info - `datasets` version: 3.3.1 - Platform: Linux-5.15.0-94-generic-x86_64-with-glibc2.35 - Python version: 3.12.9 - `huggingface_hub` version: 0.29.0 - PyArrow version: 19.0.1 - Pandas version: 2.2.3 - `fsspec` version: 2024.12.0
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7821/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7821/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7820
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7820/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7820/comments
https://api.github.com/repos/huggingface/datasets/issues/7820/events
https://github.com/huggingface/datasets/pull/7820
3,518,633,577
PR_kwDODunzps6t6suZ
7,820
Keep hffs cache in workers when streaming
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7820). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-15T15:51:28Z
2025-10-17T09:59:17Z
2025-10-17T09:59:16Z
MEMBER
null
null
null
null
(and also reorder the hffs args to improve caching) When using `DataLoader(iterable_dataset, num_workers=...)` the dataset is pickled and passed to the worker. However previously the resulting dataset would be in a process with an empty hffs cache. By keeping the cache attached to `IterableDataset`, the cached hffs instances are pickled with the dataset and re-populates the cache in the DataLoader workers this requires https://github.com/huggingface/huggingface_hub/pull/3443 to work effectively though, otherwise the unpickled hffs cache would start empty cc @andimarafioti @ltmeyer
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7820/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7820/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7820.diff", "html_url": "https://github.com/huggingface/datasets/pull/7820", "merged_at": "2025-10-17T09:59:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/7820.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7820" }
true
https://api.github.com/repos/huggingface/datasets/issues/7819
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7819/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7819/comments
https://api.github.com/repos/huggingface/datasets/issues/7819/events
https://github.com/huggingface/datasets/issues/7819
3,517,086,110
I_kwDODunzps7Ronme
7,819
Cannot download opus dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/51946663?v=4", "events_url": "https://api.github.com/users/liamsun2019/events{/privacy}", "followers_url": "https://api.github.com/users/liamsun2019/followers", "following_url": "https://api.github.com/users/liamsun2019/following{/other_user}", "gists_url": "https://api.github.com/users/liamsun2019/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/liamsun2019", "id": 51946663, "login": "liamsun2019", "node_id": "MDQ6VXNlcjUxOTQ2NjYz", "organizations_url": "https://api.github.com/users/liamsun2019/orgs", "received_events_url": "https://api.github.com/users/liamsun2019/received_events", "repos_url": "https://api.github.com/users/liamsun2019/repos", "site_admin": false, "starred_url": "https://api.github.com/users/liamsun2019/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/liamsun2019/subscriptions", "type": "User", "url": "https://api.github.com/users/liamsun2019", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Hi ! it seems \"en-zh\" doesn't exist for this dataset\n\nYou can see the list of subsets here: https://huggingface.co/datasets/Helsinki-NLP/opus_books" ]
2025-10-15T09:06:19Z
2025-10-20T13:45:16Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
When I tried to download opus_books using: from datasets import load_dataset dataset = load_dataset("Helsinki-NLP/opus_books") I got the following errors: FileNotFoundError: Couldn't find any data file at /workspace/Helsinki-NLP/opus_books. Couldn't find 'Helsinki-NLP/opus_books' on the Hugging Face Hub either: LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. I also tried: dataset = load_dataset("opus_books", "en-zh") and the errors remain the same. However, I can download "mlabonne/FineTome-100k" successfully. My datasets is version 4.2.0 Any clues? Big thanks.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7819/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7819/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7818
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7818/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7818/comments
https://api.github.com/repos/huggingface/datasets/issues/7818/events
https://github.com/huggingface/datasets/issues/7818
3,515,887,618
I_kwDODunzps7RkDAC
7,818
train_test_split and stratify breaks with Numpy 2.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/24845694?v=4", "events_url": "https://api.github.com/users/davebulaval/events{/privacy}", "followers_url": "https://api.github.com/users/davebulaval/followers", "following_url": "https://api.github.com/users/davebulaval/following{/other_user}", "gists_url": "https://api.github.com/users/davebulaval/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/davebulaval", "id": 24845694, "login": "davebulaval", "node_id": "MDQ6VXNlcjI0ODQ1Njk0", "organizations_url": "https://api.github.com/users/davebulaval/orgs", "received_events_url": "https://api.github.com/users/davebulaval/received_events", "repos_url": "https://api.github.com/users/davebulaval/repos", "site_admin": false, "starred_url": "https://api.github.com/users/davebulaval/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/davebulaval/subscriptions", "type": "User", "url": "https://api.github.com/users/davebulaval", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "I can't reproduce this. Could you pls provide an example with a public dataset/artificial dataset and show how you loaded that?\n\nThis works for me:\n\n```python\nimport numpy as np\nfrom datasets import Dataset, Features, ClassLabel, Value\n\ndata = {\"text\": [f\"sample_{i}\" for i in range(100)], \"label\": [i % 3 for i in range(100)]}\nfeatures = Features({\"text\": Value(\"string\"),\n \"label\": ClassLabel(names=[\"class_0\", \"class_1\", \"class_2\"])})\ndataset = Dataset.from_dict(data, features=features)\nsplits = dataset.train_test_split(test_size=0.2, stratify_by_column=\"label\")\nprint(f\"Success with numpy {np.__version__}\")\n```\nbut it also works for `numpy<2`", "@davebulaval tried with numpy 2.3.4, and maybe i have successfully reproduced the bug!\n```\nValueError: Unable to avoid copy while creating an array as requested.\nIf using `np.array(obj, copy=False)` replace it with `np.asarray(obj)` to allow a copy when needed (no behavior change in NumPy 1.x).\nFor more details, see https://numpy.org/devdocs/numpy_2_0_migration_guide.html#adapting-to-changes-in-the-copy-keyword.\n```\n\nAlso i downgraded to numpy 1.26.4\n```\n(hf-reproduce) F:\\Python\\Machine learning\\reproducing>python repro.py\nDatasetDict({\n train: Dataset({\n features: ['text', 'label'],\n num_rows: 16\n })\n test: Dataset({\n features: ['text', 'label'],\n num_rows: 4\n })\n})\n```", "Also @CloseChoice The bug only appears in cases where the Arrow array cannot be represented as a contiguous NumPy array without copying.\n\nSo closing the discussion loop here - \n\nThe error occurs because `train_test_split(..., stratify_by_column=...)` attempts to convert\nan Arrow column to a NumPy array using `np.array(..., copy=False)`.\n\nIn NumPy <2.0 this silently allowed a copy if needed.\nIn NumPy ≥2.0 this raises:\nValueError: Unable to avoid copy while creating an array as requested.\n\nThis only happens when the Arrow column is not contiguous in memory, which explains\nwhy some datasets reproduce it and others do not." ]
2025-10-15T00:01:19Z
2025-10-28T16:10:44Z
2025-10-28T16:10:44Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug As stated in the title, since Numpy changed in version >2.0 with copy, the stratify parameters break. e.g. `all_dataset.train_test_split(test_size=0.2,stratify_by_column="label")` returns a Numpy error. It works if you downgrade Numpy to a version lower than 2.0. ### Steps to reproduce the bug 1. Numpy > 2.0 2. `all_dataset.train_test_split(test_size=0.2,stratify_by_column="label")` ### Expected behavior It returns a stratified split as per the results of Numpy < 2.0 ### Environment info - `datasets` version: 2.14.4 - Platform: Linux-6.8.0-85-generic-x86_64-with-glibc2.35 - Python version: 3.13.7 - Huggingface_hub version: 0.34.4 - PyArrow version: 19.0.0 - Pandas version: 2.3.2
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7818/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7818/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7817
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7817/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7817/comments
https://api.github.com/repos/huggingface/datasets/issues/7817/events
https://github.com/huggingface/datasets/pull/7817
3,515,755,952
PR_kwDODunzps6tw-GG
7,817
fix: better args passthrough for `_batch_setitems()`
{ "avatar_url": "https://avatars.githubusercontent.com/u/58419736?v=4", "events_url": "https://api.github.com/users/sghng/events{/privacy}", "followers_url": "https://api.github.com/users/sghng/followers", "following_url": "https://api.github.com/users/sghng/following{/other_user}", "gists_url": "https://api.github.com/users/sghng/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sghng", "id": 58419736, "login": "sghng", "node_id": "MDQ6VXNlcjU4NDE5NzM2", "organizations_url": "https://api.github.com/users/sghng/orgs", "received_events_url": "https://api.github.com/users/sghng/received_events", "repos_url": "https://api.github.com/users/sghng/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sghng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sghng/subscriptions", "type": "User", "url": "https://api.github.com/users/sghng", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "@sghng There is a regression with python 3.13.8 when lm-eval is calling datasets load\r\n\r\n```python\r\nself = <test_llama3_2.TestLlama3_2 testMethod=test_llama3_2>\r\n\r\n def test_llama3_2(self):\r\n> self.quant_lm_eval()\r\n\r\ntests/models/test_llama3_2.py:35: \r\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\ntests/models/model_test.py:773: in quant_lm_eval\r\n self.model, _ = self.quantModel(self.NATIVE_MODEL_ID, batch_size=self.QUANT_BATCH_SIZE, trust_remote_code=self.TRUST_REMOTE_CODE, dtype=self.TORCH_DTYPE)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\ntests/models/model_test.py:588: in quantModel\r\n reuse_candidates = self.perform_post_quant_validation(path, trust_remote_code=trust_remote_code)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\ntests/models/model_test.py:250: in perform_post_quant_validation\r\n arc_records[backend] = self.run_arc_challenge_eval(model, backend, trust_remote_code=trust_remote_code)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\ntests/models/model_test.py:217: in run_arc_challenge_eval\r\n task_results = self.lm_eval(\r\ntests/models/model_test.py:753: in lm_eval\r\n raise e\r\ntests/models/model_test.py:699: in lm_eval\r\n results = GPTQModel.eval(\r\ngptqmodel/models/auto.py:474: in eval\r\n results = simple_evaluate(\r\n../lm-evaluation-harness/lm_eval/utils.py:456: in _wrapper\r\n return fn(*args, **kwargs)\r\n ^^^^^^^^^^^^^^^^^^^\r\n../lm-evaluation-harness/lm_eval/evaluator.py:283: in simple_evaluate\r\n task_dict = get_task_dict(\r\n../lm-evaluation-harness/lm_eval/tasks/__init__.py:635: in get_task_dict\r\n task_name_from_string_dict = task_manager.load_task_or_group(\r\n../lm-evaluation-harness/lm_eval/tasks/__init__.py:426: in load_task_or_group\r\n collections.ChainMap(\r\n../lm-evaluation-harness/lm_eval/tasks/__init__.py:428: in <lambda>\r\n lambda task: self._load_individual_task_or_group(task),\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n../lm-evaluation-harness/lm_eval/tasks/__init__.py:326: in _load_individual_task_or_group\r\n return _load_task(task_config, task=name_or_config)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n../lm-evaluation-harness/lm_eval/tasks/__init__.py:286: in _load_task\r\n task_object = ConfigurableTask(config=config)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n../lm-evaluation-harness/lm_eval/api/task.py:865: in __init__\r\n self.download(self.config.dataset_kwargs)\r\n../lm-evaluation-harness/lm_eval/api/task.py:997: in download\r\n self.dataset = datasets.load_dataset(\r\n../datasets/src/datasets/load.py:1397: in load_dataset\r\n builder_instance = load_dataset_builder(\r\n../datasets/src/datasets/load.py:1185: in load_dataset_builder\r\n builder_instance._use_legacy_cache_dir_if_possible(dataset_module)\r\n../datasets/src/datasets/builder.py:612: in _use_legacy_cache_dir_if_possible\r\n self._check_legacy_cache2(dataset_module) or self._check_legacy_cache() or None\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n../datasets/src/datasets/builder.py:485: in _check_legacy_cache2\r\n config_id = self.config.name + \"-\" + Hasher.hash({\"data_files\": self.config.data_files})\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n../datasets/src/datasets/fingerprint.py:188: in hash\r\n return cls.hash_bytes(dumps(value))\r\n ^^^^^^^^^^^^\r\n../datasets/src/datasets/utils/_dill.py:120: in dumps\r\n dump(obj, file)\r\n../datasets/src/datasets/utils/_dill.py:114: in dump\r\n Pickler(file, recurse=True).dump(obj)\r\n../vm313t/lib/python3.13t/site-packages/dill/_dill.py:428: in dump\r\n StockPickler.dump(self, obj)\r\n/usr/lib/python3.13/pickle.py:484: in dump\r\n self.save(obj)\r\n../datasets/src/datasets/utils/_dill.py:70: in save\r\n dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id)\r\n../vm313t/lib/python3.13t/site-packages/dill/_dill.py:422: in save\r\n StockPickler.save(self, obj, save_persistent_id)\r\n/usr/lib/python3.13/pickle.py:558: in save\r\n f(self, obj) # Call unbound method with explicit self\r\n ^^^^^^^^^^^^\r\n../vm313t/lib/python3.13t/site-packages/dill/_dill.py:1262: in save_module_dict\r\n StockPickler.save_dict(pickler, obj)\r\n_ _ _ _ _ _ _ _ _`` _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \r\n\r\nself = <datasets.utils._dill.Pickler object at 0x4eaf175e390>\r\nobj = {'data_files': {'train': ['hf://datasets/allenai/ai2_arc@210d026faf9955653af8916fad021475a3f00453/ARC-Challenge/train-...//datasets/allenai/ai2_arc@210d026faf9955653af8916fad021475a3f00453/ARC-Challenge/validation-00000-of-00001.parquet']}}\r\n\r\n def save_dict(self, obj):\r\n if self.bin:\r\n self.write(EMPTY_DICT)\r\n else: # proto 0 -- can't use EMPTY_DICT\r\n self.write(MARK + DICT)\r\n \r\n self.memoize(obj)\r\n> self._batch_setitems(obj.items())\r\nE TypeError: Pickler._batch_setitems() missing 1 required positional argument: 'obj'\r\n\r\n/usr/lib/python3.13/pickle.py:990: TypeError\r\n```\r\n\r\nPython 3.13.8\r\n```python\r\n(vm313t) root@gpu-base:~/datasets# pip show dill datasets transformers lm-eval\r\nName: dill\r\nVersion: 0.4.0\r\nSummary: serialize all of Python\r\nHome-page: https://github.com/uqfoundation/dill\r\nAuthor: Mike McKerns\r\nAuthor-email: mmckerns@uqfoundation.org\r\nLicense: BSD-3-Clause\r\nLocation: /root/vm313t/lib/python3.13t/site-packages\r\nRequires: \r\nRequired-by: datasets, evaluate, GPTQModel, lm_eval, multiprocess\r\n---\r\nName: datasets\r\nVersion: 4.2.1.dev0 <-- this PR\r\nSummary: HuggingFace community-driven open-source library of datasets\r\nHome-page: https://github.com/huggingface/datasets\r\nAuthor: HuggingFace Inc.\r\nAuthor-email: thomas@huggingface.co\r\nLicense: Apache 2.0\r\nLocation: /root/vm313t/lib/python3.13t/site-packages\r\nEditable project location: /root/datasets\r\nRequires: dill, filelock, fsspec, httpx, huggingface-hub, multiprocess, numpy, packaging, pandas, pyarrow, pyyaml, requests, tqdm, xxhash\r\nRequired-by: evaluate, GPTQModel, lm_eval\r\n---\r\nName: transformers\r\nVersion: 4.57.1\r\nSummary: State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow\r\nHome-page: https://github.com/huggingface/transformers\r\nAuthor: The Hugging Face team (past and future) with the help of all our contributors (https://github.com/huggingface/transformers/graphs/contributors)\r\nAuthor-email: transformers@huggingface.co\r\nLicense: Apache 2.0 License\r\nLocation: /root/vm313t/lib/python3.13t/site-packages\r\nRequires: filelock, huggingface-hub, numpy, packaging, pyyaml, regex, requests, safetensors, tokenizers, tqdm\r\nRequired-by: GPTQModel, lm_eval, peft, tokenicer\r\n---\r\nName: lm_eval\r\nVersion: 0.4.9.1\r\nSummary: A framework for evaluating language models\r\nHome-page: https://github.com/EleutherAI/lm-evaluation-harness\r\nAuthor: \r\nAuthor-email: EleutherAI <contact@eleuther.ai>\r\nLicense: MIT\r\nLocation: /root/vm313t/lib/python3.13t/site-packages\r\nEditable project location: /root/lm-evaluation-harness\r\nRequires: accelerate, datasets, dill, evaluate, jsonlines, more_itertools, numexpr, peft, pybind11, pytablewriter, rouge-score, sacrebleu, scikit-learn, sqlitedict, torch, tqdm-multiprocess, transformers, word2number, zstandard\r\nRequired-by: \r\n```", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7817). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "@lhoestq Using `*args` works well. In fact we might as well pass through `**kwargs` as well to make the intent clearer.\r\n\r\nIt also appears to me that there could some other changes.\r\n\r\n```python\r\ndef _batch_setitems(self, items, obj):\r\n if self._legacy_no_dict_keys_sorting:\r\n return super()._batch_setitems(items)\r\n```\r\n\r\nPython dictionaries are insertion ordered official since Python 3.7, which came to end-of-life in mid 2023 already. I tried installing `datasets` for Python 3.8 and it resolves to version `3.1`, so I think it's safe to drop this in latest version.\r\n\r\n```python\r\n dill.Pickler._batch_setitems(self, items)\r\n```\r\n\r\nThis line should simply be `return super()._batch_setitems(items)`, since we're already extending from `dill.Pickler`. (not having `return` in the original version is probably a bug)" ]
2025-10-14T22:51:51Z
2025-10-28T17:32:12Z
2025-10-27T17:08:25Z
CONTRIBUTOR
null
null
null
null
In Python 3.14, there's a change in the signature of `_Pickler._batch_setitems`. It's changed to: ```python # pickle.py def _batch_setitems(self, items, obj): # Helper to batch up SETITEMS sequences; proto >= 1 only save = self.save write = self.write ``` To accomodate this, in `dill`, we have this compatibility code: ```python if sys.hexversion < 0x30E00A1: pickler._batch_setitems(iter(source.items())) else: pickler._batch_setitems(iter(source.items()), obj=obj) ``` Thus, the datasets package will emit this error ``` │ /Users/sghuang/mamba/envs/ds/lib/python3.14/site-packages/dill/_dill.py:1262 in save_module_dict │ │ │ │ 1259 │ │ if is_dill(pickler, child=False) and pickler._session: │ │ 1260 │ │ │ # we only care about session the first pass thru │ │ 1261 │ │ │ pickler._first_pass = False │ │ ❱ 1262 │ │ StockPickler.save_dict(pickler, obj) │ │ 1263 │ │ logger.trace(pickler, "# D2") │ │ 1264 │ return │ │ 1265 │ │ │ │ /Users/sghuang/mamba/envs/ds/lib/python3.14/pickle.py:1133 in save_dict │ │ │ │ 1130 │ │ print(f"Line number: {inspect.getsourcelines(method)[1]}") │ │ 1131 │ │ print(f"Full path: {inspect.getmodule(method)}") │ │ 1132 │ │ print(f"Class: {method.__qualname__}") │ │ ❱ 1133 │ │ self._batch_setitems(obj.items(), obj) │ │ 1134 │ │ │ 1135 │ dispatch[dict] = save_dict │ │ 1136 │ ╰──────────────────────────────────────────────────────────────────────────────────────────────────╯ TypeError: Pickler._batch_setitems() takes 2 positional arguments but 3 were given [NOTE] when serializing datasets.table.InMemoryTable state [NOTE] when serializing datasets.table.InMemoryTable object ``` To fix it, we update the signature of the `_batch_setitems` method defined in `utils/_dill.py`. This fix should be backward compatible, since the compatibility is handled by `dill`. This should close #7813. Similar to https://github.com/joblib/joblib/issues/1658. Related to https://github.com/uqfoundation/dill/pull/724.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/7817/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7817/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7817.diff", "html_url": "https://github.com/huggingface/datasets/pull/7817", "merged_at": "2025-10-27T17:08:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/7817.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7817" }
true
https://api.github.com/repos/huggingface/datasets/issues/7816
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7816/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7816/comments
https://api.github.com/repos/huggingface/datasets/issues/7816/events
https://github.com/huggingface/datasets/issues/7816
3,512,210,206
I_kwDODunzps7RWBMe
7,816
disable_progress_bar() not working as expected
{ "avatar_url": "https://avatars.githubusercontent.com/u/5577741?v=4", "events_url": "https://api.github.com/users/windmaple/events{/privacy}", "followers_url": "https://api.github.com/users/windmaple/followers", "following_url": "https://api.github.com/users/windmaple/following{/other_user}", "gists_url": "https://api.github.com/users/windmaple/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/windmaple", "id": 5577741, "login": "windmaple", "node_id": "MDQ6VXNlcjU1Nzc3NDE=", "organizations_url": "https://api.github.com/users/windmaple/orgs", "received_events_url": "https://api.github.com/users/windmaple/received_events", "repos_url": "https://api.github.com/users/windmaple/repos", "site_admin": false, "starred_url": "https://api.github.com/users/windmaple/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/windmaple/subscriptions", "type": "User", "url": "https://api.github.com/users/windmaple", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "@xianbaoqian ", "Closing this one since it's a Xet issue." ]
2025-10-14T03:25:39Z
2025-10-14T23:49:26Z
2025-10-14T23:49:26Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug Hi, I'm trying to load a dataset on Kaggle TPU image. There is some known compat issue with progress bar on Kaggle, so I'm trying to disable the progress bar globally. This does not work as you can see in [here](https://www.kaggle.com/code/windmaple/hf-datasets-issue). In contract, disabling progress bar for snapshot_download() works as expected as in [here](https://www.kaggle.com/code/windmaple/snapshot-download-error). ### Steps to reproduce the bug See this [notebook](https://www.kaggle.com/code/windmaple/hf-datasets-issue). There is sth. wrong with `shell_paraent`. ### Expected behavior The downloader should disable progress bar and move forward w/ no error. ### Environment info The latest version as I did: !pip install -U datasets ipywidgets ipykernel
{ "avatar_url": "https://avatars.githubusercontent.com/u/5577741?v=4", "events_url": "https://api.github.com/users/windmaple/events{/privacy}", "followers_url": "https://api.github.com/users/windmaple/followers", "following_url": "https://api.github.com/users/windmaple/following{/other_user}", "gists_url": "https://api.github.com/users/windmaple/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/windmaple", "id": 5577741, "login": "windmaple", "node_id": "MDQ6VXNlcjU1Nzc3NDE=", "organizations_url": "https://api.github.com/users/windmaple/orgs", "received_events_url": "https://api.github.com/users/windmaple/received_events", "repos_url": "https://api.github.com/users/windmaple/repos", "site_admin": false, "starred_url": "https://api.github.com/users/windmaple/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/windmaple/subscriptions", "type": "User", "url": "https://api.github.com/users/windmaple", "user_view_type": "public" }
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7816/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7816/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7815
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7815/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7815/comments
https://api.github.com/repos/huggingface/datasets/issues/7815/events
https://github.com/huggingface/datasets/pull/7815
3,511,338,522
PR_kwDODunzps6tiDIT
7,815
Add nifti support
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7815). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "Btw I couldn't resist but share your PR with the community online on twitter already, I hope this is fine !", "> Btw I couldn't resist but share your PR with the community online on twitter already, I hope this is fine !\r\n\r\nWow, that was quick! Thanks, already liked your comment, I appreciate it!", "NIfTI support is out in `datasets==4.4.0` ! 🥳 \r\n\r\nBtw do you know a good NIfTI vizualizer in HTML/JS or using python ? We could add something like `.to_html()` (or equivalent) to view data in a notebook and enable the Dataset Viewer on HF if it can be useful\r\n\r\ncc @cfahlgren1 @georgiachanning for viz", "Hi, I have a quick question while testing the new NIfTI support.\r\n\r\nI cloned the latest `main` branch, installed it locally using `pip install -e .`, and ran the following:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\r\n \"TobiasPitters/NIfTI-SIRF-exercises-geometry\",\r\n split=\"train\"\r\n)\r\ndataset[0]['nifti'].get_fdata()[0].shape\r\n```\r\n\r\nHowever, I’m getting the following error:\r\n\r\n```\r\nFileNotFoundError: No such file or no access: 'data/nifti/OBJECT_phantom_T2W_TSE_Sag_18_1.nii'\r\n```\r\n\r\nWhen I manually place the NIfTI file in that local path, it works fine.\r\nBut I assume the intended behavior is for the `.nii` file to be included in the dataset hosted on the Hub, so that `load_dataset()` automatically loads it without relying on a local file path.\r\n\r\nInterestingly, after adding the `embed_storage` method, it started working properly.\r\nCould you please confirm whether this is the expected behavior, or if my previous setup was missing something?", "> ```python\r\n> from datasets import load_dataset\r\n> \r\n> dataset = load_dataset(\r\n> \"TobiasPitters/NIfTI-SIRF-exercises-geometry\",\r\n> split=\"train\"\r\n> )\r\n> dataset[0]['nifti'].get_fdata()[0].shape\r\n> ```\r\n\r\nThanks for the report, I can confirm this. It's a problem with the dataset. Can you try this:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nimport nibabel as nib\r\n\r\ndataset = load_dataset(\r\n \"TobiasPitters/test-nifti-unzipped\",\r\n split=\"train\" # Load as single Dataset, not DatasetDict\r\n)\r\n\r\nprint(\"length dataset:\", len(dataset))\r\nfor item in dataset:\r\n assert isinstance(item[\"nifti\"], nib.nifti1.Nifti1Image)\r\n```\r\nIf your interested in `\"TobiasPitters/NIfTI-SIRF-exercises-geometry\"` I can give it a shot to reupload correctly, otherwise I'd take it down.", "> NIfTI\r\n\r\nI would suggest https://github.com/rii-mango/Papaya, just tested it and it looks quite good. How would it work to add that to the dataset-viewer?\r\n\r\nAnd I assume you'd like to have the `to_html` feature on the `NifTI` class?\r\n\r\nEDIT: do we have anything like this already for other features? Couldn't find anything. I mean we can do this in different ways, simply inlining papaya or building custom components (like e.g. SHAP is doing). If we decide for the latter, this means that we'll need to build js components in datasets, so we'll need a bundler, etc. but this provides the highest flexibility. If that's of interest, I can take a look into this.", "Hi, thanks for your help earlier. I tested the dataset you shared (TobiasPitters/test-nifti-unzipped), and it works perfectly — all NIfTI files load correctly and get_fdata() returns valid arrays.\r\n\r\nHowever, when I upload my own dataset to the Hugging Face Hub using my code, it doesn’t work properly. The NIfTI files seem not to decode correctly. can you check it?\r\n\r\n```\r\ntrain_dataset = Dataset.from_pandas(train_df)\r\n\r\ndef cast_dataset(dataset):\r\n dataset = dataset.cast_column(\"nifti\", Nifti(decode=True))\r\n dataset = dataset.cast_column(\"label\", ClassLabel(num_classes=10, names=[str(i) for i in range(10)]))\r\n\r\ntrain_dataset = cast_dataset(train_dataset)\r\n```", "> Hi, thanks for your help earlier. I tested the dataset you shared (TobiasPitters/test-nifti-unzipped), and it works perfectly — all NIfTI files load correctly and get_fdata() returns valid arrays.\r\n> \r\n> However, when I upload my own dataset to the Hugging Face Hub using my code, it doesn’t work properly. The NIfTI files seem not to decode correctly. can you check it?\r\n> \r\n> ```\r\n> train_dataset = Dataset.from_pandas(train_df)\r\n> \r\n> def cast_dataset(dataset):\r\n> dataset = dataset.cast_column(\"nifti\", Nifti(decode=True))\r\n> dataset = dataset.cast_column(\"label\", ClassLabel(num_classes=10, names=[str(i) for i in range(10)]))\r\n> \r\n> train_dataset = cast_dataset(train_dataset)\r\n> ```\r\n\r\nAre you using zipped Nifti files? It seems like there is an issue with that. I found that this creates problems, that in `decode_example` the path is something like `'gzip://T1.nii::/home/tobias/programming/github/datasets/nitest-balls1/NIFTI/T1.nii.gz'`, and then we go down the remote path which results in an `KeyError` since repo_id is not specified. The root cause for this is in the [DownloadManager.extract](https://github.com/huggingface/datasets/blob/main/src/datasets/download/download_manager.py#L297-L306) method, where we extract compressed files. \r\n\r\n@lhoestq : what do you suggest here? We could probably do something like this in the [`decode_example`](https://github.com/huggingface/datasets/blob/main/src/datasets/features/nifti.py#L144):\r\n```python\r\nif path.startswith(\"gzip:\"):\r\n path = path.split(\"::\")[-1]\r\n```\r\nThough I would need to test if this is actually OS agnostic.\r\n ", "I think the issue with gzip can be fixed using the same code as in Image() imo:\r\n```diff\r\n- try:\r\n- repo_id = string_to_dict(source_url, pattern)[\"repo_id\"]\r\n- token = token_per_repo_id.get(repo_id)\r\n- except ValueError:\r\n- token = None\r\n+ source_url_fields = string_to_dict(source_url, pattern)\r\n+ token = (\r\n+ token_per_repo_id.get(source_url_fields[\"repo_id\"]) if source_url_fields is not None else None\r\n+ )\r\n```\r\n", "> Hi, thanks for your help earlier. I tested the dataset you shared (TobiasPitters/test-nifti-unzipped), and it works perfectly — all NIfTI files load correctly and get_fdata() returns valid arrays.\r\n> \r\n> However, when I upload my own dataset to the Hugging Face Hub using my code, it doesn’t work properly. The NIfTI files seem not to decode correctly. can you check it?\r\n> \r\n> ```\r\n> train_dataset = Dataset.from_pandas(train_df)\r\n> \r\n> def cast_dataset(dataset):\r\n> dataset = dataset.cast_column(\"nifti\", Nifti(decode=True))\r\n> dataset = dataset.cast_column(\"label\", ClassLabel(num_classes=10, names=[str(i) for i in range(10)]))\r\n> \r\n> train_dataset = cast_dataset(train_dataset)\r\n> ```\r\n\r\nCan you pls try with [this branch](https://github.com/CloseChoice/datasets/tree/fix-embed-storage-nifti): \r\n```\r\npip install git+https://github.com/CloseChoice/datasets.git@fix-embed-storage-nifti\r\n```\r\nThis should fix the existing problems with NifTI", "I checked and it looks like the fix-embed-storage-nifti branch has already been merged into the main. And it worked fine. thanks." ]
2025-10-13T20:07:32Z
2025-11-07T00:52:01Z
2025-10-24T13:22:18Z
CONTRIBUTOR
null
null
null
null
Add support for NIfTI. supports #7804 This PR follows https://github.com/huggingface/datasets/pull/7325 very closely I am a bit unsure what we need to add to the `document_dataset.mdx` and `document_load.mdx`. I should probably create a dataset on the hub first to create this guide instead of copy+pasting from PDF. Open todos: - [x] create nifti dataset on the hub - ~[ ] update `document_dataset.mdx` and `document_load.mdx`~ EDIT: I tested with two datasets I created on the hub: - https://huggingface.co/datasets/TobiasPitters/test-nifti-unzipped - https://huggingface.co/datasets/TobiasPitters/test-nifti for zipped (file extension `.nii.gz` and unzipped `.nii`) files and both seem to work fine. Also tested loading locally and that seems to work as well. Here is the scriptsthat I ran against the hub: ```python from pathlib import Path from datasets import load_dataset import nibabel as nib dataset = load_dataset( "TobiasPitters/test-nifti-unzipped", split="test" # Load as single Dataset, not DatasetDict ) print("length dataset unzipped:", len(dataset)) for item in dataset: isinstance(item["nifti"], nib.nifti1.Nifti1Image) dataset = load_dataset( "TobiasPitters/test-nifti", split="train" # Load as single Dataset, not DatasetDict ) print("length dataset zipped:", len(dataset)) for item in dataset: isinstance(item["nifti"], nib.nifti1.Nifti1Image) ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7815/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7815/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7815.diff", "html_url": "https://github.com/huggingface/datasets/pull/7815", "merged_at": "2025-10-24T13:22:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/7815.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7815" }
true
https://api.github.com/repos/huggingface/datasets/issues/7814
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7814/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7814/comments
https://api.github.com/repos/huggingface/datasets/issues/7814/events
https://github.com/huggingface/datasets/pull/7814
3,510,488,792
PR_kwDODunzps6tfJCm
7,814
Allow streaming hdf5 files
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7814). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-13T15:25:44Z
2025-10-13T15:28:51Z
2025-10-13T15:28:49Z
MEMBER
null
null
null
null
Add streaming support after https://github.com/huggingface/datasets/pull/7690, cc @klamike :) ## Details in `datasets` loaders, `open()` is extended to work with files that are on disk but also on HF. Files on HF are streamed using HTTP range requests using the `HfFileSystem` implementation in the `huggingface_hub` library.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7814/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7814/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7814.diff", "html_url": "https://github.com/huggingface/datasets/pull/7814", "merged_at": "2025-10-13T15:28:49Z", "patch_url": "https://github.com/huggingface/datasets/pull/7814.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7814" }
true
https://api.github.com/repos/huggingface/datasets/issues/7813
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7813/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7813/comments
https://api.github.com/repos/huggingface/datasets/issues/7813/events
https://github.com/huggingface/datasets/issues/7813
3,503,446,288
I_kwDODunzps7Q0lkQ
7,813
Caching does not work when using python3.14
{ "avatar_url": "https://avatars.githubusercontent.com/u/142020129?v=4", "events_url": "https://api.github.com/users/intexcor/events{/privacy}", "followers_url": "https://api.github.com/users/intexcor/followers", "following_url": "https://api.github.com/users/intexcor/following{/other_user}", "gists_url": "https://api.github.com/users/intexcor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/intexcor", "id": 142020129, "login": "intexcor", "node_id": "U_kgDOCHcOIQ", "organizations_url": "https://api.github.com/users/intexcor/orgs", "received_events_url": "https://api.github.com/users/intexcor/received_events", "repos_url": "https://api.github.com/users/intexcor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/intexcor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/intexcor/subscriptions", "type": "User", "url": "https://api.github.com/users/intexcor", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "https://github.com/uqfoundation/dill/issues/725", "@intexcor does #7817 fix your problem?" ]
2025-10-10T15:36:46Z
2025-10-27T17:08:26Z
2025-10-27T17:08:26Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug Traceback (most recent call last): File "/workspace/ctn.py", line 8, in <module> ds = load_dataset(f"naver-clova-ix/synthdog-{lang}") # или "synthdog-zh" для китайского File "/workspace/.venv/lib/python3.14/site-packages/datasets/load.py", line 1397, in load_dataset builder_instance = load_dataset_builder( path=path, ...<10 lines>... **config_kwargs, ) File "/workspace/.venv/lib/python3.14/site-packages/datasets/load.py", line 1185, in load_dataset_builder builder_instance._use_legacy_cache_dir_if_possible(dataset_module) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/builder.py", line 612, in _use_legacy_cache_dir_if_possible self._check_legacy_cache2(dataset_module) or self._check_legacy_cache() or None ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/builder.py", line 485, in _check_legacy_cache2 config_id = self.config.name + "-" + Hasher.hash({"data_files": self.config.data_files}) ~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/fingerprint.py", line 188, in hash return cls.hash_bytes(dumps(value)) ~~~~~^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/utils/_dill.py", line 120, in dumps dump(obj, file) ~~~~^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/utils/_dill.py", line 114, in dump Pickler(file, recurse=True).dump(obj) ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/dill/_dill.py", line 428, in dump StockPickler.dump(self, obj) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^ File "/usr/lib/python3.14/pickle.py", line 498, in dump self.save(obj) ~~~~~~~~~^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/datasets/utils/_dill.py", line 70, in save dill.Pickler.save(self, obj, save_persistent_id=save_persistent_id) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/dill/_dill.py", line 422, in save StockPickler.save(self, obj, save_persistent_id) ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/lib/python3.14/pickle.py", line 572, in save f(self, obj) # Call unbound method with explicit self ~^^^^^^^^^^^ File "/workspace/.venv/lib/python3.14/site-packages/dill/_dill.py", line 1262, in save_module_dict StockPickler.save_dict(pickler, obj) ~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^ File "/usr/lib/python3.14/pickle.py", line 1064, in save_dict self._batch_setitems(obj.items(), obj) ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^ TypeError: Pickler._batch_setitems() takes 2 positional arguments but 3 were given ### Steps to reproduce the bug ds_train = ds["train"].map(lambda x: {**x, "lang": lang}) ### Expected behavior Fixed bugs ### Environment info - `datasets` version: 4.2.0 - Platform: Linux-6.8.0-85-generic-x86_64-with-glibc2.39 - Python version: 3.14.0 - `huggingface_hub` version: 0.35.3 - PyArrow version: 21.0.0 - Pandas version: 2.3.3 - `fsspec` version: 2025.9.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7813/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7813/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7812
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7812/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7812/comments
https://api.github.com/repos/huggingface/datasets/issues/7812/events
https://github.com/huggingface/datasets/pull/7812
3,500,901,422
PR_kwDODunzps6s_New
7,812
docs: document_dataset PDFs & OCR
{ "avatar_url": "https://avatars.githubusercontent.com/u/34215814?v=4", "events_url": "https://api.github.com/users/ethanknights/events{/privacy}", "followers_url": "https://api.github.com/users/ethanknights/followers", "following_url": "https://api.github.com/users/ethanknights/following{/other_user}", "gists_url": "https://api.github.com/users/ethanknights/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ethanknights", "id": 34215814, "login": "ethanknights", "node_id": "MDQ6VXNlcjM0MjE1ODE0", "organizations_url": "https://api.github.com/users/ethanknights/orgs", "received_events_url": "https://api.github.com/users/ethanknights/received_events", "repos_url": "https://api.github.com/users/ethanknights/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ethanknights/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ethanknights/subscriptions", "type": "User", "url": "https://api.github.com/users/ethanknights", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
2025-10-09T23:31:41Z
2025-10-20T14:03:53Z
2025-10-20T14:03:53Z
CONTRIBUTOR
null
null
null
null
Use acronyms consistently across document_dataset docs.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7812/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7812/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7812.diff", "html_url": "https://github.com/huggingface/datasets/pull/7812", "merged_at": "2025-10-20T14:03:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/7812.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7812" }
true
https://api.github.com/repos/huggingface/datasets/issues/7811
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7811/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7811/comments
https://api.github.com/repos/huggingface/datasets/issues/7811/events
https://github.com/huggingface/datasets/issues/7811
3,500,741,658
I_kwDODunzps7QqRQa
7,811
SIGSEGV when Python exits due to near null deref
{ "avatar_url": "https://avatars.githubusercontent.com/u/5192353?v=4", "events_url": "https://api.github.com/users/iankronquist/events{/privacy}", "followers_url": "https://api.github.com/users/iankronquist/followers", "following_url": "https://api.github.com/users/iankronquist/following{/other_user}", "gists_url": "https://api.github.com/users/iankronquist/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/iankronquist", "id": 5192353, "login": "iankronquist", "node_id": "MDQ6VXNlcjUxOTIzNTM=", "organizations_url": "https://api.github.com/users/iankronquist/orgs", "received_events_url": "https://api.github.com/users/iankronquist/received_events", "repos_url": "https://api.github.com/users/iankronquist/repos", "site_admin": false, "starred_url": "https://api.github.com/users/iankronquist/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iankronquist/subscriptions", "type": "User", "url": "https://api.github.com/users/iankronquist", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "The issue seems to come from `dill` which is a `datasets` dependency, e.g. this segfaults:\n\n```python\nimport dill\nfrom tqdm import tqdm\nprogress_bar = tqdm(total=(1000), unit='cols', desc='cols ')\nprogress_bar.update(1)\n```\n\n`tqdm` seems to segfault when `dill` is imported. I only found this about segfault but it's maybe not related https://github.com/tqdm/tqdm/issues/1678 ?", "After more investigation it seems to be because of it imports `__main__`. This segfaults:\n\n```python\nimport __main__\nfrom tqdm import tqdm\nprogress_bar = tqdm(total=(1000), unit='cols', desc='cols ')\nprogress_bar.update(1)\n```\n\nI opened an issue at https://github.com/tqdm/tqdm/issues/1687", "Here is a workaround. You can run your code as long as the progress bar is closed before exiting.\n\n```python\nfrom datasets import load_dataset\nfrom tqdm import tqdm\n\nprogress_bar = tqdm(total=(1000), unit='cols', desc='cols ')\nprogress_bar.update(1)\nprogress_bar.close() # avoids the segfault\n```", "https://github.com/tqdm/tqdm/issues/1687#issuecomment-3392457094" ]
2025-10-09T22:00:11Z
2025-10-10T22:09:24Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug When I run the following python script using datasets I get a segfault. ```python from datasets import load_dataset from tqdm import tqdm progress_bar = tqdm(total=(1000), unit='cols', desc='cols ') progress_bar.update(1) ``` ``` % lldb -- python3 crashmin.py (lldb) target create "python3" Current executable set to '/Users/ian/bug/venv/bin/python3' (arm64). (lldb) settings set -- target.run-args "crashmin.py" (lldb) r Process 8095 launched: '/Users/ian/bug/venv/bin/python3' (arm64) Process 8095 stopped * thread #2, stop reason = exec frame #0: 0x0000000100014b30 dyld`_dyld_start dyld`_dyld_start: -> 0x100014b30 <+0>: mov x0, sp 0x100014b34 <+4>: and sp, x0, #0xfffffffffffffff0 0x100014b38 <+8>: mov x29, #0x0 ; =0 Target 0: (Python) stopped. (lldb) c Process 8095 resuming cols : 0% 0/1000 [00:00<?, ?cols/s]Process 8095 stopped * thread #2, queue = 'com.apple.main-thread', stop reason = EXC_BAD_ACCESS (code=1, address=0x10) frame #0: 0x0000000101783454 _datetime.cpython-313-darwin.so`delta_new + 188 _datetime.cpython-313-darwin.so`delta_new: -> 0x101783454 <+188>: ldr x3, [x20, #0x10] 0x101783458 <+192>: adrp x0, 10 0x10178345c <+196>: add x0, x0, #0x6fc ; "seconds" Target 0: (Python) stopped. (lldb) bt * thread #2, queue = 'com.apple.main-thread', stop reason = EXC_BAD_ACCESS (code=1, address=0x10) * frame #0: 0x0000000101783454 _datetime.cpython-313-darwin.so`delta_new + 188 frame #1: 0x0000000100704b60 Python`type_call + 96 frame #2: 0x000000010067ba34 Python`_PyObject_MakeTpCall + 120 frame #3: 0x00000001007aae3c Python`_PyEval_EvalFrameDefault + 30236 frame #4: 0x000000010067c900 Python`PyObject_CallOneArg + 112 frame #5: 0x000000010070f0a0 Python`slot_tp_finalize + 116 frame #6: 0x000000010070c3b4 Python`subtype_dealloc + 788 frame #7: 0x00000001006c378c Python`insertdict + 756 frame #8: 0x00000001006db2b0 Python`_PyModule_ClearDict + 660 frame #9: 0x000000010080a9a8 Python`finalize_modules + 1772 frame #10: 0x0000000100809a44 Python`_Py_Finalize + 264 frame #11: 0x0000000100837630 Python`Py_RunMain + 252 frame #12: 0x0000000100837ef8 Python`pymain_main + 304 frame #13: 0x0000000100837f98 Python`Py_BytesMain + 40 frame #14: 0x000000019cfcc274 dyld`start + 2840 (lldb) register read x20 x20 = 0x0000000000000000 (lldb) ``` ### Steps to reproduce the bug Run the script above, and observe the segfault. ### Expected behavior No segfault ### Environment info ``` % pip freeze datasets | grep -i datasets datasets==4.2.0 (venv) 0 ~/bug 14:58:06 % pip freeze tqdm | grep -i tqdm tqdm==4.67.1 (venv) 0 ~/bug 14:58:16 % python --version Python 3.13.7 ```
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7811/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7811/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7810
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7810/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7810/comments
https://api.github.com/repos/huggingface/datasets/issues/7810/events
https://github.com/huggingface/datasets/pull/7810
3,499,855,569
PR_kwDODunzps6s7wHa
7,810
fix conda deps
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7810). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T16:32:04Z
2025-10-09T16:35:15Z
2025-10-09T16:35:14Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7810/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7810/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7810.diff", "html_url": "https://github.com/huggingface/datasets/pull/7810", "merged_at": "2025-10-09T16:35:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/7810.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7810" }
true
https://api.github.com/repos/huggingface/datasets/issues/7809
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7809/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7809/comments
https://api.github.com/repos/huggingface/datasets/issues/7809/events
https://github.com/huggingface/datasets/pull/7809
3,499,811,179
PR_kwDODunzps6s7mwb
7,809
Set dev version
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7809). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T16:19:19Z
2025-10-09T16:22:12Z
2025-10-09T16:19:31Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7809/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7809/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7809.diff", "html_url": "https://github.com/huggingface/datasets/pull/7809", "merged_at": "2025-10-09T16:19:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/7809.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7809" }
true
https://api.github.com/repos/huggingface/datasets/issues/7808
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7808/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7808/comments
https://api.github.com/repos/huggingface/datasets/issues/7808/events
https://github.com/huggingface/datasets/pull/7808
3,499,779,993
PR_kwDODunzps6s7gBq
7,808
release: 4.2.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7808). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T16:10:53Z
2025-10-09T16:21:01Z
2025-10-09T16:11:08Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7808/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7808/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7808.diff", "html_url": "https://github.com/huggingface/datasets/pull/7808", "merged_at": "2025-10-09T16:11:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/7808.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7808" }
true
https://api.github.com/repos/huggingface/datasets/issues/7807
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7807/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7807/comments
https://api.github.com/repos/huggingface/datasets/issues/7807/events
https://github.com/huggingface/datasets/pull/7807
3,499,765,725
PR_kwDODunzps6s7c_U
7,807
typo
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7807). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T16:06:47Z
2025-10-09T16:16:31Z
2025-10-09T16:06:58Z
MEMBER
null
null
null
null
add an s to be consistent with pandas' on_bad_lines
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7807/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7807/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7807.diff", "html_url": "https://github.com/huggingface/datasets/pull/7807", "merged_at": "2025-10-09T16:06:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/7807.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7807" }
true
https://api.github.com/repos/huggingface/datasets/issues/7806
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7806/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7806/comments
https://api.github.com/repos/huggingface/datasets/issues/7806/events
https://github.com/huggingface/datasets/pull/7806
3,499,483,246
PR_kwDODunzps6s6gnr
7,806
Parquet: add `on_bad_file` argument to error/warn/skip bad files
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7806). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T14:41:46Z
2025-10-09T16:04:35Z
2025-10-09T16:04:33Z
MEMBER
null
null
null
null
```python from datasets import load_dataset on_bad_file = "error" # default # on_bad_file = "warn" # warn and skip bad file # on_bad_file = "skip" # skip bad file ds = load_dataset(parquet_dataset_id, on_bad_file=on_bad_file) ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7806/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7806/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7806.diff", "html_url": "https://github.com/huggingface/datasets/pull/7806", "merged_at": "2025-10-09T16:04:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/7806.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7806" }
true
https://api.github.com/repos/huggingface/datasets/issues/7805
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7805/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7805/comments
https://api.github.com/repos/huggingface/datasets/issues/7805/events
https://github.com/huggingface/datasets/pull/7805
3,499,286,947
PR_kwDODunzps6s52Ew
7,805
Less api calls when resolving data_files
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7805). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T13:53:06Z
2025-10-09T14:01:57Z
2025-10-09T14:01:56Z
MEMBER
null
null
null
null
There ~10 unnecessary `/api/datasets/user/dataset/revision`calls due to multithreading in data files resolution. I disabled multithreading, which was actually not useful anymore since `HfFileSystem` has been using caching for a while now.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 1, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7805/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7805/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7805.diff", "html_url": "https://github.com/huggingface/datasets/pull/7805", "merged_at": "2025-10-09T14:01:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/7805.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7805" }
true
https://api.github.com/repos/huggingface/datasets/issues/7804
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7804/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7804/comments
https://api.github.com/repos/huggingface/datasets/issues/7804/events
https://github.com/huggingface/datasets/issues/7804
3,498,534,596
I_kwDODunzps7Qh2bE
7,804
Support scientific data formats
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Please add the support for `Zarr`! That's what we use in the Bioimaging community. It is crucial, because raw upload of a *single* bio image can take _terrabytes in memory_!\n\nThe python library would be `bioio` or `zarr`:\n- [ ] Zarr: `bioio` or `zarr`\n\nSee a [Zarr example](https://ome.github.io/ome-ngff-validator/?source=https://uk1s3.embassy.ebi.ac.uk/bia-integrator-data/S-BIAD845/796b9fb8-f4ec-4c4b-bfc3-5cb00ccf19fe/796b9fb8-f4ec-4c4b-bfc3-5cb00ccf19fe.zarr)\n\ncc @joshmoore", "@stefanches7 `zarr` is already usable with the hf hub as an array store. See this example from the [docs](https://huggingface.co/docs/huggingface_hub/en/guides/hf_file_system):\n\n```python\nimport numpy as np\nimport zarr\n\nembeddings = np.random.randn(50000, 1000).astype(\"float32\")\n\n# Write an array to a repo\nwith zarr.open_group(\"hf://my-username/my-model-repo/array-store\", mode=\"w\") as root:\n foo = root.create_group(\"embeddings\")\n foobar = foo.zeros('experiment_0', shape=(50000, 1000), chunks=(10000, 1000), dtype='f4')\n foobar[:] = embeddings\n\n# Read an array from a repo\nwith zarr.open_group(\"hf://my-username/my-model-repo/array-store\", mode=\"r\") as root:\n first_row = root[\"embeddings/experiment_0\"][0]\n```\n\nIs there additional functionality that would not be covered by this?", "@cakiki I think some tiling capabilities, as well as metadata / labels handling. Consult ome-zarr doc here: https://ome-zarr.readthedocs.io/en/stable/python.html\nVisualization would be the cherry on the top. \n\ncc @joshmoore @lubianat @St3V0Bay: curious what you think", "zarr-specific dataset viewer would be very cool", "A support for BIDS it would be perfect, I think it's possible to do all the biosinal can be done with mne. There's a cool community for decoding brain signals, and now with EMG. The new META bracelet EMG is saving things in BIDS.\n\nI can help to interface, coding and try to make this happen. I am available at hugging face discord with the username aristimunha, if some 1-to-1 discuss it would be necessary :)", "@lhoestq , @cakiki , do you think we can make this happen?", "If you give me the OK, I'll create the PR to make everything for a Biosignal Reader logic, I already studied the nilabel PR :)", "That would be an amazing addition ! Feel free to ping me in your PR for review or if you have questions / if I can help", "@bruAristimunha @lhoestq I've recalled a gold of a resource for BIDS: https://openneuro.org/\n\nDo you think there is a data-easy way to make those visible here on HuggingFace? Afaik they use `datalad` to fetch the data. Maybe the best way is to leave OpenNeuro as-is, not connecting it to HuggingFace at all - just an idea I had spontaneously.", "I know an \"easy\" way to create interoperability with all biosignal datasets from OpenNeuro =) \n\nFor biosignal data, we can use [EEGDash](https://eegdash.org/) to create a Pytorch dataset, which automates fetch, lazy read, and converts to a pytorch dataset. \n\nI have a question about the best serialization for a Hugging Face dataset, but I can discuss it with some of you on Discord; my username is aristimunha.", "I can explain it publicly too, but I think a short 5-minute conversation would be better than many, many texts to explain the details.", "It's ok to have discussions in one place here (or in a separate issue if it's needed) - I also generally check github more often than discord ^^'", "Hi @bruAristimunha @lhoestq any way we could proceed on this?\nI see someone posted a Nifti vizualization PR: https://github.com/huggingface/datasets/pull/7874 - I think it would be a shame if we couldn't accompany that by a neat way to import BIDS Nifti!", "@stefanches7 author of #7874 here, would be open to expand the current support to BIDS as well after having a brief look. \nMaybe having a brief call over Discord (my username: TobiasPitters on the huggingface discord server) might help sorting things out, since I am not familiar with BIDS. So getting an understanding over test cases needed, etc. would be great!", "Hey!!\n\nFrom a bids perspective, I can provide full support for all biosignal types (EEG, iEEG, MEG, EMG). BIDS is a well-established contract format; I believe we can design something that supports the entire medical domain. I think it just requires a few details to be aligned.\n\nFrom my perspective, the tricky part is how to best adapt and serialize from the Hugging Face perspective.\n\nUnder the hood, for the biosignal part, I think I would use [mne](https://mne.tools/) for interoperability and [eegdash](https://eegdash.org/dataset_summary.html) to create the serialized dataset, but we can definitely discuss this further. I will ping you @CloseChoice on Discord.", "had a discussion with @neurolabusc and here's a quick wrap-up:\n - BIDS support would be huge (@bruAristimunha would be great if we could catch up on that)\n - DICOM support as well, but that might be harder due to a lot of variety in how headers are handled, vendor specifics etc. So to have a reliable pipeline to interact with whole folders of DICOM files (including metadata) would require a lot of work and a lot of testing. Therefore I set https://github.com/huggingface/datasets/pull/7835 back to draft mode. But there are tools that ease the way, especially https://github.com/ImagingDataCommons/highdicom (or potentially https://github.com/QIICR/dcmqi). \n - Getting users would help in order to understand what other formats/features are required therefore loading a bunch of open datasets to the hub using the new Nifti feature would be great. Some tutorials might help here as well.", "Hi @CloseChoice and @bruAristimunha, glad to meet you both! We could appoint a call; I am currently moving to a new job, so the time slots are limited, but let's connect over Discord and see what we could do.\n\n* BIDS: our hackathon team @zuazo @ekarrieta @lakshya16157 put up a BIDS format converter: https://huggingface.co/spaces/stefanches/OpenBIDSifier. Might be useful for imaging dataset conversion to BIDS.\n* DICOM support: cc @St3V0Bay, the author of DICOM support in CroissantML (https://github.com/mlcommons/croissant/pull/942)\n\ncc @nolden", "my username is aristimunha within the huggieng face discord to discuss more" ]
2025-10-09T10:18:24Z
2025-11-26T16:09:43Z
null
MEMBER
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
List of formats and libraries we can use to load the data in `datasets`: - [ ] DICOMs: pydicom - [x] NIfTIs: nibabel - [ ] WFDB: wfdb cc @zaRizk7 for viz Feel free to comment / suggest other formats and libs you'd like to see or to share your interest in one of the mentioned format
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 5, "hooray": 4, "laugh": 0, "rocket": 0, "total_count": 10, "url": "https://api.github.com/repos/huggingface/datasets/issues/7804/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7804/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7803
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7803/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7803/comments
https://api.github.com/repos/huggingface/datasets/issues/7803/events
https://github.com/huggingface/datasets/pull/7803
3,498,395,879
PR_kwDODunzps6s2zyO
7,803
More Parquet streaming docs
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7803). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-09T09:39:11Z
2025-10-09T10:01:46Z
2025-10-09T10:01:43Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7803/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7803/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7803.diff", "html_url": "https://github.com/huggingface/datasets/pull/7803", "merged_at": "2025-10-09T10:01:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/7803.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7803" }
true
https://api.github.com/repos/huggingface/datasets/issues/7802
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7802/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7802/comments
https://api.github.com/repos/huggingface/datasets/issues/7802/events
https://github.com/huggingface/datasets/issues/7802
3,497,454,119
I_kwDODunzps7Qduon
7,802
[Docs] Missing documentation for `Dataset.from_dict`
{ "avatar_url": "https://avatars.githubusercontent.com/u/69421545?v=4", "events_url": "https://api.github.com/users/aaronshenhao/events{/privacy}", "followers_url": "https://api.github.com/users/aaronshenhao/followers", "following_url": "https://api.github.com/users/aaronshenhao/following{/other_user}", "gists_url": "https://api.github.com/users/aaronshenhao/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/aaronshenhao", "id": 69421545, "login": "aaronshenhao", "node_id": "MDQ6VXNlcjY5NDIxNTQ1", "organizations_url": "https://api.github.com/users/aaronshenhao/orgs", "received_events_url": "https://api.github.com/users/aaronshenhao/received_events", "repos_url": "https://api.github.com/users/aaronshenhao/repos", "site_admin": false, "starred_url": "https://api.github.com/users/aaronshenhao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/aaronshenhao/subscriptions", "type": "User", "url": "https://api.github.com/users/aaronshenhao", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "I'd like to work on this documentation issue.", "Hi I'd like to work on this. I can see the docstring is already in the code. \nCould you confirm:\n1. Is this still available?\n2. Should I add this to the main_classes.md file, or is there a specific \n documentation file I should update?\n3. Are there any formatting guidelines I should follow?\n\nI'm new to contributing but eager to learn!" ]
2025-10-09T02:54:41Z
2025-10-19T16:09:33Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
Documentation link: https://huggingface.co/docs/datasets/en/package_reference/main_classes Link to method (docstring present): https://github.com/huggingface/datasets/blob/6f2502c5a026caa89839713f6f7c8b958e5e83eb/src/datasets/arrow_dataset.py#L1029 The docstring is present for the function, but seems missing from the official documentation for the `Dataset` class on HuggingFace. The method in question: ```python @classmethod def from_dict( cls, mapping: dict, features: Optional[Features] = None, info: Optional[DatasetInfo] = None, split: Optional[NamedSplit] = None, ) -> "Dataset": """ Convert `dict` to a `pyarrow.Table` to create a [`Dataset`]. Important: a dataset created with from_dict() lives in memory and therefore doesn't have an associated cache directory. This may change in the future, but in the meantime if you want to reduce memory usage you should write it back on disk and reload using e.g. save_to_disk / load_from_disk. Args: mapping (`Mapping`): Mapping of strings to Arrays or Python lists. features ([`Features`], *optional*): Dataset features. info (`DatasetInfo`, *optional*): Dataset information, like description, citation, etc. split (`NamedSplit`, *optional*): Name of the dataset split. Returns: [`Dataset`] """ ```
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7802/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7802/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7801
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7801/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7801/comments
https://api.github.com/repos/huggingface/datasets/issues/7801/events
https://github.com/huggingface/datasets/pull/7801
3,496,388,063
PR_kwDODunzps6swITn
7,801
Add parquet scan options and docs
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7801). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-08T18:04:52Z
2025-10-09T07:55:58Z
2025-10-09T07:55:56Z
MEMBER
null
null
null
null
I added scan options, useful to control buffering and caching when streaming and docs, including how to select a subset of columns and apply filters
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7801/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7801/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7801.diff", "html_url": "https://github.com/huggingface/datasets/pull/7801", "merged_at": "2025-10-09T07:55:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/7801.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7801" }
true
https://api.github.com/repos/huggingface/datasets/issues/7800
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7800/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7800/comments
https://api.github.com/repos/huggingface/datasets/issues/7800/events
https://github.com/huggingface/datasets/pull/7800
3,494,747,495
PR_kwDODunzps6sqkmT
7,800
Fix polars cast column image
{ "avatar_url": "https://avatars.githubusercontent.com/u/31857876?v=4", "events_url": "https://api.github.com/users/CloseChoice/events{/privacy}", "followers_url": "https://api.github.com/users/CloseChoice/followers", "following_url": "https://api.github.com/users/CloseChoice/following{/other_user}", "gists_url": "https://api.github.com/users/CloseChoice/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CloseChoice", "id": 31857876, "login": "CloseChoice", "node_id": "MDQ6VXNlcjMxODU3ODc2", "organizations_url": "https://api.github.com/users/CloseChoice/orgs", "received_events_url": "https://api.github.com/users/CloseChoice/received_events", "repos_url": "https://api.github.com/users/CloseChoice/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CloseChoice/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CloseChoice/subscriptions", "type": "User", "url": "https://api.github.com/users/CloseChoice", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The Image() type is set to have a storage of `string` for \"path\" and not `large_string`. Therefore while your change does work to do the conversion, it can create issues in other places. For example I'm pretty sure you wouldn't be able to concatenate the resulting dataset with a dataset with Image() using `string`.\r\n\r\nMaybe we can convert `large_string` data to `string` somehow to make this work ?", "@lhoestq thanks for the review. Just to be thorough I checked the concat example and this seems to work:\r\n```python\r\nimport sys\r\nfrom pathlib import Path\r\nsys.path.insert(0, str(Path(__file__).parent.parent / \"src\"))\r\n\r\nimport pandas as pd\r\nimport polars as pl\r\nfrom datasets import Dataset, Image, concatenate_datasets\r\nimport pyarrow as pa\r\n\r\nimage_path = \"tests/features/data/test_image_rgb.jpg\"\r\n\r\n\r\ndf_pl = pl.DataFrame({\"image\": [image_path]})\r\ndset_pl = Dataset.from_polars(df_pl).cast_column(\"image\", Image())\r\n\r\n\r\ndf_pd = pd.DataFrame({\"image\": [image_path]})\r\ndset_pd = Dataset.from_pandas(df_pd).cast_column(\"image\", Image())\r\n\r\n\r\nconcatenated = concatenate_datasets([dset_pl, dset_pd])\r\nprint(concatenated._data)\r\n```\r\noutputs:\r\n```bash\r\nConcatenationTable\r\nimage: struct<bytes: binary, path: string>\r\n child 0, bytes: binary\r\n child 1, path: string\r\n----\r\nimage: [\r\n -- is_valid: all not null\r\n -- child 0 type: binary\r\n[null]\r\n -- child 1 type: string\r\n[\"tests/features/data/test_image_rgb.jpg\"],\r\n -- is_valid: all not null\r\n -- child 0 type: binary\r\n[null]\r\n -- child 1 type: string\r\n[\"tests/features/data/test_image_rgb.jpg\"]]\r\n```\r\n(not quite sure though if this is a really what you meant). I agree that there could be pro a lot of problems if we rely on implicit conversion therefore I updated the PR. I also checked the exception handling locally and it works, am unsure though if we want to create such large objects in the CI, if desired I can add a test for that.", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7800). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "Apologies @lhoestq @CloseChoice , I unintentionally reverted this PR earlier. Leaving it as is." ]
2025-10-08T10:01:18Z
2025-10-18T13:48:37Z
2025-10-13T14:39:47Z
CONTRIBUTOR
null
null
null
null
Fixes #7765 The problem here is that polars uses pyarrow large_string for images, while pandas and others just use the string type. This PR solves that and adds a test. ```python import polars as pl from datasets import Dataset import pandas as pd import pyarrow as pa from pathlib import Path shared_datadir = Path("tests/features/data") image_path = str(shared_datadir / "test_image_rgb.jpg") # Load via polars df_polars = pl.DataFrame({"image_path": [image_path]}) dataset_polars = Dataset.from_polars(df_polars) print("Polars DF is large string:", pa.types.is_large_string(df_polars.to_arrow().schema[0].type)) print("Polars DF is string:", pa.types.is_string(df_polars.to_arrow().schema[0].type)) # Load via pandas df_pandas = pd.DataFrame({"image_path": [image_path]}) dataset_pandas = Dataset.from_pandas(df_pandas) arrow_table_pd = pa.Table.from_pandas(df_pandas) print("Pandas DF is large string", pa.types.is_large_string(arrow_table_pd.schema[0].type)) print("Pandas DF is string", pa.types.is_string(arrow_table_pd.schema[0].type)) ``` Outputs: ```bash Polars DF is large string: True Polars DF is string: False Pandas DF is large string False Pandas DF is string True ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7800/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7800/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7800.diff", "html_url": "https://github.com/huggingface/datasets/pull/7800", "merged_at": "2025-10-13T14:39:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/7800.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7800" }
true
https://api.github.com/repos/huggingface/datasets/issues/7799
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7799/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7799/comments
https://api.github.com/repos/huggingface/datasets/issues/7799/events
https://github.com/huggingface/datasets/pull/7799
3,487,791,741
PR_kwDODunzps6sTJKA
7,799
Define CI future
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7799). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-10-06T15:15:45Z
2025-10-07T14:30:21Z
2025-10-07T14:30:19Z
MEMBER
null
null
null
null
this should fix the CI which currently uses transformers on 3.9 while it's now unsupported
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7799/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7799/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7799.diff", "html_url": "https://github.com/huggingface/datasets/pull/7799", "merged_at": "2025-10-07T14:30:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/7799.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7799" }
true
https://api.github.com/repos/huggingface/datasets/issues/7798
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7798/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7798/comments
https://api.github.com/repos/huggingface/datasets/issues/7798/events
https://github.com/huggingface/datasets/issues/7798
3,484,470,782
I_kwDODunzps7PsM3-
7,798
Audio dataset is not decoding on 4.1.1
{ "avatar_url": "https://avatars.githubusercontent.com/u/61390950?v=4", "events_url": "https://api.github.com/users/thewh1teagle/events{/privacy}", "followers_url": "https://api.github.com/users/thewh1teagle/followers", "following_url": "https://api.github.com/users/thewh1teagle/following{/other_user}", "gists_url": "https://api.github.com/users/thewh1teagle/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thewh1teagle", "id": 61390950, "login": "thewh1teagle", "node_id": "MDQ6VXNlcjYxMzkwOTUw", "organizations_url": "https://api.github.com/users/thewh1teagle/orgs", "received_events_url": "https://api.github.com/users/thewh1teagle/received_events", "repos_url": "https://api.github.com/users/thewh1teagle/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thewh1teagle/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thewh1teagle/subscriptions", "type": "User", "url": "https://api.github.com/users/thewh1teagle", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Previously (datasets<=3.6.0), audio columns were decoded automatically when accessing a row. Now, for performance reasons, audio decoding is lazy by default: you just see the file path unless you explicitly cast the column to Audio.\n\nHere’s the fix (following the current [datasets audio docs](https://huggingface.co/docs/datasets/en/audio_load)\n):\n\n```\nfrom datasets import load_dataset, Audio\n\ndataset = load_dataset(\"MrDragonFox/Elise\", split=\"train\")\n\n# Explicitly decode the audio column\ndataset = dataset.cast_column(\"audio\", Audio(sampling_rate=16_000))\n\nprint(dataset[0][\"audio\"])\n# {'path': '...', 'array': array([...], dtype=float32), 'sampling_rate': 16000}\n```", "@haitam03-yo's comment is right that the data is not decoded by default anymore indeed, but here is how it works in practice now:\n\nFrom `datasets` v4, audio data are read as [AudioDecoder](https://meta-pytorch.org/torchcodec/0.4/generated/torchcodec.decoders.AudioDecoder.html) objects from torchcodec. This doesn't decode the data by default, but you can call `audio.get_all_samples()` to decode the audio.\n\nSee the documentation on how to process audio data here: https://huggingface.co/docs/datasets/audio_process", "To resolve this, you need to explicitly cast the audio column to the Audio feature. This will decode the audio data and make it accessible as an array. Here is the corrected code snippet\n\n\nfrom datasets import load_dataset, Audio\n\n# Load your dataset\ndataset = load_dataset(\"MrDragonFox/Elise\", split=\"train\")\n\n# Explicitly cast the 'audio' column to the Audio feature\ndataset = dataset.cast_column(\"audio\", Audio(sampling_rate=16_000))\n\n# Now you can access the decoded audio array\nprint(dataset[0][\"audio\"])\n\nBy adding the cast_column step, you are telling the datasets library to decode the audio data with the specified sampling rate, and you will then be able to access the audio array as you were used to in previous versions." ]
2025-10-05T06:37:50Z
2025-10-06T14:07:55Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug The audio column remain as non-decoded objects even when accessing them. ```python dataset = load_dataset("MrDragonFox/Elise", split = "train") dataset[0] # see that it doesn't show 'array' etc... ``` Works fine with `datasets==3.6.0` Followed the docs in - https://huggingface.co/docs/datasets/en/audio_load ### Steps to reproduce the bug ```python dataset = load_dataset("MrDragonFox/Elise", split = "train") dataset[0] # see that it doesn't show 'array' etc... ``` ### Expected behavior It should decode when accessing the elemenet ### Environment info 4.1.1 ubuntu 22.04 Related - https://github.com/huggingface/datasets/issues/7707
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7798/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7798/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7797
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7797/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7797/comments
https://api.github.com/repos/huggingface/datasets/issues/7797/events
https://github.com/huggingface/datasets/pull/7797
3,473,011,621
PR_kwDODunzps6rhtf_
7,797
Datasets: Add WMT21 & WMT22 loaders (basic TSV loaders, sample data, tests)
{ "avatar_url": "https://avatars.githubusercontent.com/u/164366940?v=4", "events_url": "https://api.github.com/users/tanisha-samant/events{/privacy}", "followers_url": "https://api.github.com/users/tanisha-samant/followers", "following_url": "https://api.github.com/users/tanisha-samant/following{/other_user}", "gists_url": "https://api.github.com/users/tanisha-samant/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/tanisha-samant", "id": 164366940, "login": "tanisha-samant", "node_id": "U_kgDOCcwKXA", "organizations_url": "https://api.github.com/users/tanisha-samant/orgs", "received_events_url": "https://api.github.com/users/tanisha-samant/received_events", "repos_url": "https://api.github.com/users/tanisha-samant/repos", "site_admin": false, "starred_url": "https://api.github.com/users/tanisha-samant/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tanisha-samant/subscriptions", "type": "User", "url": "https://api.github.com/users/tanisha-samant", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "closing since datasets should be added on https://huggingface.co directly" ]
2025-10-01T10:46:01Z
2025-10-10T15:33:25Z
2025-10-10T15:33:25Z
NONE
null
null
null
null
- Implemented TSV-based dataset loaders: - WMT21Dataset (local_datasets/wmt21/wmt21_dataset.py) - WMT22Dataset (local_datasets/wmt22/wmt22_dataset.py) These classes load source-target pairs from .tsv files for train, validation, and test splits. - Created sample dummy data for both datasets: - dummy_data/train.tsv, dummy_data/validation.tsv, dummy_data/test.tsv - Includes a few realistic example lines to allow CI and local tests to pass without downloading full datasets. - Added automated tests for robust validation: - tests/test_wmt21.py and tests/test_wmt22.py - Checks that all splits load correctly, empty lines are ignored, and the number of examples matches the number of lines in the .tsv files. - Edge cases handled: empty lines, malformed lines, extra tabs. - Added README.md files for both datasets: - Provides dataset structure, usage instructions, and placeholders for citation & license information. - Ensures that other developers and reviewers can understand dataset usage immediately. - Ensured easy local testing: - Load datasets programmatically using WMT21Dataset / WMT22Dataset. - Verified train/validation/test splits are correctly returned as Python dictionaries of Dataset objects. -Provides initial support for WMT21 and WMT22 NLP/translation experiments. -Allows contributors and reviewers to test dataset loading locally or in CI without downloading large datasets. -Serves as a template to extend to other WMT datasets in the future. Testing instructions: ``` # Activate your environment pytest tests/test_wmt21.py -v pytest tests/test_wmt22.py -v ``` ``` from local_datasets.wmt21.wmt21_dataset import WMT21Dataset from local_datasets.wmt22.wmt22_dataset import WMT22Dataset # WMT21 wmt21 = WMT21Dataset("local_datasets/wmt21/dummy_data") ds21 = wmt21.load() print(ds21["train"][0]) print(ds21["validation"][0]) print(ds21["test"][0]) # WMT22 wmt22 = WMT22Dataset("local_datasets/wmt22/dummy_data") ds22 = wmt22.load() print(ds22["train"][0]) print(ds22["validation"][0]) print(ds22["test"][0]) ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7797/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7797/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7797.diff", "html_url": "https://github.com/huggingface/datasets/pull/7797", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7797.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7797" }
true
https://api.github.com/repos/huggingface/datasets/issues/7796
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7796/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7796/comments
https://api.github.com/repos/huggingface/datasets/issues/7796/events
https://github.com/huggingface/datasets/pull/7796
3,470,616,799
PR_kwDODunzps6rZjrW
7,796
Docs: fix typo, improve readability, add code comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/164366940?v=4", "events_url": "https://api.github.com/users/tanisha-samant/events{/privacy}", "followers_url": "https://api.github.com/users/tanisha-samant/followers", "following_url": "https://api.github.com/users/tanisha-samant/following{/other_user}", "gists_url": "https://api.github.com/users/tanisha-samant/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/tanisha-samant", "id": 164366940, "login": "tanisha-samant", "node_id": "U_kgDOCcwKXA", "organizations_url": "https://api.github.com/users/tanisha-samant/orgs", "received_events_url": "https://api.github.com/users/tanisha-samant/received_events", "repos_url": "https://api.github.com/users/tanisha-samant/repos", "site_admin": false, "starred_url": "https://api.github.com/users/tanisha-samant/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tanisha-samant/subscriptions", "type": "User", "url": "https://api.github.com/users/tanisha-samant", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-09-30T18:34:16Z
2025-10-10T18:44:12Z
null
NONE
null
null
null
null
What I did: - Fixed a small typo in README to improve clarity - Fixed repeated word "frameworks frameworks" - Split long paragraphs into shorter sentences for readability - Added # Example comments before code blocks for clarity Why: - Helps new users avoid confusion How I tested: - Checked locally in Markdown preview
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7796/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7796/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7796.diff", "html_url": "https://github.com/huggingface/datasets/pull/7796", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7796.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7796" }
true
https://api.github.com/repos/huggingface/datasets/issues/7795
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7795/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7795/comments
https://api.github.com/repos/huggingface/datasets/issues/7795/events
https://github.com/huggingface/datasets/pull/7795
3,463,990,654
PR_kwDODunzps6rDEce
7,795
Add pyarrow's binary view to features
{ "avatar_url": "https://avatars.githubusercontent.com/u/6834061?v=4", "events_url": "https://api.github.com/users/delta003/events{/privacy}", "followers_url": "https://api.github.com/users/delta003/followers", "following_url": "https://api.github.com/users/delta003/following{/other_user}", "gists_url": "https://api.github.com/users/delta003/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/delta003", "id": 6834061, "login": "delta003", "node_id": "MDQ6VXNlcjY4MzQwNjE=", "organizations_url": "https://api.github.com/users/delta003/orgs", "received_events_url": "https://api.github.com/users/delta003/received_events", "repos_url": "https://api.github.com/users/delta003/repos", "site_admin": false, "starred_url": "https://api.github.com/users/delta003/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/delta003/subscriptions", "type": "User", "url": "https://api.github.com/users/delta003", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "@lhoestq 🙏 ", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7795). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-09-29T09:12:55Z
2025-10-10T16:04:21Z
2025-10-10T16:04:21Z
CONTRIBUTOR
null
null
null
null
Basically https://github.com/huggingface/datasets/pull/7718 just for binary view instead of string view
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 2, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/7795/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7795/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7795.diff", "html_url": "https://github.com/huggingface/datasets/pull/7795", "merged_at": "2025-10-10T16:04:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/7795.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7795" }
true
https://api.github.com/repos/huggingface/datasets/issues/7794
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7794/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7794/comments
https://api.github.com/repos/huggingface/datasets/issues/7794/events
https://github.com/huggingface/datasets/pull/7794
3,460,793,966
PR_kwDODunzps6q4XyU
7,794
Fix nested data conversions error in parquet loading (fixes #7793)
{ "avatar_url": "https://avatars.githubusercontent.com/u/41635755?v=4", "events_url": "https://api.github.com/users/Aishwarya0811/events{/privacy}", "followers_url": "https://api.github.com/users/Aishwarya0811/followers", "following_url": "https://api.github.com/users/Aishwarya0811/following{/other_user}", "gists_url": "https://api.github.com/users/Aishwarya0811/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Aishwarya0811", "id": 41635755, "login": "Aishwarya0811", "node_id": "MDQ6VXNlcjQxNjM1NzU1", "organizations_url": "https://api.github.com/users/Aishwarya0811/orgs", "received_events_url": "https://api.github.com/users/Aishwarya0811/received_events", "repos_url": "https://api.github.com/users/Aishwarya0811/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Aishwarya0811/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Aishwarya0811/subscriptions", "type": "User", "url": "https://api.github.com/users/Aishwarya0811", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Unfortunately, I'm running into this error:\r\n```\r\n~/scratch » uv run python test_hf.py \r\nResolving data files: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 42/42 [00:00<00:00, 149.18it/s]\r\nResolving data files: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 102/102 [00:00<00:00, 317608.77it/s]\r\nDownloading data: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 102/102 [00:00<00:00, 337.74files/s]\r\nGenerating public split: 77%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▌ | 5506/7179 [00:19<00:10, 156.43 examples/s]Using fallback for nested data in file '/Users/neev/.cache/huggingface/hub/datasets--metr-evals--malt-public/snapshots/86f8dcf09084458117b16a8f83256097d27fe35b/irrelevant_detail/public-00081-of-00102.parquet': Nested data conversions not implemented for chunked array outputs\r\nGenerating public split: 77%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▌ | 5506/7179 [00:21<00:06, 256.72 examples/s]\r\nTraceback (most recent call last):\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/packaged_modules/parquet/parquet.py\", line 134, in _generate_tables\r\n for batch_idx, record_batch in enumerate(\r\n ~~~~~~~~~^\r\n parquet_fragment.to_batches(\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n ...<5 lines>...\r\n )\r\n ^\r\n ):\r\n ^\r\n File \"pyarrow/_dataset.pyx\", line 3904, in _iterator\r\n File \"pyarrow/_dataset.pyx\", line 3494, in pyarrow._dataset.TaggedRecordBatchIterator.__next__\r\n File \"pyarrow/error.pxi\", line 155, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow/error.pxi\", line 92, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowNotImplementedError: Nested data conversions not implemented for chunked array outputs\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py\", line 1815, in _prepare_split_single\r\n for _, table in generator:\r\n ^^^^^^^^^\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/packaged_modules/parquet/parquet.py\", line 152, in _generate_tables\r\n full_table = parquet_fragment.to_table(\r\n columns=self.config.columns,\r\n filter=filter_expr,\r\n )\r\n File \"pyarrow/_dataset.pyx\", line 1743, in pyarrow._dataset.Fragment.to_table\r\n File \"pyarrow/_dataset.pyx\", line 3939, in pyarrow._dataset.Scanner.to_table\r\n File \"pyarrow/error.pxi\", line 155, in pyarrow.lib.pyarrow_internal_check_status\r\n File \"pyarrow/error.pxi\", line 92, in pyarrow.lib.check_status\r\npyarrow.lib.ArrowNotImplementedError: Nested data conversions not implemented for chunked array outputs\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nTraceback (most recent call last):\r\n File \"/Users/neev/scratch/test_hf.py\", line 3, in <module>\r\n ds = datasets.load_dataset(path=\"metr-evals/malt-public\", name=\"irrelevant_detail\")\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/load.py\", line 1412, in load_dataset\r\n builder_instance.download_and_prepare(\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^\r\n download_config=download_config,\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n ...<3 lines>...\r\n storage_options=storage_options,\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n )\r\n ^\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py\", line 894, in download_and_prepare\r\n self._download_and_prepare(\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~^\r\n dl_manager=dl_manager,\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\n ...<2 lines>...\r\n **download_and_prepare_kwargs,\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n )\r\n ^\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py\", line 970, in _download_and_prepare\r\n self._prepare_split(split_generator, **prepare_split_kwargs)\r\n ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py\", line 1702, in _prepare_split\r\n for job_id, done, content in self._prepare_split_single(\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~^\r\n gen_kwargs=gen_kwargs, job_id=job_id, **_prepare_split_args\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n ):\r\n ^\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py\", line 1858, in _prepare_split_single\r\n raise DatasetGenerationError(\"An error occurred while generating the dataset\") from e\r\ndatasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset\r\n```", "Also the gated dataset has automatic approval so you should feel free to sign in and test if you'd like!", "hi @neevparikh I've updated the fix based on your feedback. The new approach uses row group reading as a fallback when both to_batches() and to_table() fail. I've successfully tested it with an actual file from your dataset and it loads correctly. Could you test the updated version?\r\n\r\n", "Now we're failing with this error:\r\n\r\n```Resolving data files: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 42/42 [00:00<00:00, 79.30it/s]\r\nResolving data files: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 102/102 [00:00<00:00, 646252.28it/s]\r\nDownloading data: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 102/102 [00:00<00:00, 781.32files/s]\r\nGenerating public split: 77%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▌ | 5506/7179 [00:23<00:10, 156.37 examples/s]Using fallback for nested data in file '/Users/neev/.cache/huggingface/hub/datasets--metr-evals--malt-public/snapshots/86f8dcf09084458117b16a8f83256097d27fe35b/irrelevant_detail/public-00081-of-00102.parquet': Nested data conversions not implemented for chunked array outputs\r\nSkipping row group 0 due to nested data issues: Nested data conversions not implemented for chunked array outputs\r\nCould not read any row groups from file '/Users/neev/.cache/huggingface/hub/datasets--metr-evals--malt-public/snapshots/86f8dcf09084458117b16a8f83256097d27fe35b/irrelevant_detail/public-00081-of-00102.parquet'\r\nGenerating public split: 99%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▏ | 7099/7179 [00:38<00:00, 182.59 examples/s]\r\nTraceback (most recent call last):\r\n File \"/Users/neev/scratch/test_hf.py\", line 3, in <module>\r\n ds = datasets.load_dataset(\r\n path=\"metr-evals/malt-public\",\r\n name=\"irrelevant_detail\",\r\n )\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/load.py\", line 1412, in load_dataset\r\n builder_instance.download_and_prepare(\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^\r\n download_config=download_config,\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n ...<3 lines>...\r\n storage_options=storage_options,\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n )\r\n ^\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py\", line 894, in download_and_prepare\r\n self._download_and_prepare(\r\n ~~~~~~~~~~~~~~~~~~~~~~~~~~^\r\n dl_manager=dl_manager,\r\n ^^^^^^^^^^^^^^^^^^^^^^\r\n ...<2 lines>...\r\n **download_and_prepare_kwargs,\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n )\r\n ^\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py\", line 988, in _download_and_prepare\r\n verify_splits(self.info.splits, split_dict)\r\n ~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/utils/info_utils.py\", line 77, in verify_splits\r\n raise NonMatchingSplitsSizesError(str(bad_splits))\r\ndatasets.exceptions.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='public', num_bytes=25417866585, num_examples=7179, shard_lengths=None, dataset_name=None), 'recorded': SplitInfo(name='public', num_bytes=22946940147, num_examples=7099, shard_lengths=[300, 240, 180, 300, 600, 779, 359, 358, 239, 80, 80, 239, 79, 80, 159, 239, 399, 239, 398, 159, 159, 80, 80, 398, 80, 637, 80, 79], dataset_name='malt-public')}]```", "it seems to me that we dropped the ones we couldn't read?", "@Aishwarya0811 let me know if there's helpful things here I can do?" ]
2025-09-27T22:04:13Z
2025-10-01T16:56:20Z
null
NONE
null
null
null
null
Fixes #7793 ## Problem Loading datasets with deeply nested structures (like `metr-evals/malt-public`) fails with: ArrowNotImplementedError: Nested data conversions not implemented for chunked array outputs This occurs when parquet files contain nested data (lists, structs, maps) that exceed PyArrow's 16MB chunk limit. ## Root Cause PyArrow's C++ implementation explicitly rejects nested data conversions when data is split across multiple chunks. The limitation exists in the `WrapIntoListArray` function where repetition levels cannot be reconstructed across chunk boundaries. ## Solution - **Fallback mechanism**: Catches the specific PyArrow error and switches to non-chunked reading - **Selective optimization**: Only combines chunks for problematic nested columns to minimize memory impact - **Manual batching**: Maintains batching behavior even in fallback mode - **Backward compatibility**: Zero impact on existing datasets ## Implementation Details - Added `_is_nested_type()` helper to detect nested PyArrow types - Added `_handle_nested_chunked_conversion()` for selective chunk combining - Modified `_generate_tables()` to catch and handle the specific error - Preserves all existing error handling and logging ## Testing - [x] No regressions: Normal parquet datasets continue working - [x] Code follows existing patterns in the datasets codebase - [x] tested by original reporter (gated dataset access needed) **Note**: This fix is based on thorough research of PyArrow limitations and similar issues in the ecosystem. While we cannot test with the original dataset due to access restrictions, the implementation follows established patterns for handling this PyArrow limitation. ## Request for Testing @neevparikh Could you please test this fix with your original failing dataset? The implementation should resolve the nested data conversion error you encountered.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7794/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7794/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7794.diff", "html_url": "https://github.com/huggingface/datasets/pull/7794", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7794.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7794" }
true
https://api.github.com/repos/huggingface/datasets/issues/7793
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7793/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7793/comments
https://api.github.com/repos/huggingface/datasets/issues/7793/events
https://github.com/huggingface/datasets/issues/7793
3,459,496,971
I_kwDODunzps7OM7wL
7,793
Cannot load dataset, fails with nested data conversions not implemented for chunked array outputs
{ "avatar_url": "https://avatars.githubusercontent.com/u/41182432?v=4", "events_url": "https://api.github.com/users/neevparikh/events{/privacy}", "followers_url": "https://api.github.com/users/neevparikh/followers", "following_url": "https://api.github.com/users/neevparikh/following{/other_user}", "gists_url": "https://api.github.com/users/neevparikh/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/neevparikh", "id": 41182432, "login": "neevparikh", "node_id": "MDQ6VXNlcjQxMTgyNDMy", "organizations_url": "https://api.github.com/users/neevparikh/orgs", "received_events_url": "https://api.github.com/users/neevparikh/received_events", "repos_url": "https://api.github.com/users/neevparikh/repos", "site_admin": false, "starred_url": "https://api.github.com/users/neevparikh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neevparikh/subscriptions", "type": "User", "url": "https://api.github.com/users/neevparikh", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Hey @neevparikh,\nThanks for reporting this! I can reproduce the issue and have identified the root cause.\nProblem: The metr-evals/malt-public dataset contains deeply nested conversation data that exceeds PyArrow's 16MB chunk limit. When PyArrow tries to read it in chunks, it hits a fundamental limitation: \"Nested data conversions not implemented for chunked array outputs\".\nRoot Cause: Your dataset has large nested arrays (conversation trees with 4k-87k elements) that get automatically chunked by PyArrow, but the nested data conversion logic can't handle repetition levels across chunk boundaries\n I'm preparing a PR that adds a fallback mechanism to the parquet reader. When this specific error occurs, it will:\n\nDetect the nested data issue\nCombine chunks selectively for problematic columns\nContinue processing normally\n\nThis maintains backward compatibility while fixing the issue for nested datasets like yours.\nWorkaround (if you need immediate access): Try loading with smaller batch sizes:\npythonds = datasets.load_dataset(\"metr-evals/malt-public\", name=\"irrelevant_detail\", \n download_config=datasets.DownloadConfig(\n parquet_batch_size=1000\n ))" ]
2025-09-27T01:03:12Z
2025-09-27T21:35:31Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Describe the bug Hi! When I load this dataset, it fails with a pyarrow error. I'm using datasets 4.1.1, though I also see this with datasets 4.1.2 To reproduce: ``` import datasets ds = datasets.load_dataset(path="metr-evals/malt-public", name="irrelevant_detail") ``` Error: ``` Traceback (most recent call last): File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py", line 1815, in _prepare_split_single for _, table in generator: ^^^^^^^^^ File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/packaged_modules/parquet/parquet.py", line 93, in _generate_tables for batch_idx, record_batch in enumerate( ~~~~~~~~~^ parquet_fragment.to_batches( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ...<5 lines>... ) ^ ): ^ File "pyarrow/_dataset.pyx", line 3904, in _iterator File "pyarrow/_dataset.pyx", line 3494, in pyarrow._dataset.TaggedRecordBatchIterator.__next__ File "pyarrow/error.pxi", line 155, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Nested data conversions not implemented for chunked array outputs The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Users/neev/scratch/test_hf.py", line 3, in <module> ds = datasets.load_dataset(path="metr-evals/malt-public", name="irrelevant_detail") File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/load.py", line 1412, in load_dataset builder_instance.download_and_prepare( ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^ download_config=download_config, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ...<3 lines>... storage_options=storage_options, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py", line 894, in download_and_prepare self._download_and_prepare( ~~~~~~~~~~~~~~~~~~~~~~~~~~^ dl_manager=dl_manager, ^^^^^^^^^^^^^^^^^^^^^^ ...<2 lines>... **download_and_prepare_kwargs, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ) ^ File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py", line 970, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py", line 1702, in _prepare_split for job_id, done, content in self._prepare_split_single( ~~~~~~~~~~~~~~~~~~~~~~~~~~^ gen_kwargs=gen_kwargs, job_id=job_id, **_prepare_split_args ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ): ^ File "/Users/neev/scratch/.venv/lib/python3.13/site-packages/datasets/builder.py", line 1858, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset ``` ### Steps to reproduce the bug To reproduce: ``` import datasets ds = datasets.load_dataset(path="metr-evals/malt-public", name="irrelevant_detail") ``` ### Expected behavior The dataset loads ### Environment info Datasets: 4.1.1 Python: 3.13 Platform: Macos
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7793/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7793/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7792
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7792/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7792/comments
https://api.github.com/repos/huggingface/datasets/issues/7792/events
https://github.com/huggingface/datasets/issues/7792
3,456,802,210
I_kwDODunzps7OCp2i
7,792
Concatenate IterableDataset instances and distribute underlying shards in a RoundRobin manner
{ "avatar_url": "https://avatars.githubusercontent.com/u/13559010?v=4", "events_url": "https://api.github.com/users/LTMeyer/events{/privacy}", "followers_url": "https://api.github.com/users/LTMeyer/followers", "following_url": "https://api.github.com/users/LTMeyer/following{/other_user}", "gists_url": "https://api.github.com/users/LTMeyer/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/LTMeyer", "id": 13559010, "login": "LTMeyer", "node_id": "MDQ6VXNlcjEzNTU5MDEw", "organizations_url": "https://api.github.com/users/LTMeyer/orgs", "received_events_url": "https://api.github.com/users/LTMeyer/received_events", "repos_url": "https://api.github.com/users/LTMeyer/repos", "site_admin": false, "starred_url": "https://api.github.com/users/LTMeyer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LTMeyer/subscriptions", "type": "User", "url": "https://api.github.com/users/LTMeyer", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[ "# With `datasets.Dataset`\n\nHere is an small script that shows the distribution differences of samples between `interleave_datasets`, `concatenate_datasets` and `concatenate_datasets` + shuffling.\n\n```python\nimport datasets as hf_datasets\n\ndef gen(dataset: int, n_samples: int):\n for i in range(n_samples):\n yield {\"dataset\": dataset, \"sample\": i}\n\nds_1 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 0, \"n_samples\": 2})\nds_2 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 1, \"n_samples\": 1})\nds_3 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 2, \"n_samples\": 3})\n\nn_workers = 3\nprint(f\"Simulate run with {n_workers} workers\")\n\nprint(\"Interleave datasets\")\nfor w in range(n_workers):\n ds_interleave = hf_datasets.interleave_datasets([ds_1, ds_2, ds_3]).shard(n_workers, w)\n for i, sample in enumerate(ds_interleave):\n print(f\"Worker {w} process sample {i} {sample}\")\n\nprint(\"Concatenate datasets\")\nfor w in range(n_workers):\n ds_concatenate = hf_datasets.concatenate_datasets([ds_1, ds_2, ds_3]).shard(n_workers, w)\n for i, sample in enumerate(ds_concatenate):\n print(f\"Worker {w} process sample {i} {sample}\")\n\nprint(\"Concated and shuffled datasets\")\nfor w in range(n_workers):\n ds_concatenate = hf_datasets.concatenate_datasets([ds_1, ds_2, ds_3]).shuffle().shard(n_workers, w)\n for i, sample in enumerate(ds_concatenate):\n print(f\"Worker {w} process sample {i} {sample}\")\n```\n\n> Interleave datasets\nWorker 0 process sample 0 {'dataset': 0, 'sample': 0}\nWorker 1 process sample 0 {'dataset': 1, 'sample': 0}\nWorker 2 process sample 0 {'dataset': 2, 'sample': 0}\n\n> Concatenate datasets\nWorker 0 process sample 0 {'dataset': 0, 'sample': 0}\nWorker 0 process sample 1 {'dataset': 0, 'sample': 1}\nWorker 1 process sample 0 {'dataset': 1, 'sample': 0}\nWorker 1 process sample 1 {'dataset': 2, 'sample': 0}\nWorker 2 process sample 0 {'dataset': 2, 'sample': 1}\nWorker 2 process sample 1 {'dataset': 2, 'sample': 2}\n\n> Concated and shuffled datasets\nWorker 0 process sample 0 {'dataset': 2, 'sample': 2}\nWorker 0 process sample 1 {'dataset': 2, 'sample': 0}\nWorker 1 process sample 0 {'dataset': 0, 'sample': 1}\nWorker 1 process sample 1 {'dataset': 2, 'sample': 1}\nWorker 2 process sample 0 {'dataset': 2, 'sample': 2}\nWorker 2 process sample 1 {'dataset': 0, 'sample': 0}\n\nWithout shuffling, round robin would yield:\n> Worker 0 process sample 0 {'dataset': 0, 'sample': 0}\nWorker 0 process sample 1 {'dataset': 2, 'sample': 0}\nWorker 1 process sample 0 {'dataset': 0, 'sample': 1}\nWorker 1 process sample 1 {'dataset': 2, 'sample': 1}\nWorker 2 process sample 0 {'dataset': 1, 'sample': 0}\nWorker 2 process sample 1 {'dataset': 2, 'sample': 2}", "# With `datasets.IterableDataset`\n\nThe above works for `Dataset`, but with a sharded `IterableDataset` some data get discarded. See the following results obtained with the script below.\n\n> Simulate run with 3 workers\n\n> Interleave datasets\nWorker 0 process sample 0 {'dataset': 0, 'sample': 0}\nWorker 1 fails with list index out of range.\nWorker 2 fails with list index out of range.\nWith dataloader\nToo many dataloader workers: 3 (max is dataset.num_shards=1). Stopping 2 dataloader workers.\n{'dataset': tensor([0]), 'sample': tensor([0])}\n\n> Concatenate datasets\nWorker 0 process sample 0 {'dataset': 0, 'sample': 0}\nWorker 0 process sample 1 {'dataset': 1, 'sample': 0}\nWorker 0 process sample 2 {'dataset': 2, 'sample': 0}\nWorker 1 fails with list index out of range\nWorker 2 fails with list index out of range\nWith dataloader\nToo many dataloader workers: 3 (max is dataset.num_shards=1). Stopping 2 dataloader workers.\n{'dataset': tensor([0]), 'sample': tensor([0])}\n{'dataset': tensor([1]), 'sample': tensor([0])}\n{'dataset': tensor([2]), 'sample': tensor([0])}\n\n> Concated and shuffled datasets\nWorker 0 process sample 0 {'dataset': 0, 'sample': 0}\nWorker 0 process sample 1 {'dataset': 1, 'sample': 0}\nWorker 0 process sample 2 {'dataset': 2, 'sample': 0}\nWorker 1 fails with list index out of range\nWorker 2 fails with list index out of range\nWith dataloader\nToo many dataloader workers: 3 (max is dataset.num_shards=1). Stopping 2 dataloader workers.\n{'dataset': tensor([0]), 'sample': tensor([0])}\n{'dataset': tensor([1]), 'sample': tensor([0])}\n{'dataset': tensor([2]), 'sample': tensor([0])}\n\n<details>\n\n<summary>Experiment script</summary>\n\n```python\nds_1 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 0, \"n_samples\": 2}).to_iterable_dataset(\n num_shards=2\n)\nds_2 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 1, \"n_samples\": 1}).to_iterable_dataset(\n num_shards=1\n)\nds_3 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 2, \"n_samples\": 3}).to_iterable_dataset(\n num_shards=3\n)\n\nn_workers = 3\nprint(f\"Simulate run with {n_workers} workers\")\n\nprint(\"\\nInterleave datasets\")\nds_interleave = hf_datasets.interleave_datasets([ds_1, ds_2, ds_3])\nfor w in range(n_workers):\n try:\n for i, sample in enumerate(ds_interleave.shard(n_workers, w)):\n print(f\"Worker {w} process sample {i} {sample}\")\n except IndexError as e:\n print(f\"Worker {w} fails with {e}.\")\n\nprint(\"With dataloader\")\nfor sample in torch.utils.data.DataLoader(ds_interleave, num_workers=n_workers):\n print(f\"{sample}\")\n\nprint(\"\\nConcatenate datasets\")\nds_concatenate = hf_datasets.concatenate_datasets([ds_1, ds_2, ds_3])\nfor w in range(n_workers):\n try:\n for i, sample in enumerate(ds_concatenate.shard(n_workers, w)):\n print(f\"Worker {w} process sample {i} {sample}\")\n except IndexError as e:\n print(f\"Worker {w} fails with {e}\")\n\nprint(\"With dataloader\")\nfor sample in torch.utils.data.DataLoader(ds_concatenate, num_workers=n_workers):\n print(f\"{sample}\")\n\nprint(\"\\nConcated and shuffled datasets\")\nds_concatenate = hf_datasets.concatenate_datasets([ds_1, ds_2, ds_3]).shuffle()\nfor w in range(n_workers):\n try:\n for i, sample in enumerate(ds_concatenate.shard(n_workers, w)):\n print(f\"Worker {w} process sample {i} {sample}\")\n except IndexError as e:\n print(f\"Worker {w} fails with {e}\")\n\nprint(\"With dataloader\")\nfor sample in torch.utils.data.DataLoader(ds_concatenate, num_workers=n_workers):\n print(f\"{sample}\")\n```\n\n</details>\n\n# Round Robin with fixed logic\n\n> I started implementing the following, but I'm afraid my sharding logic is incorrect.\n\nHere is a solution for mixing the data in a round robin fashion that allows to distribute the data to all workers. In the previous example above only 1 worker over 3 was actually retrieving data, which resulted in discarding some data.\n\n```python\ndef shard_data_sources(self, num_shards: int, index: int, contiguous=True) -> \"MixMultiSourceExampleIterable\":\n \"\"\"Shard the underlying iterables in a roundrobin manner.\n\n Let's consider we have our iterables as [[s0_0, s0_1], [s1_0], [s2_0, s2_1, s2_3]],\n and we request 3 shards.\n index 0 gets s0_0 s2_0\n index 1 gets s0_1 s2_1\n index 2 gets s1_0 s2_3\n \"\"\"\n return MixMultiSourcesExampleIterable(\n list(\n islice(\n # flatten all underlying iterables (fixed logic)\n [\n ex_iterable.shard_data_sources(ex_iterable.num_shards, index)\n for ex_iterable in self.ex_iterables\n for index in range(ex_iterable.num_shards)\n ],\n # offset the starting point by the index\n index,\n # take over the full list, so exhaust the iterators\n None,\n # step by the number of shards requested\n num_shards,\n )\n )\n )\n```\n\nEditing the example above with the following we obtain the expected result:\n```python\nprint(\"\\nMix datasets\")\nds_mix = mix_dataset([ds_1, ds_2, ds_3])\nfor w in range(n_workers):\n try:\n for i, sample in enumerate(ds_mix.shard(n_workers, w)):\n print(f\"Worker {w} process sample {i} {sample}\")\n except IndexError as e:\n print(f\"Worker {w} fails with {e}\")\n\nprint(\"With dataloader\")\nfor sample in torch.utils.data.DataLoader(ds_mix, num_workers=n_workers):\n print(f\"{sample}\")\n```\n> Mix datasets\nMix datasets\nWorker 0 process sample 0 {'dataset': 0, 'sample': 0}\nWorker 0 process sample 1 {'dataset': 2, 'sample': 0}\nWorker 1 process sample 0 {'dataset': 0, 'sample': 1}\nWorker 1 process sample 1 {'dataset': 2, 'sample': 1}\nWorker 2 process sample 0 {'dataset': 1, 'sample': 0}\nWorker 2 process sample 1 {'dataset': 2, 'sample': 2}\nWith dataloader\n{'dataset': tensor([0]), 'sample': tensor([0])}\n{'dataset': tensor([0]), 'sample': tensor([1])}\n{'dataset': tensor([1]), 'sample': tensor([0])}\n{'dataset': tensor([2]), 'sample': tensor([0])}\n{'dataset': tensor([2]), 'sample': tensor([1])}\n{'dataset': tensor([2]), 'sample': tensor([2])}\n\n# Questions \n\n- The example is quite small, showing that some data get discarded, but on large datasets is this significant?\n- How does the suggested solution interplays with shuffling?\n\n\n\n\n", "# Larger Experiment\n\n> The example is quite small, showing that some data get discarded, but on large datasets is this significant?\n\nContinuing the experiment above, but with 3 larger and unbalanced datasets, with respectively 1000, 150, and 300 samples, and a dataloader with 4 workers:\n \n> Interleave datasets\nWith dataloader\nToo many dataloader workers: 4 (max is dataset.num_shards=1). Stopping 3 dataloader workers.\nYield 300 samples\n\n> Concatenate datasets\nWith dataloader\nToo many dataloader workers: 4 (max is dataset.num_shards=1). Stopping 3 dataloader workers.\nYield 705 samples\n\n> Concated and shuffled datasets\nWith dataloader\nToo many dataloader workers: 4 (max is dataset.num_shards=1). Stopping 3 dataloader workers.\nYield 705 samples\n\n> Mix datasets\nWith dataloader\nYield 1405 samples\n\nThe dataset mixing proposed above is the only one that yields all the samples while using all the dataloaders.\nAdditional checks should include training metrics (does it improve training quality to mix the data like this), and behavior check in a DDP settings, we don't want to face any deadlock due to some GPU having more batches than other. But this later point should be already handled by the iterator of the `IterableDataset`.\n\n# Follow up?\n\n@lhoestq would there be any interest in making a PR of it? Otherwise I can close the issue as I found a solution to my problem. ", "I believe this PR could solve your issue? :)\n\nhttps://github.com/huggingface/datasets/pull/7786", "> I believe this PR could solve your issue? :)\n\nThank you @lhoestq for the reply.\nI have just tested it with the script above. It gives:\n\n> Interleave datasets without replacement\nWith dataloader\nToo many dataloader workers: 4 (max is dataset.num_shards=1). Stopping 3 dataloader workers.\nYield 705 samples\n\nIf we compare with the original `interleave_dataset` method it produces 405 samples more. However, it only uses 1 worker on the 4 available. Moreover it doesn't yield all the samples as the mixing strategy with RoundRobin above does (1405 samples vs 705).", "@LTMeyer With the following script and using the code from #7786 I get all 1450 samples\n\n```\nimport datasets as hf_datasets\n\n\ndef gen(dataset: int, n_samples: int):\n for i in range(n_samples):\n yield {\"dataset\": dataset, \"sample\": i}\n\n\nds_1 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 0, \"n_samples\": 1000}).to_iterable_dataset()\nds_2 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 1, \"n_samples\": 150}).to_iterable_dataset()\nds_3 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 2, \"n_samples\": 300}).to_iterable_dataset()\n\nprint(\"Interleave datasets\")\nds_interleave = hf_datasets.interleave_datasets(\n [ds_1, ds_2, ds_3],\n probabilities=[1 / 3, 1 / 3, 1 / 3],\n stopping_strategy=\"all_exhausted_without_replacement\",\n)\nfor i, sample in enumerate(ds_interleave):\n print(f\"process sample {i} {sample}\")\n```\nI'm not sure on the workers side how many will be spawned and so on. ", "> [@LTMeyer](https://github.com/LTMeyer) With the following script and using the code from [#7786](https://github.com/huggingface/datasets/pull/7786) I get all 1450 samples\n\nThis depends on the number of shards and the number of processes being used.\nIn the example below there is only one shard per dataset (the default of `to_iterable_dataset` method). Then, the for loop is running in the main process. It thus consumes all the shards, hence the 1450 samples.\n\n> \n> ```\n> import datasets as hf_datasets\n> \n> \n> def gen(dataset: int, n_samples: int):\n> for i in range(n_samples):\n> yield {\"dataset\": dataset, \"sample\": i}\n> \n> \n> ds_1 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 0, \"n_samples\": 1000}).to_iterable_dataset()\n> ds_2 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 1, \"n_samples\": 150}).to_iterable_dataset()\n> ds_3 = hf_datasets.Dataset.from_generator(gen, gen_kwargs={\"dataset\": 2, \"n_samples\": 300}).to_iterable_dataset()\n> \n> print(\"Interleave datasets\")\n> ds_interleave = hf_datasets.interleave_datasets(\n> [ds_1, ds_2, ds_3],\n> probabilities=[1 / 3, 1 / 3, 1 / 3],\n> stopping_strategy=\"all_exhausted_without_replacement\",\n> )\n> for i, sample in enumerate(ds_interleave):\n> print(f\"process sample {i} {sample}\")\n> ```\n> \n\n\n> I'm not sure on the workers side how many will be spawned and so on.\n\nWhile using the data to train a model, I would like to use the `torch.utils.data.DataLoader` to feed batches of data to my model. To make the data loading fast, it is common to use `num_workers>0` in the dataloader. This will consume data in parallel. In practice, it copies the dataset instance and read in parallel different chunks of data. These chunks correspond to the underlying shards of the iterable dataset.\n\nIf we have 1 shard per dataset, as it is the case in the example above, the dataloading will indeed get all the 1450 samples, but it will run only in one process even if multiple are available. This is inefficient because it doesn't utilize all available resources. See the script and results below.\n\n```python\nfor num_workers in [0, 1, 2, 3, 4]:\n print(f\"Dataloader with {num_workers} workers.\")\n dataloader = DataLoader(ds_interleave, num_workers=num_workers, batch_size=1)\n for i, sample in enumerate(dataloader, start=1):\n pass\n print(f\"{i} processed samples\")\n```\n\n```\nDataloader with 0 workers.\n1450 processed samples\nDataloader with 1 workers.\n1450 processed samples\nDataloader with 2 workers.\nToo many dataloader workers: 2 (max is dataset.num_shards=1). Stopping 1 dataloader workers.\n1450 processed samples\nDataloader with 3 workers.\nToo many dataloader workers: 3 (max is dataset.num_shards=1). Stopping 2 dataloader workers.\n1450 processed samples\nDataloader with 4 workers.\nToo many dataloader workers: 4 (max is dataset.num_shards=1). Stopping 3 dataloader workers.\n1450 processed samples\n```\n\nNow if we shard our data differently, like 2, 1, and 3 for each dataset respectively as the [previous example](https://github.com/huggingface/datasets/issues/7792#issuecomment-3345970293), and use a dataloader with different number of workers (same script as above), we obtain:\n\n```\nDataloader with 0 workers.\n1450 processed samples\nDataloader with 1 workers.\n1450 processed samples\nDataloader with 2 workers.\nToo many dataloader workers: 2 (max is dataset.num_shards=1). Stopping 1 dataloader workers.\n850 processed samples\nDataloader with 3 workers.\nToo many dataloader workers: 3 (max is dataset.num_shards=1). Stopping 2 dataloader workers.\n750 processed samples\nDataloader with 4 workers.\nToo many dataloader workers: 4 (max is dataset.num_shards=1). Stopping 3 dataloader workers.\n750 processed samples\n```", "I added a small fix to your PR @radulescupetru to try to make @LTMeyer 's example work :)\n\nCan you confirm it works for you now @LTMeyer ?\n\nNote that maximum parallelism requires each subset to have num_shards >= num_workers, otherwise there aren't enough shards to distribute to every worker for interleaving. In your example one of the subsets has only 1 shard, so only 1 worker can take care of interleaving.", "> Can you confirm it works for you now [@LTMeyer](https://github.com/LTMeyer) ?\n\nResult with https://github.com/huggingface/datasets/pull/7786/commits/a547d81469128bea4acc3bcc2a4a6a95968936ee:\n```\nDataloader with 0 workers.\n1450 processed samples\nDataloader with 1 workers.\n1450 processed samples\nDataloader with 2 workers.\nToo many dataloader workers: 2 (max is dataset.num_shards=1). Stopping 1 dataloader workers.\n1450 processed samples\nDataloader with 3 workers.\nToo many dataloader workers: 3 (max is dataset.num_shards=1). Stopping 2 dataloader workers.\n1450 processed samples\nDataloader with 4 workers.\nToo many dataloader workers: 4 (max is dataset.num_shards=1). Stopping 3 dataloader workers.\n1450 processed samples\n```\n\n I have checked with the script above and I confirm that all samples are now correctly returned, thank you @lhoestq .\n\n> Note that maximum parallelism requires each subset to have num_shards >= num_workers, otherwise there aren't enough shards to distribute to every worker for interleaving. In your example one of the subsets has only 1 shard, so only 1 worker can take care of interleaving.\n\nThis point I'm not sure I understand. That is maybe where @radulescupetru's intent and mine differ. Why should we limit the number of workers to the minimum number of shards? My initial goal was to distribute shards among workers to maximize data loading speed, and to mix the data so batches are representative of the whole dataset and diverse enough (hence the round-robin). \n\nIn the example above, we have 6 shards in total, can we not distribute these shards among workers? That what the `MixMultiSourcesExampleIterable` in https://github.com/huggingface/datasets/issues/7792#issuecomment-3345970293 above does.\n- If 2 workers, 3 shards for each. \n- If 3 workers, 2 shards for each.\n- If 4 workers, the 2 first ones get 2 shards while the two last ones get only 1.\n- Above 6 workers, the 6 first ones get 1 shard each, and the remaining workers get none.\n\n\n", "@LTMeyer I think it's just a design choice that datasets library took. From my interaction with it, it seems that even when concatenating or interleaving, individual components are still treated individually (for example, num_shards is not summed).\n\nI guess in a real scenario you wouldn't end up with 1 shard only, but it's true that you need to be a bit careful with the setup. For workers it's a bit more automated in the sense that if you have more it will stop the extra ones, but when distributing a dataset over multiple gpus it's even more tricky as if the number of shards is not a factor of world size iterating is slower.", "> [@LTMeyer](https://github.com/LTMeyer) I think it's just a design choice that datasets library took. From my interaction with it, it seems that even when concatenating or interleaving, individual components are still treated individually (for example, num_shards is not summed).\n\nIndeed. I am curious to know if there is any explanation for this choice that I am missing.\n\n> I guess in a real scenario you wouldn't end up with 1 shard only, but it's true that you need to be a bit careful with the setup. \n\nIn my case I would like to mix many small datasets which are individually based on only few shards. So it's actually close to the case with 1 shard only.\n\n> For workers it's a bit more automated in the sense that if you have more it will stop the extra ones, but when distributing a dataset over multiple gpus it's even more tricky as if the number of shards is not a factor of world size iterating is slower.\n\nMy understanding is that, in a multi-gpu settings, we want each GPU to receive the same number of batches to avoid deadlock in any synchronization process. \nMulti-GPU related sharding of the `IterableDataset` is managed there https://github.com/huggingface/datasets/blob/4.1.1/src/datasets/iterable_dataset.py#L2371-L2392,\nwhile the sharding for dataloaders with multiple workers is handled there https://github.com/huggingface/datasets/blob/4.1.1/src/datasets/iterable_dataset.py#L2292-L2314.\n\nHere is a script to check the behavior in case of multi-gpus, using `split_dataset_by_node`. In the example I consider just 2 GPUs.\n\n```python\nworld_size = 2\nfor num_workers in [0, 1, 2, 3, 4]:\n for rank in range(world_size):\n print(f\"Rank {rank}\")\n ds_interleave_rank = split_dataset_by_node(ds_interleave, rank, world_size)\n print(f\"Dataloader with {num_workers} workers.\")\n dataloader = DataLoader(ds_interleave_rank, num_workers=num_workers, batch_size=1)\n for i in enumerate(dataloader, start=1):\n pass\n print(f\"{i} processed samples\")\n print(\"\\n\")\n```\n\nThe results using https://github.com/huggingface/datasets/pull/7786/commits/455bfaaa6d574aa9d9c9592baee390017512cc5f:\n```\nRank 0\nDataloader with 0 workers.\n725 processed samples\nRank 1\nDataloader with 0 workers.\n725 processed samples\n\n\nRank 0\nDataloader with 1 workers.\n725 processed samples\nRank 1\nDataloader with 1 workers.\n725 processed samples\n\n\nRank 0\nDataloader with 2 workers.\nToo many dataloader workers: 2 (max is dataset.num_shards=1). Stopping 1 dataloader workers.\n725 processed samples\nRank 1\nDataloader with 2 workers.\n725 processed samples\n\n\nRank 0\nDataloader with 3 workers.\nToo many dataloader workers: 3 (max is dataset.num_shards=1). Stopping 2 dataloader workers.\n725 processed samples\nRank 1\nDataloader with 3 workers.\n725 processed samples\n\n\nRank 0\nDataloader with 4 workers.\nToo many dataloader workers: 4 (max is dataset.num_shards=1). Stopping 3 dataloader workers.\n725 processed samples\nRank 1\nDataloader with 4 workers.\n725 processed samples\n```\n\nIf now I use the mixing described above the results are:\n```\nRank 0\nDataloader with 0 workers.\n750 processed samples\nRank 1\nDataloader with 0 workers.\n700 processed samples\n\n\nRank 0\nDataloader with 1 workers.\n750 processed samples\nRank 1\nDataloader with 1 workers.\n700 processed samples\n\n\nRank 0\nDataloader with 2 workers.\n750 processed samples\nRank 1\nDataloader with 2 workers.\n700 processed samples\n\n\nRank 0\nDataloader with 3 workers.\n750 processed samples\nRank 1\nDataloader with 3 workers.\n700 processed samples\n\n\nRank 0\nDataloader with 4 workers.\n750 processed samples\nRank 1\nDataloader with 4 workers.\n700 processed samples\n```\n\nDifferent GPUs received different number of batches which is problematic. The interleave method, on the other hand, feeds each GPU with the same number of batches. Nonetheless, it doesn't leverage all available workers.\nI'll check if I can fix the distribution of shards across GPU in the last configuration.", "When concatenating or interleaving, the resulting `num_shards` is the *minimum `num_shards` of the input datasets*. This allows each new shard to always contain data from every input dataset. This ensures in every shard the right sampling when interleaving and the right data order when concatenating.\n\nSumming the dataset shards isn't ideal since each shard would contain data from only one of the dataset and would not contain any interleaved/concatenated data.", "Thank you @lhoestq, it makes perfect sense. The part I am missing is that if I concatenate many datasets with small number of shards it will result in a global dataset with not so many shards, thus limiting the use of available workers. Data loading will be consequently inefficient. I was looking for a solution to leverage all parallelism available to maximize data loading speed.\n\nMy original use case was:\nI want to use a dataset stored on the HF hub. It is composed of many subfolders. Each of this subfolder contain only a few shards. I would like to use the dataset but only on a subset of folders, while keeping information about the origin of each sample (i.e. from which subfolder they come from).\nThe first part would possible with the `data_files` argument of `load_dataset` method. However, I would not have the origin information about the sample, as it is not provided in the original dataset. I was thus thinking about considering each subfolder as an independent HF iterable dataset and concatenate them. This method does not work because it drastically reduces the dataloading efficiency due to the low number of shards.\n\n> Summing the dataset shards isn't ideal `since` each shard would contain data from only one of the dataset and would not contain any interleaved/concatenated data.\n\nThis is not necessarily a problem for my use case. It will be the case for the original dataset anyway.", "Also, I notice in the example above that if we modify the number of shards, we get different number of samples per GPU and workers even with the implementation of @radulescupetru. This will cause a deadlock in the DDP. So I guess HF expects all shards to contain the same number of samples. Is that a correct assumption @lhoestq?\n\nSetting the number of shards for the datasets above to 2, 2 and 3. Using the `interleave_datasets` I get the following:\n```\nRank 0\nAssigning 1 shard (or data source) of the dataset to each node.\nDataloader with 0 workers.\nAssigning 1 shard (or data source) of the dataset to each node.\n775 processed samples\nRank 1\nDataloader with 0 workers.\n675 processed samples\n\n\nRank 0\nAssigning 1 shard (or data source) of the dataset to each node.\nDataloader with 1 workers.\nAssigning 1 shard (or data source) of the dataset to each node.\n775 processed samples\nRank 1\nDataloader with 1 workers.\n675 processed samples\n\n\nRank 0\nAssigning 1 shard (or data source) of the dataset to each node.\nDataloader with 2 workers.\nToo many dataloader workers: 2 (max is dataset.num_shards=1). Stopping 1 dataloader workers.\nWARNING:datasets.iterable_dataset:Too many dataloader workers: 2 (max is dataset.num_shards=1). Stopping 1 dataloader workers.\nAssigning 1 shard (or data source) of the dataset to each node.\n775 processed samples\nRank 1\nDataloader with 2 workers.\n675 processed samples\n\n\nRank 0\nAssigning 1 shard (or data source) of the dataset to each node.\nDataloader with 3 workers.\nToo many dataloader workers: 3 (max is dataset.num_shards=1). Stopping 2 dataloader workers.\nWARNING:datasets.iterable_dataset:Too many dataloader workers: 3 (max is dataset.num_shards=1). Stopping 2 dataloader workers.\nAssigning 1 shard (or data source) of the dataset to each node.\n775 processed samples\nRank 1\nDataloader with 3 workers.\n675 processed samples\n\n\nRank 0\nAssigning 1 shard (or data source) of the dataset to each node.\nDataloader with 4 workers.\nToo many dataloader workers: 4 (max is dataset.num_shards=1). Stopping 3 dataloader workers.\nWARNING:datasets.iterable_dataset:Too many dataloader workers: 4 (max is dataset.num_shards=1). Stopping 3 dataloader workers.\nAssigning 1 shard (or data source) of the dataset to each node.\n775 processed samples\nRank 1\nDataloader with 4 workers.\n675 processed samples\n```", "I see @LTMeyer, that makes sense. Do you think we should sum the shards by default for concatenating then ? I feel like your use case is more important than ensuring each worker has data of every subdataset in order.\n\n(I wouldn't touch the interleaving logic though)\n\n> Also, I notice in the example above that if we modify the number of shards, we get different number of samples per GPU and workers even with the implementation of @radulescupetru. This will cause a deadlock in the DDP. So I guess HF expects all shards to contain the same number of samples. Is that a correct assumption @lhoestq?\n\nShards rarely have the same number of samples, so the DDP algorithm itself should be able to stop on its own or have a strategy to circumvent this. For example it can loop until all the nodes have exhausted their data:\n\n```python\ndef loop():\n while True:\n yield from dataloader\n yield \"end\"\n\nfor x in loop():\n if x == \"end\":\n exhausted[rank] = True\n continue\n # stop once the data from all the ranks are exhausted\n dist.all_reduce(exhausted)\n if torch.all(exhausted):\n break\n # do your forward pass + loss here\n # model.forward(...)\n```\n\nI made a full example here: https://github.com/huggingface/datasets/issues/6623#issuecomment-2379458138", "To summarize, and highlight the distinction with https://github.com/huggingface/datasets/pull/7786, there are actually two feature requests:\n1. Similarly to `interleave_datasets`, we want to interleave the longest dataset without repetition. This is handled by https://github.com/huggingface/datasets/pull/7786, and is consistant with the rest of the HF features (i.e. `concatenate_datasets` and `interleave_datasets`);\n2. We want to be able to _fuse_ datasets and distribute their shards across workers to maximize data loading speed.\n\n > I feel like your use case is more important than ensuring each worker has data of every subdataset in order.\n\nIndeed my use case, pointed as 2. above is first about maximizing data loading speed and second about mixing the data. The order of priority seems to be the opposite in 1.\n\n> Do you think we should sum the shards by default for concatenating then?\n\nI think the library should at least provide a method for this. Users can then decide what matters the most for their use case (data order or dataloading speed). What do you think?\n\n> Shards rarely have the same number of samples, so the DDP algorithm itself should be able to stop on its own or have a strategy to circumvent this.\n\nIf imbalanced data stream in a DDP context is not the responsibility of the datasets library, it is, for me, a reason more to provides a fuse or mix dataset method that sum the shards.\n\n> I made a full example here: https://github.com/huggingface/datasets/issues/6623#issuecomment-2379458138 \n\nThank you for the example. Pytorch now provides also utilities to handle this problematic case, see [Join context manager in DDP](https://docs.pytorch.org/tutorials/advanced/generic_join.html#:%7E:text=The%20context%20manager%20allows%20the,shadowed%20are%20specified%20by%20hooks)", "I'm closing this issue because of several existing solutions:\n- https://github.com/huggingface/datasets/pull/7786 allows to interleave datasets without replacement.\n- Using [`.shard`](https://huggingface.co/docs/datasets/v4.2.0/en/package_reference/main_classes#datasets.IterableDataset.shard) instead of [`split_dataset_by_node`](https://huggingface.co/docs/datasets/v4.2.0/en/package_reference/main_classes#datasets.distributed.split_dataset_by_node). Given _m_ shards and _n_ ranks, if m % n != 0, the later function will make each of the _n_ ranks go through all of the _m_ shards, although not fetching the same data. On the other hand, the former function can distribute the _m_ shards across the _n_ ranks and make better use of parallel reads.\n\nThank you @lhoestq and @radulescupetru for the help." ]
2025-09-26T10:05:19Z
2025-10-15T18:05:23Z
2025-10-15T18:05:23Z
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
### Feature request I would like to be able to concatenate multiple `IterableDataset` with possibly different features. I would like to then be able to stream the results in parallel (both using DDP and multiple workers in the pytorch DataLoader). I want the merge of datasets to be well balanced between the different processes. ### Motivation I want to train a model on a combination of datasets, which I can convert to a single representation. This applies to converting different datasets items to the same Python class, as using a tokenizer on multiple modalities. Assuming that my original datasets are not necessarily well balanced as they may have different size and thus different number of shards, I would like the merged dataset to be distributed evenly over the multiple processes. I don't mind if it's not perfectly balanced, and as result, some workers of the torch DataLoader do nothing, as long as the DDP is properly handled causing no deadlock. ### What I've tried I've tried the two functions already provided in datasets, namely `interleave_datasets` and `concatenate_datasets`. - Interleave seems to be the best approach of what I'm trying to do. However, it doesn't suit my purpose because as I understand it, it stops as soon as one of the dataset source is exhausted, or repeat the smallest source items until the largest is exhausted. I would like something in-between, similarly to what [roundrobin does](https://more-itertools.readthedocs.io/en/stable/api.html#more_itertools.roundrobin). - Concatenate does not mix the data enough and one dataset may be overrepresented in some early batches. Let's consider we have 3 datasets composed of different number of shards as follow [[s0_0, s0_1], [s1_0], [s2_0, s2_1, s2_3]], where s denotes the underlying shard, the first index the dataset and the second the shard number. If we request 3 shards in the `shard_data_source` we should obtain the following: index 0 gets s0_0 s2_0 index 1 gets s0_1 s2_1 index 2 gets s1_0 s2_3 I started implementing the following, but I'm afraid my sharding logic is incorrect. ```python from copy import deepcopy from itertools import chain, islice import datasets import numpy as np from datasets import IterableDataset from datasets.iterable_dataset import _BaseExamplesIterable from more_itertools import roundrobin class MixMultiSourcesExampleIterable(_BaseExamplesIterable): def __init__(self, ex_iterables: list[_BaseExamplesIterable]): super().__init__() self.ex_iterables = ex_iterables def _init_state_dict(self) -> dict: self._state_dict = { "ex_iterables": [ex_iterable._init_state_dict() for ex_iterable in self.ex_iterables], "type": self.__class__.__name__, } return self._state_dict @property def num_shards(self) -> int: return sum(ex_iterable.num_shards for ex_iterable in self.ex_iterables) def __iter__(self): yield from roundrobin(*self.ex_iterables) def shuffle_data_sources(self, generator: np.random.Generator) -> "MixMultiSourcesExampleIterable": """Shuffle the list of examples iterable, as well as each underlying examples iterable.""" rng = deepcopy(generator) ex_iterables = list(self.ex_iterables) rng.shuffle(ex_iterables) ex_iterables = [ex_iterable.shuffle_data_sources(generator) for ex_iterable in ex_iterables] return MixMultiSourcesExampleIterable(ex_iterables) def shard_data_sources(self, num_shards: int, index: int, contiguous=True) -> "MixMultiSourceExampleIterable": """Shard the underlying iterables in a roundrobin manner. Let's consider we have our iterables as [[s0_0, s0_1], [s1_0], [s2_0, s2_1, s2_3]], and we request 3 shards. index 0 gets s0_0 s2_0 index 1 gets s0_1 s2_1 index 2 gets s1_0 s2_3 """ return MixMultiSourcesExampleIterable( list( islice( # flatten all underlying iterables chain.from_iterable([ex_iterable.shard_data_sources(1, 0) for ex_iterable in self.ex_iterables]), # offset the starting point by the index index, # take over the full list, so exhaust the iterators None, # step by the number of shards requested num_shards, ) ) ) def mix_dataset(iterable_datasets: list[datasets.IterableDataset]) -> IterableDataset: ex_iterable = MixMultiSourcesExampleIterable([ds._ex_iterable for ds in iterable_datasets]) return IterableDataset( ex_iterable, distributed=iterable_datasets[0]._distributed, formatting=iterable_datasets[0]._formatting ) ``` ### Questions - Am I missing something? Is there a way to use `interleave_datasets` or `concatenate_datasets` to fit my purpose? - Would it be the right approach to spread the maximum number of underlying shards across my different processes? ### Your contribution As much as I can.
{ "avatar_url": "https://avatars.githubusercontent.com/u/13559010?v=4", "events_url": "https://api.github.com/users/LTMeyer/events{/privacy}", "followers_url": "https://api.github.com/users/LTMeyer/followers", "following_url": "https://api.github.com/users/LTMeyer/following{/other_user}", "gists_url": "https://api.github.com/users/LTMeyer/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/LTMeyer", "id": 13559010, "login": "LTMeyer", "node_id": "MDQ6VXNlcjEzNTU5MDEw", "organizations_url": "https://api.github.com/users/LTMeyer/orgs", "received_events_url": "https://api.github.com/users/LTMeyer/received_events", "repos_url": "https://api.github.com/users/LTMeyer/repos", "site_admin": false, "starred_url": "https://api.github.com/users/LTMeyer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LTMeyer/subscriptions", "type": "User", "url": "https://api.github.com/users/LTMeyer", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7792/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7792/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7791
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7791/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7791/comments
https://api.github.com/repos/huggingface/datasets/issues/7791/events
https://github.com/huggingface/datasets/pull/7791
3,454,046,306
PR_kwDODunzps6qh_2W
7,791
fix: add `num_proc` argument to `Dataset.to_sql`
{ "avatar_url": "https://avatars.githubusercontent.com/u/100021446?v=4", "events_url": "https://api.github.com/users/EricSaikali/events{/privacy}", "followers_url": "https://api.github.com/users/EricSaikali/followers", "following_url": "https://api.github.com/users/EricSaikali/following{/other_user}", "gists_url": "https://api.github.com/users/EricSaikali/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/EricSaikali", "id": 100021446, "login": "EricSaikali", "node_id": "U_kgDOBfY0xg", "organizations_url": "https://api.github.com/users/EricSaikali/orgs", "received_events_url": "https://api.github.com/users/EricSaikali/received_events", "repos_url": "https://api.github.com/users/EricSaikali/repos", "site_admin": false, "starred_url": "https://api.github.com/users/EricSaikali/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/EricSaikali/subscriptions", "type": "User", "url": "https://api.github.com/users/EricSaikali", "user_view_type": "public" }
[]
open
false
null
[]
null
[ "Hi ! could you also write a test to make sure this works fine ?\r\n\r\n(in case there needs to be a special logic to handle the concurrent writes to the database)", "Hi @lhoestq \r\nDone! Let me know if more is needed :)", "Hi @lhoestq could you please review my changes?", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7791). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-09-25T15:02:46Z
2025-11-10T16:25:57Z
null
NONE
null
null
null
null
**Task Done:** - Resolve issue #7788 : Add the missing argument mapping in Dataset.to_sql (`src/datasets/arrow_dataset.py`)
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7791/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7791/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7791.diff", "html_url": "https://github.com/huggingface/datasets/pull/7791", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7791.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7791" }
true
https://api.github.com/repos/huggingface/datasets/issues/7790
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7790/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7790/comments
https://api.github.com/repos/huggingface/datasets/issues/7790/events
https://github.com/huggingface/datasets/pull/7790
3,453,679,876
PR_kwDODunzps6qgvjv
7,790
update tips in docs
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7790). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "cc @mishig25" ]
2025-09-25T13:36:02Z
2025-09-25T13:39:28Z
2025-09-25T13:39:22Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7790/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7790/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7790.diff", "html_url": "https://github.com/huggingface/datasets/pull/7790", "merged_at": "2025-09-25T13:39:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/7790.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7790" }
true
https://api.github.com/repos/huggingface/datasets/issues/7789
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7789/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7789/comments
https://api.github.com/repos/huggingface/datasets/issues/7789/events
https://github.com/huggingface/datasets/pull/7789
3,453,273,059
PR_kwDODunzps6qfZUc
7,789
fix link for rotten_tomatoes dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/8176079?v=4", "events_url": "https://api.github.com/users/0xmohit/events{/privacy}", "followers_url": "https://api.github.com/users/0xmohit/followers", "following_url": "https://api.github.com/users/0xmohit/following{/other_user}", "gists_url": "https://api.github.com/users/0xmohit/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/0xmohit", "id": 8176079, "login": "0xmohit", "node_id": "MDQ6VXNlcjgxNzYwNzk=", "organizations_url": "https://api.github.com/users/0xmohit/orgs", "received_events_url": "https://api.github.com/users/0xmohit/received_events", "repos_url": "https://api.github.com/users/0xmohit/repos", "site_admin": false, "starred_url": "https://api.github.com/users/0xmohit/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/0xmohit/subscriptions", "type": "User", "url": "https://api.github.com/users/0xmohit", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-09-25T11:51:36Z
2025-09-25T11:51:36Z
null
NONE
null
null
null
null
The current link leads to a 404 page.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7789/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7789/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7789.diff", "html_url": "https://github.com/huggingface/datasets/pull/7789", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7789.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7789" }
true
https://api.github.com/repos/huggingface/datasets/issues/7788
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7788/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7788/comments
https://api.github.com/repos/huggingface/datasets/issues/7788/events
https://github.com/huggingface/datasets/issues/7788
3,450,913,796
I_kwDODunzps7NsMQE
7,788
`Dataset.to_sql` doesn't utilize `num_proc`
{ "avatar_url": "https://avatars.githubusercontent.com/u/30357072?v=4", "events_url": "https://api.github.com/users/tcsmaster/events{/privacy}", "followers_url": "https://api.github.com/users/tcsmaster/followers", "following_url": "https://api.github.com/users/tcsmaster/following{/other_user}", "gists_url": "https://api.github.com/users/tcsmaster/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/tcsmaster", "id": 30357072, "login": "tcsmaster", "node_id": "MDQ6VXNlcjMwMzU3MDcy", "organizations_url": "https://api.github.com/users/tcsmaster/orgs", "received_events_url": "https://api.github.com/users/tcsmaster/received_events", "repos_url": "https://api.github.com/users/tcsmaster/repos", "site_admin": false, "starred_url": "https://api.github.com/users/tcsmaster/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tcsmaster/subscriptions", "type": "User", "url": "https://api.github.com/users/tcsmaster", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
2025-09-24T20:34:47Z
2025-09-24T20:35:01Z
null
NONE
null
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
The underlying `SqlDatasetWriter` has `num_proc` as an available argument [here](https://github.com/huggingface/datasets/blob/5dc1a179783dff868b0547c8486268cfaea1ea1f/src/datasets/io/sql.py#L63) , but `Dataset.to_sql()` does not accept it, therefore it is always using one process for the SQL conversion.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7788/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7788/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7787
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7787/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7787/comments
https://api.github.com/repos/huggingface/datasets/issues/7787/events
https://github.com/huggingface/datasets/pull/7787
3,450,858,674
PR_kwDODunzps6qXRo-
7,787
feat: avoid some copies in torch formatter
{ "avatar_url": "https://avatars.githubusercontent.com/u/9896130?v=4", "events_url": "https://api.github.com/users/drbh/events{/privacy}", "followers_url": "https://api.github.com/users/drbh/followers", "following_url": "https://api.github.com/users/drbh/following{/other_user}", "gists_url": "https://api.github.com/users/drbh/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/drbh", "id": 9896130, "login": "drbh", "node_id": "MDQ6VXNlcjk4OTYxMzA=", "organizations_url": "https://api.github.com/users/drbh/orgs", "received_events_url": "https://api.github.com/users/drbh/received_events", "repos_url": "https://api.github.com/users/drbh/repos", "site_admin": false, "starred_url": "https://api.github.com/users/drbh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/drbh/subscriptions", "type": "User", "url": "https://api.github.com/users/drbh", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7787). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "can you re-read your PR please ?" ]
2025-09-24T20:19:44Z
2025-09-26T15:04:25Z
2025-09-26T15:04:23Z
CONTRIBUTOR
null
null
null
null
## perf: reduce copies in TorchFormatter This PR make changes the torch formatter to avoid unnecessary copies and casts when converting decoded batches to tensors. Because many arrays are already in a torch-friendly memory layout and dtype, we can do zero‑copy conversions (`torch.from_numpy`) and only fall back to `as_tensor` when a dtype/device change is required. We also consolidate lists of same‑shape tensors with a cheap `stack` only when safe. Why it helps - Avoids extra materialization and dtype churn during batched map and indexing. - Preserves API and outputs; only changes internal conversion logic. Small benchmark script (based on https://github.com/huggingface/datasets/issues/6104) ```python import time from datasets import load_dataset def main(): dataset = load_dataset("NightMachinery/hf_datasets_bug1") dataset = dataset["train"] if "train" in dataset else dataset t0 = time.time() dataset.set_format(type="torch") # identity map with small batches dataset = dataset.map(lambda x: x, batched=True, batch_size=20) # force materialization data = dataset[:300] print(len(data.keys())) t1 = time.time() print(f"Duration: {t1 - t0:.2f} s") if __name__ == "__main__": main() ``` Without changes ```bash uv run bench.py ``` ```bash # 303 # Duration: 7.26 s ``` With changes ```bash uv run bench.py ``` ```bash # 303 # Duration: 4.43 s ``` # Updated reproduction scripts Below are some simple test cases using `main` and this `refactor-torch-formatter` branch. I've included the two scripts and output when running on a local machine. ```python # /// script # requires-python = ">=3.10" # dependencies = [ # "torch", # "datasets", # "pillow", # ] # # [tool.uv.sources] # datasets = { git = "https://github.com/huggingface/datasets.git" } # /// import time import random import numpy as np from PIL import Image from datasets import Dataset, load_dataset import torch def create_mock_images_dataset(num_samples=5000): """Create a deterministic mock dataset with PIL images.""" random.seed(42) np.random.seed(42) images = [] labels = [] for i in range(num_samples): # Create deterministic RGB image width, height = 64, 64 rgb_array = np.random.randint(0, 256, (height, width, 3), dtype=np.uint8) image = Image.fromarray(rgb_array) images.append(image) labels.append(i % 10) # 10 classes return Dataset.from_dict({"image": images, "label": labels}) def create_mock_text_dataset(num_samples=5000): """Create a deterministic mock dataset with text.""" random.seed(42) words = ["apple", "banana", "cherry", "date", "elderberry", "fig", "grape", "honeydew"] texts = [] labels = [] for i in range(num_samples): # Create deterministic text text_length = 5 + (i % 20) # 5-24 words text = " ".join(random.choices(words, k=text_length)) texts.append(text) labels.append(i % 3) # 3 classes return Dataset.from_dict({"text": texts, "label": labels}) def create_mock_ints_dataset(num_samples=5000): """Create a deterministic mock dataset with integers.""" random.seed(42) data = [] labels = [] for i in range(num_samples): # Create deterministic integer arrays arr = [random.randint(0, 1000) for _ in range(50)] # 50 integers each data.append(arr) labels.append(i % 5) # 5 classes return Dataset.from_dict({"data": data, "label": labels}) def create_mock_floats_dataset(num_samples=5000): """Create a deterministic mock dataset with floats.""" random.seed(42) data = [] labels = [] for i in range(num_samples): # Create deterministic float arrays arr = [random.uniform(0.0, 100.0) for _ in range(30)] # 30 floats each data.append(arr) labels.append(i % 4) # 4 classes return Dataset.from_dict({"data": data, "label": labels}) def benchmark_dataset(name, dataset, num_samples=1000): """Benchmark dataset access speed.""" print(f"\n=== {name} Dataset Benchmark ===") t0 = time.time() dataset.set_format(type="torch") # identity map with small batches dataset = dataset.map(lambda x: x, batched=True, batch_size=20) # force materialization data = dataset[:num_samples] print(f"Keys: {list(data.keys())}") print(f"Sample count: {len(data[list(data.keys())[0]])}") t1 = time.time() print(f"Duration: {t1 - t0:.2f} s") print(f"Speed: {num_samples / (t1 - t0):.1f} samples/s") def main(): # PIL Images benchmark images_dataset = create_mock_images_dataset() benchmark_dataset("PIL Images", images_dataset) # Text benchmark text_dataset = create_mock_text_dataset() benchmark_dataset("Text", text_dataset) # Integers benchmark ints_dataset = create_mock_ints_dataset() benchmark_dataset("Integers", ints_dataset) # Floats benchmark floats_dataset = create_mock_floats_dataset() benchmark_dataset("Floats", floats_dataset) if __name__ == "__main__": main() ``` output ```bash uv run --refresh example1.py ``` ```text === PIL Images Dataset Benchmark === Map: 0%| | 0/5000 [00:00<?, ? examples/s]/Users/drbh/.cache/uv/environments-v2/example1-2aca1a30e84bdead/lib/python3.10/site-packages/datasets/features/image.py:352: UserWarning: Downcasting array dtype int64 to uint8 to be compatible with 'Pillow' warnings.warn(f"Downcasting array dtype {dtype} to {dest_dtype} to be compatible with 'Pillow'") Map: 100%|█████████████████████████████████████████████| 5000/5000 [00:01<00:00, 3669.15 examples/s] Keys: ['image', 'label'] Sample count: 1000 Duration: 2.14 s Speed: 466.5 samples/s === Text Dataset Benchmark === Map: 100%|███████████████████████████████████████████| 5000/5000 [00:00<00:00, 141327.04 examples/s] Keys: ['text', 'label'] Sample count: 1000 Duration: 0.04 s Speed: 27004.3 samples/s === Integers Dataset Benchmark === Map: 100%|███████████████████████████████████████████| 5000/5000 [00:00<00:00, 112904.90 examples/s] Keys: ['data', 'label'] Sample count: 1000 Duration: 0.05 s Speed: 21680.6 samples/s === Floats Dataset Benchmark === Map: 100%|███████████████████████████████████████████| 5000/5000 [00:00<00:00, 104084.25 examples/s] Keys: ['data', 'label'] Sample count: 1000 Duration: 0.05 s Speed: 20215.1 samples/s ``` and this branch specifically ```python # /// script # requires-python = ">=3.10" # dependencies = [ # "torch", # "datasets", # "pillow", # ] # # [tool.uv.sources] # datasets = { git = "https://github.com/huggingface/datasets.git", rev = "refactor-torch-formatter" } # /// import time import random import numpy as np from PIL import Image from datasets import Dataset, load_dataset import torch def create_mock_images_dataset(num_samples=5000): """Create a deterministic mock dataset with PIL images.""" random.seed(42) np.random.seed(42) images = [] labels = [] for i in range(num_samples): # Create deterministic RGB image width, height = 64, 64 rgb_array = np.random.randint(0, 256, (height, width, 3), dtype=np.uint8) image = Image.fromarray(rgb_array) images.append(image) labels.append(i % 10) # 10 classes return Dataset.from_dict({"image": images, "label": labels}) def create_mock_text_dataset(num_samples=5000): """Create a deterministic mock dataset with text.""" random.seed(42) words = [ "apple", "banana", "cherry", "date", "elderberry", "fig", "grape", "honeydew", ] texts = [] labels = [] for i in range(num_samples): # Create deterministic text text_length = 5 + (i % 20) # 5-24 words text = " ".join(random.choices(words, k=text_length)) texts.append(text) labels.append(i % 3) # 3 classes return Dataset.from_dict({"text": texts, "label": labels}) def create_mock_ints_dataset(num_samples=5000): """Create a deterministic mock dataset with integers.""" random.seed(42) data = [] labels = [] for i in range(num_samples): # Create deterministic integer arrays arr = [random.randint(0, 1000) for _ in range(50)] # 50 integers each data.append(arr) labels.append(i % 5) # 5 classes return Dataset.from_dict({"data": data, "label": labels}) def create_mock_floats_dataset(num_samples=5000): """Create a deterministic mock dataset with floats.""" random.seed(42) data = [] labels = [] for i in range(num_samples): # Create deterministic float arrays arr = [random.uniform(0.0, 100.0) for _ in range(30)] # 30 floats each data.append(arr) labels.append(i % 4) # 4 classes return Dataset.from_dict({"data": data, "label": labels}) def benchmark_dataset(name, dataset, num_samples=1000): """Benchmark dataset access speed.""" print(f"\n=== {name} Dataset Benchmark ===") t0 = time.time() dataset.set_format(type="torch") # identity map with small batches dataset = dataset.map(lambda x: x, batched=True, batch_size=20) # force materialization data = dataset[:num_samples] print(f"Keys: {list(data.keys())}") print(f"Sample count: {len(data[list(data.keys())[0]])}") t1 = time.time() print(f"Duration: {t1 - t0:.2f} s") print(f"Speed: {num_samples / (t1 - t0):.1f} samples/s") def main(): # PIL Images benchmark images_dataset = create_mock_images_dataset() benchmark_dataset("PIL Images", images_dataset) # Text benchmark text_dataset = create_mock_text_dataset() benchmark_dataset("Text", text_dataset) # Integers benchmark ints_dataset = create_mock_ints_dataset() benchmark_dataset("Integers", ints_dataset) # Floats benchmark floats_dataset = create_mock_floats_dataset() benchmark_dataset("Floats", floats_dataset) if __name__ == "__main__": main() ``` ```bash uv run --refresh example2.py ``` ```text Updated https://github.com/huggingface/datasets.git (2cb64d1b6503afb49d822b20979760efe4519d03) Built datasets @ git+https://github.com/huggingface/datasets.git@2cb64d1b6503afb49d822b20979760efe Uninstalled 1 package in 20ms Installed 1 package in 5ms === PIL Images Dataset Benchmark === Map: 0%| | 0/5000 [00:00<?, ? examples/s]/Users/drbh/.cache/uv/environments-v2/example2-d4af608668b706ec/lib/python3.10/site-packages/datasets/features/image.py:352: UserWarning: Downcasting array dtype int64 to uint8 to be compatible with 'Pillow' warnings.warn(f"Downcasting array dtype {dtype} to {dest_dtype} to be compatible with 'Pillow'") Map: 100%|█████████████████████████████████████████████| 5000/5000 [00:01<00:00, 3645.14 examples/s] Keys: ['image', 'label'] Sample count: 1000 Duration: 2.04 s Speed: 491.2 samples/s === Text Dataset Benchmark === Map: 100%|████████████████████████████████████████████████████| 5000/5000 [00:00<00:00, 169877.28 examples/s] Keys: ['text', 'label'] Sample count: 1000 Duration: 0.03 s Speed: 32236.1 samples/s === Integers Dataset Benchmark === Map: 100%|████████████████████████████████████████████████████| 5000/5000 [00:00<00:00, 131940.33 examples/s] Keys: ['data', 'label'] Sample count: 1000 Duration: 0.04 s Speed: 25493.3 samples/s === Floats Dataset Benchmark === Map: 100%|████████████████████████████████████████████████████| 5000/5000 [00:00<00:00, 120621.64 examples/s] Keys: ['data', 'label'] Sample count: 1000 Duration: 0.04 s Speed: 23370.6 samples/s ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7787/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7787/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7787.diff", "html_url": "https://github.com/huggingface/datasets/pull/7787", "merged_at": "2025-09-26T15:04:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/7787.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7787" }
true
https://api.github.com/repos/huggingface/datasets/issues/7786
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7786/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7786/comments
https://api.github.com/repos/huggingface/datasets/issues/7786/events
https://github.com/huggingface/datasets/pull/7786
3,448,506,148
PR_kwDODunzps6qPTgs
7,786
Sample without replacement option when interleaving datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/26553095?v=4", "events_url": "https://api.github.com/users/radulescupetru/events{/privacy}", "followers_url": "https://api.github.com/users/radulescupetru/followers", "following_url": "https://api.github.com/users/radulescupetru/following{/other_user}", "gists_url": "https://api.github.com/users/radulescupetru/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/radulescupetru", "id": 26553095, "login": "radulescupetru", "node_id": "MDQ6VXNlcjI2NTUzMDk1", "organizations_url": "https://api.github.com/users/radulescupetru/orgs", "received_events_url": "https://api.github.com/users/radulescupetru/received_events", "repos_url": "https://api.github.com/users/radulescupetru/repos", "site_admin": false, "starred_url": "https://api.github.com/users/radulescupetru/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/radulescupetru/subscriptions", "type": "User", "url": "https://api.github.com/users/radulescupetru", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "@lhoestq Continuing on the idea from https://github.com/huggingface/datasets/issues/217 \r\nThis doesn't add a new stopping criteria, but a new argument to interleave_datasets method. Let me know what you think and if you see a better way of doing this I'm open to suggestions.", "Great ! this is a cool additions :)\r\n\r\nIMO sample_with_replacement as a new argument doesn't make sense if the strategy is \"first_exhausted\", which is the default, and since disabling replacement affects the stopping strategy, I would be in favor of having it as a new strategy instead", "Makes sense, here's a revised implementation with that argument removed and adding a new stopping strategy.", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7786). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "@lhoestq Let me know if there's anything on my side that I can do!", "Hi @radulescupetru, I'm commenting here after @lhoestq mentioned this PR in #7792. I'm facing a similar problem and I was wondering if there was a common a solution. Let me know if we share the same problem.\r\n\r\nAs described in the issue, my problem is that I want to mix unbalanced datasets, distribute the samples on multiple workers and ranks, without repeating samples and while retrieving most samples as I can (i.e. without discarding samples whenever they could actually be used). I also noticed that the current approaches `interleave_dataset` or `concatenate_dataset` do not leverage all the workers if the number of shards do not align with the number of workers.", "I pushed a small update @radulescupetru related to @LTMeyer 's issue, I hope you don't mind.\r\n\r\nThe logic looks all good to me now :) could you also update `_interleave_map_style_datasets()` in `arrow_dataset.py` before we merge ? This way the `Dataset` objects will also benefit from this new stopping strategy.", "@lhoestq Thanks for that fix. I've pushed updates to support the new stopping strategy for map style datasets as well." ]
2025-09-24T09:18:14Z
2025-10-07T14:50:16Z
2025-10-07T14:50:16Z
CONTRIBUTOR
null
null
null
null
Right now, `interleave_datasets` function with probabilities will sample with replacement. The PR adds the ability to sample without replacement. ``` import datasets # Create datasets of different sizes to test exhaustion data_a = [{"value": i, "source": "A"} for i in range(5)] data_b = [{"value": i, "source": "B"} for i in range(10, 15)] ds_a = datasets.Dataset.from_list(data_a).to_iterable_dataset() ds_b = datasets.Dataset.from_list(data_b).to_iterable_dataset() # Interleave with probabilities ds_interleaved = datasets.interleave_datasets( [ds_a, ds_b], probabilities=[0.6, 0.4], seed=42, stopping_strategy="all_exhausted", sample_with_replacement=True, ) for i, example in enumerate(ds_interleaved): print(f"Sample:{i}: value:{example['value']:02d} source:{example['source']}") ``` In this example, `sample_with_replacement=True` and it prints: ``` Sample:0: value:10 source:B Sample:1: value:00 source:A Sample:2: value:11 source:B Sample:3: value:12 source:B Sample:4: value:01 source:A Sample:5: value:13 source:B Sample:6: value:14 source:B Sample:7: value:10 source:B Sample:8: value:02 source:A Sample:9: value:03 source:A Sample:10: value:04 source:A ``` Note that sample with value:10 source: B is sampled twice (Sample:0 and Sample:7) Re-running with `sample_with_replacement=False` in prints: ``` Sample:0: value:10 source:B Sample:1: value:00 source:A Sample:2: value:11 source:B Sample:3: value:12 source:B Sample:4: value:01 source:A Sample:5: value:13 source:B Sample:6: value:14 source:B Sample:7: value:02 source:A Sample:8: value:03 source:A Sample:9: value:04 source:A ``` Note that we don't see any repeated items.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7786/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7786/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7786.diff", "html_url": "https://github.com/huggingface/datasets/pull/7786", "merged_at": "2025-10-07T14:50:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/7786.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7786" }
true
https://api.github.com/repos/huggingface/datasets/issues/7785
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7785/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7785/comments
https://api.github.com/repos/huggingface/datasets/issues/7785/events
https://github.com/huggingface/datasets/pull/7785
3,439,897,018
PR_kwDODunzps6pyTM_
7,785
Fix Audio docstring by removing unsupported mono argument
{ "avatar_url": "https://avatars.githubusercontent.com/u/84439872?v=4", "events_url": "https://api.github.com/users/tanuj-rai/events{/privacy}", "followers_url": "https://api.github.com/users/tanuj-rai/followers", "following_url": "https://api.github.com/users/tanuj-rai/following{/other_user}", "gists_url": "https://api.github.com/users/tanuj-rai/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/tanuj-rai", "id": 84439872, "login": "tanuj-rai", "node_id": "MDQ6VXNlcjg0NDM5ODcy", "organizations_url": "https://api.github.com/users/tanuj-rai/orgs", "received_events_url": "https://api.github.com/users/tanuj-rai/received_events", "repos_url": "https://api.github.com/users/tanuj-rai/repos", "site_admin": false, "starred_url": "https://api.github.com/users/tanuj-rai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tanuj-rai/subscriptions", "type": "User", "url": "https://api.github.com/users/tanuj-rai", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "I think we can keep the arg and add the missing torch.mean() in the Audio.decode_example method", "> I think we can keep the arg and add the missing torch.mean() in the Audio.decode_example method\r\n\r\nThank you @lhoestq. I will add torch.mean().", "fixed by #7840 " ]
2025-09-22T09:06:52Z
2025-11-03T14:52:28Z
2025-11-03T14:52:27Z
CONTRIBUTOR
null
null
null
null
This PR fixes issue #7745. Who can review: @lhoestq
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7785/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7785/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7785.diff", "html_url": "https://github.com/huggingface/datasets/pull/7785", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7785.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7785" }
true
https://api.github.com/repos/huggingface/datasets/issues/7783
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7783/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7783/comments
https://api.github.com/repos/huggingface/datasets/issues/7783/events
https://github.com/huggingface/datasets/pull/7783
3,430,715,779
PR_kwDODunzps6pT7pg
7,783
Support huggingface_hub v0.x and v1.x
{ "avatar_url": "https://avatars.githubusercontent.com/u/11801849?v=4", "events_url": "https://api.github.com/users/Wauplin/events{/privacy}", "followers_url": "https://api.github.com/users/Wauplin/followers", "following_url": "https://api.github.com/users/Wauplin/following{/other_user}", "gists_url": "https://api.github.com/users/Wauplin/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Wauplin", "id": 11801849, "login": "Wauplin", "node_id": "MDQ6VXNlcjExODAxODQ5", "organizations_url": "https://api.github.com/users/Wauplin/orgs", "received_events_url": "https://api.github.com/users/Wauplin/received_events", "repos_url": "https://api.github.com/users/Wauplin/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Wauplin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wauplin/subscriptions", "type": "User", "url": "https://api.github.com/users/Wauplin", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7783). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "@lhoestq could you have a look at this PR please? It makes `datasets` compatible with the upcoming `huggingface_hub` v1.0 release while staying compatible with 0.x. Look at PR description for more details.\r\n\r\nThe CI is currently failing because of 429 rate limit errors but otherwise everything should be fine (made extensive local and ci tests to ensure that). Let me know if you notice anything weird. PR is ready to be merged \"as-is\" in my opinion." ]
2025-09-18T14:45:20Z
2025-10-01T13:56:05Z
2025-10-01T13:56:03Z
CONTRIBUTOR
null
null
null
null
Related to https://github.com/huggingface/huggingface_hub/issues/3340. This PR adapts `datasets` to be compatible with both huggingface_hub v0.x and v1.x. In practice nothing else should change (I've checked the codebase). The `HfHubHTTPError` is a base error defined in `huggingface_hub` that inherits from `requests.HTTPError` in v0.x and will inherit from `httpx.HTTPError` in v1.x. It has been introduced ~2 years ago so it's fine to use it right now (i.e. no need to wait for v1.x release or bump minimal version). Most of the changes have been around the test suite to make sure that tests are passing with both `requests` and `httpx` backends. Mid-term it would be good to completely remove the `requests` dependency from `datasets` but that's an orthogonal topic.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7783/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7783/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7783.diff", "html_url": "https://github.com/huggingface/datasets/pull/7783", "merged_at": "2025-10-01T13:56:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/7783.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7783" }
true
https://api.github.com/repos/huggingface/datasets/issues/7782
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7782/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7782/comments
https://api.github.com/repos/huggingface/datasets/issues/7782/events
https://github.com/huggingface/datasets/pull/7782
3,430,341,875
PR_kwDODunzps6pSozj
7,782
set dev version
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7782). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-09-18T13:15:56Z
2025-09-18T13:20:03Z
2025-09-18T13:16:04Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7782/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7782/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7782.diff", "html_url": "https://github.com/huggingface/datasets/pull/7782", "merged_at": "2025-09-18T13:16:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/7782.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7782" }
true
https://api.github.com/repos/huggingface/datasets/issues/7781
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7781/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7781/comments
https://api.github.com/repos/huggingface/datasets/issues/7781/events
https://github.com/huggingface/datasets/pull/7781
3,430,332,841
PR_kwDODunzps6pSm0C
7,781
release: 4.1.1
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7781). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2025-09-18T13:13:47Z
2025-09-18T13:16:48Z
2025-09-18T13:14:47Z
MEMBER
null
null
null
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7781/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7781/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7781.diff", "html_url": "https://github.com/huggingface/datasets/pull/7781", "merged_at": "2025-09-18T13:14:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/7781.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7781" }
true