id
int64 953M
3.35B
| number
int64 2.72k
7.75k
| title
stringlengths 1
290
| state
stringclasses 2
values | created_at
timestamp[s]date 2021-07-26 12:21:17
2025-08-23 00:18:43
| updated_at
timestamp[s]date 2021-07-26 13:27:59
2025-08-23 12:34:39
| closed_at
timestamp[s]date 2021-07-26 13:27:59
2025-08-20 16:35:55
⌀ | html_url
stringlengths 49
51
| pull_request
dict | user_login
stringlengths 3
26
| is_pull_request
bool 2
classes | comments
listlengths 0
30
|
|---|---|---|---|---|---|---|---|---|---|---|---|
2,743,437,260
| 7,335
|
Too many open files: '/root/.cache/huggingface/token'
|
open
| 2024-12-16T21:30:24
| 2024-12-16T21:30:24
| null |
https://github.com/huggingface/datasets/issues/7335
| null |
kopyl
| false
|
[] |
2,740,266,503
| 7,334
|
TypeError: Value.__init__() missing 1 required positional argument: 'dtype'
|
open
| 2024-12-15T04:08:46
| 2025-07-10T03:32:36
| null |
https://github.com/huggingface/datasets/issues/7334
| null |
ghost
| false
|
[
"same error \n```\ndata = load_dataset('/opt/deepseek_R1_finetune/hf_datasets/openai/gsm8k', 'main')[split] \n```",
"> same error\n> \n> ```\n> data = load_dataset('/opt/deepseek_R1_finetune/hf_datasets/openai/gsm8k', 'main')[split] \n> ```\n\nhttps://github.com/huggingface/open-r1/issues/204 this help me",
"Solved by delete `dataset_infos.json` file in dataset dir, or you can transfer datasets from Hugginface to Modelscope by [hf-ms-transfer](https://github.com/wa008/hf-ms-transfer), which will solve this problem by default. "
] |
2,738,626,593
| 7,328
|
Fix typo in arrow_dataset
|
closed
| 2024-12-13T15:17:09
| 2024-12-19T17:10:27
| 2024-12-19T17:10:25
|
https://github.com/huggingface/datasets/pull/7328
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7328",
"html_url": "https://github.com/huggingface/datasets/pull/7328",
"diff_url": "https://github.com/huggingface/datasets/pull/7328.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7328.patch",
"merged_at": "2024-12-19T17:10:25"
}
|
AndreaFrancis
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7328). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,738,514,909
| 7,327
|
.map() is not caching and ram goes OOM
|
open
| 2024-12-13T14:22:56
| 2025-02-10T10:42:38
| null |
https://github.com/huggingface/datasets/issues/7327
| null |
simeneide
| false
|
[
"I have the same issue - any update on this?"
] |
2,738,188,902
| 7,326
|
Remove upper bound for fsspec
|
open
| 2024-12-13T11:35:12
| 2025-01-03T15:34:37
| null |
https://github.com/huggingface/datasets/issues/7326
| null |
fellhorn
| false
|
[
"Unfortunately `fsspec` versioning allows breaking changes across version and there is no way we can keep it without constrains at the moment. It already broke `datasets` once in the past. Maybe one day once `fsspec` decides on a stable and future proof API but I don't think this will happen anytime soon\r\n\r\nedit: bumped to 2024.10.0 in https://github.com/huggingface/datasets/pull/7352"
] |
2,736,618,054
| 7,325
|
Introduce pdf support (#7318)
|
closed
| 2024-12-12T18:31:18
| 2025-03-18T14:00:36
| 2025-03-18T14:00:36
|
https://github.com/huggingface/datasets/pull/7325
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7325",
"html_url": "https://github.com/huggingface/datasets/pull/7325",
"diff_url": "https://github.com/huggingface/datasets/pull/7325.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7325.patch",
"merged_at": "2025-03-18T14:00:36"
}
|
yabramuvdi
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7325). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Hi @AndreaFrancis and @lhoestq ! Thanks for looking at the code and for all the changes and suggestions. I have worked on all your suggestions. I need to work a bit more on the tests, but I created a first version and uploaded a simple pdf to use for testing. The pdf has a couple of pages some of which have images, figures, and tables (this is exactly the pdfs for which I think keeping the pdf format is very interesting because they have multiple types of content). Will try to finish the tests as soon as possible.",
"(I updated the CI a bit, I took the liberty to update your branch and re-run it on your PR @yabramuvdi)"
] |
2,736,008,698
| 7,323
|
Unexpected cache behaviour using load_dataset
|
closed
| 2024-12-12T14:03:00
| 2025-01-31T11:34:24
| 2025-01-31T11:34:24
|
https://github.com/huggingface/datasets/issues/7323
| null |
Moritz-Wirth
| false
|
[
"Hi ! Since `datasets` 3.x, the `datasets` specific files are in `cache_dir=` and the HF files are cached using `huggingface_hub` and you can set its cache directory using the `HF_HOME` environment variable.\r\n\r\nThey are independent, for example you can delete the Hub cache (containing downloaded files) but still reload your cached datasets from the `datasets` cache (containing prepared datasets in Arrow format)"
] |
2,732,254,868
| 7,322
|
ArrowInvalid: JSON parse error: Column() changed from object to array in row 0
|
open
| 2024-12-11T08:41:39
| 2025-07-15T13:06:55
| null |
https://github.com/huggingface/datasets/issues/7322
| null |
Polarisamoon
| false
|
[
"Hi ! `datasets` uses Arrow under the hood which expects each column and array to have fixed types that don't change across rows of a dataset, which is why we get this error. This dataset in particular doesn't have a format compatible with Arrow unfortunately. Don't hesitate to open a discussion or PR on HF to fix the dataset",
"@lhoestq Is it correct to assume that most multimodal datasets with variable number of images across conversations are not compatible with Arrow ? \n\nI’m running into a problem while trying to format multimodal datasets (image + text) using the Hugging Face datasets library. Specifically, I’m working with a structure where conversations include both images and text messages. When I convert my dataset from a Python dict to a Hugging Face Dataset, I’m seeing unexpected None values being inserted for some fields that aren’t relevant for a given message (e.g., \"text\": None in image messages). Here’s what seems to be happening:\n\n🔍 What’s going wrong\n\nHugging Face datasets (backed by Apache Arrow) tries to flatten the schema across all samples. That means it enforces a fixed set of fields across the dataset – even if some of them are None for a given entry. Since my dataset contains heterogeneous conversation messages (some with images, others with text), Arrow is injecting None for whichever attributes don’t exist in each message type to preserve a unified schema.\n\nThis results in a lot of meaningless or misleading Nones across the dataset, and breaks logic further down the pipeline – in particular, it causes the qwen_vl_utils functions (like process_vision_info) to crash or misbehave.\n\n\n🤔 Workarounds considered\n\nI could stick with raw JSON, which preserves the heterogeneous structure properly. But that means giving up all the nice features of datasets and Arrow (e.g. streaming, map/filter, etc.), which feels like a shame.\n\n❓My question\n\nDo you know of a clean way to define a flexible schema with Hugging Face Datasets – maybe using nested structures or dynamic fields – so that each conversation message doesn’t get forced into a flat structure with irrelevant keys? Or should I just stick with JSON for this kind of multimodal case?\n\nLet me know if you’ve run into this before or have any tips!\n",
"Datasets without fixed types are harder to use in many data frameworks unfortunately. You will have to handle the case with None values if you want to use Arrow/`datasets` (or any other framework using Arrow like spark, ray, dask). **IMO the short term solution is to fix qwen_vl_utils / process_vision_info.**\n\nAlternatively we can explore adding the Arrow `Json` type to `datasets`, but doesn't allow having image types in the Json object so I don't think this is the right solution.",
"I totally share the analysis, I am happy to try to help with this! I am currently using Unsloth so it is the `UnslothVisionDataCollator` which has almost the same logic as `qwen_vl_utils.process_vision_info`. Is there a Huggingface class I could patch too ? "
] |
2,731,626,760
| 7,321
|
ImportError: cannot import name 'set_caching_enabled' from 'datasets'
|
open
| 2024-12-11T01:58:46
| 2024-12-11T13:32:15
| null |
https://github.com/huggingface/datasets/issues/7321
| null |
sankexin
| false
|
[
"pip install datasets==2.18.0",
"Hi ! I think you need to update axolotl"
] |
2,731,112,100
| 7,320
|
ValueError: You should supply an encoding or a list of encodings to this method that includes input_ids, but you provided ['label']
|
closed
| 2024-12-10T20:23:11
| 2024-12-10T23:22:23
| 2024-12-10T23:22:23
|
https://github.com/huggingface/datasets/issues/7320
| null |
atrompeterog
| false
|
[
"Now i have other error"
] |
2,730,679,980
| 7,319
|
set dev version
|
closed
| 2024-12-10T17:01:34
| 2024-12-10T17:04:04
| 2024-12-10T17:01:45
|
https://github.com/huggingface/datasets/pull/7319
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7319",
"html_url": "https://github.com/huggingface/datasets/pull/7319",
"diff_url": "https://github.com/huggingface/datasets/pull/7319.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7319.patch",
"merged_at": "2024-12-10T17:01:45"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7319). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,730,676,278
| 7,318
|
Introduce support for PDFs
|
open
| 2024-12-10T16:59:48
| 2024-12-12T18:38:13
| null |
https://github.com/huggingface/datasets/issues/7318
| null |
yabramuvdi
| false
|
[
"#self-assign",
"Awesome ! Let me know if you have any question or if I can help :)\r\n\r\ncc @AndreaFrancis as well for viz",
"Other candidates libraries for the Pdf type: PyMuPDF pypdf and pdfplumber\r\n\r\nEDIT: Pymupdf looks like a good choice when it comes to maturity + performance + versatility BUT the license is maybe an issue, and pypdf, pypdfium2 or pdfplumber are good options imo",
"Related to https://github.com/huggingface/datasets/issues/7058",
"PyMuPDF is AGPL licensed, so we can't use it. I will move forward with [pdfplumber](https://github.com/jsvine/pdfplumber?tab=readme-ov-file#python-library).",
"Hi both! I have made a pull request with a first basic implementation of the Pdf feature. I followed closely what I saw on the Video and Image features. It is my first time contributing so any comments are very welcomed. I think it would be useful to outline together what additional things we can implement (e.g. enabling parsing of the pdf). Thanks :) "
] |
2,730,661,237
| 7,317
|
Release: 3.2.0
|
closed
| 2024-12-10T16:53:20
| 2024-12-10T16:56:58
| 2024-12-10T16:56:56
|
https://github.com/huggingface/datasets/pull/7317
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7317",
"html_url": "https://github.com/huggingface/datasets/pull/7317",
"diff_url": "https://github.com/huggingface/datasets/pull/7317.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7317.patch",
"merged_at": "2024-12-10T16:56:56"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7317). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,730,196,085
| 7,316
|
More docs to from_dict to mention that the result lives in RAM
|
closed
| 2024-12-10T13:56:01
| 2024-12-10T13:58:32
| 2024-12-10T13:57:02
|
https://github.com/huggingface/datasets/pull/7316
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7316",
"html_url": "https://github.com/huggingface/datasets/pull/7316",
"diff_url": "https://github.com/huggingface/datasets/pull/7316.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7316.patch",
"merged_at": "2024-12-10T13:57:02"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7316). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,727,502,630
| 7,314
|
Resolved for empty datafiles
|
open
| 2024-12-09T15:47:22
| 2024-12-27T18:20:21
| null |
https://github.com/huggingface/datasets/pull/7314
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7314",
"html_url": "https://github.com/huggingface/datasets/pull/7314",
"diff_url": "https://github.com/huggingface/datasets/pull/7314.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7314.patch",
"merged_at": null
}
|
sahillihas
| true
|
[
"Closes #6152 ",
"@mariosasko I hope this resolves #6152 "
] |
2,726,240,634
| 7,313
|
Cannot create a dataset with relative audio path
|
open
| 2024-12-09T07:34:20
| 2025-04-19T07:13:08
| null |
https://github.com/huggingface/datasets/issues/7313
| null |
sedol1339
| false
|
[
"Hello ! when you `cast_column` you need the paths to be absolute paths or relative paths to your working directory, not the original dataset directory.\r\n\r\nThough I'd recommend structuring your dataset as an AudioFolder which automatically links a metadata.jsonl or csv to the audio files via relative paths **within** the dataset repository: https://huggingface.co/docs/datasets/v3.2.0/en/audio_load#audiofolder",
"@lhoestq thank you, but there are two problems with using AudioFolder:\r\n1. It is said that AudioFolder requires metadata.csv. However, my datset is too large and contains nested and np.ndarray fields, so I can't use csv.\r\n2. It is said that I need to load the dataset with `load_dataset(\"audiofolder\", ...)`. However, if possible, I want my dataset to be loaded as usual with `load_dataset(dataset_name)` after I upload if to HF.",
"You can use metadata.jsonl if you have nested data :)\r\n\r\nAnd actually if you have a dataset structured as an AudioFolder then `load_dataset(dataset_name)` does work after uploading to HF",
"I have created an audio dataset. In my repo, I have explained the steps and structure. An example dataset is also available in the repo. https://github.com/pr0mila/ParquetToHuggingFace "
] |
2,725,103,094
| 7,312
|
[Audio Features - DO NOT MERGE] PoC for adding an offset+sliced reading to audio file.
|
open
| 2024-12-08T10:27:31
| 2024-12-08T10:27:31
| null |
https://github.com/huggingface/datasets/pull/7312
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7312",
"html_url": "https://github.com/huggingface/datasets/pull/7312",
"diff_url": "https://github.com/huggingface/datasets/pull/7312.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7312.patch",
"merged_at": null
}
|
TParcollet
| true
|
[] |
2,725,002,630
| 7,311
|
How to get the original dataset name with username?
|
open
| 2024-12-08T07:18:14
| 2025-01-09T10:48:02
| null |
https://github.com/huggingface/datasets/issues/7311
| null |
npuichigo
| false
|
[
"Hi ! why not pass the dataset id to Ray and let it check the parquet files ? Or pass the parquet files lists directly ?",
"I'm not sure why ray design an API like this to accept a `Dataset` object, so they need to verify the `Dataset` is the original one and use the `DatasetInfo` to query the huggingface hub. I'll advise the ray data team to use dataset id instead of dataset for the `HuggingFaceDatasource` API."
] |
2,724,830,603
| 7,310
|
Enable the Audio Feature to decode / read with an offset + duration
|
open
| 2024-12-07T22:01:44
| 2024-12-09T21:09:46
| null |
https://github.com/huggingface/datasets/issues/7310
| null |
TParcollet
| false
|
[
"Hi ! What about having audio + start + duration columns and enable something like this ?\r\n\r\n```python\r\nfor example in ds:\r\n array = example[\"audio\"].read(start=example[\"start\"], frames=example[\"duration\"])\r\n```",
"Hi @lhoestq, this would work with a file-based dataset but would be terrible for a sharded one as it would duplicate the large audio file many times. Also, very long audio files are not embedded very well in the parquet file, even with large_binary(). It crashed a few times for me until I switched to one sample == one file :-( "
] |
2,729,738,963
| 7,315
|
Allow manual configuration of Dataset Viewer for datasets not created with the `datasets` library
|
open
| 2024-12-07T16:37:12
| 2024-12-11T11:05:22
| null |
https://github.com/huggingface/datasets/issues/7315
| null |
diarray-hub
| false
|
[
"Hi @diarray-hub , thanks for opening the issue :) Let me ping @lhoestq and @severo from the dataset viewer team :hugs: ",
"amazing :)",
"Hi ! why not modify the manifest.json file directly ? this way users see in the viewer the dataset as is instead which makes it easier to use using e.g. the `datasets` library",
"Can I create and push the dataset with the dataset library while also pushing the dataset directory, mainting its structure and all the files as with git? ",
"(I transferred to the issue to the `datasets` repo as it's not related to `huggingface_hub`)",
"> Can I create and push the dataset with the dataset library while also pushing the dataset directory, mainting its structure and all the files as with git?\r\n\r\nyes push_to_hub simply uploads Parquet files in a directory named \"data\" in the git repository\r\n",
"That's the problem actually, I need that the data stays in the same format and the directory they are in keep the same structure in order to go quick with Nemo training so users of Nvidia's Nemo framework don't need to write any preprocessing code before starting training. That's why I used git instead of push_to_hub so me and other users working with Nemo can just:\r\n1. git clone\r\n2. asr_model.setup_training_data(train_data_config={'manifest_filepath': training_manifest_filepath})\r\n\r\nAnd start training already. It may be not very kind of me to prioritize users of a specific framework but I noticed that it take much more code to convert an huggingFace dataset with the parquet file to Nemo manifest format than the inverse :haha: ",
"Happy to help if you think the Nemo dataset format should be supported in `datasets` (and therefore in the HF Viewer that is based on `datasets`). Maybe the Nemo team could help as well\r\n\r\nThough I'm not sure if there is only one but actually many formats/structure in Nemo depending on the task ?",
"Yeah, you're right Quentin, it depends of the task. This one is for ASR. And, yes maybe they can help. I noticed that they already share their models through HF. Maybe someone in your teams already have a contact point there. Anyway it's not really a big issues since people can easily understand the dataset and its format with the dataset card but it's a little annoying for those who wanna visually explore each features with the viewer as for regular HF datasets",
"In that case I'd recommend you to upload the dataset in Nemo format and \r\n1) add the \"nemo\" tag\r\n2) add how to use the dataset in Nemo in the dataset README.md\r\n\r\nThe viewer is likely to show the audio content by default but without the transcriptions. You can also configure the viewer to show the transcriptions instead (without the audio).",
"I already did, it's just a little bit \"dommage\" (Hope you'll understand, you speak french right? Cause I don't know any english word for this) that I have to choose which one the viewer displays. But it's no problem for the usability of the dataset. Thanks Quentin :+1: ",
"It's \"dommage\" for now, but feel free to ping the Nemo people if you think there is room for making this better together :)\r\n\r\nKinda related, but the `datasets` AudioFolder structure looks similar and simply asks for a `metadata.jsonl` with a field named \"file_name\" to link the transcriptions to the audio files - you could also add this file to your repository to make the viewer show audio + transcripts.\r\n\r\nAlternatively maybe we can expand the AudioFolder configuration to allow you to set the metadata file to be the \"manifest.json\" and the linking field to be \"audio_file_name\" (we just need to agree on something general - not just for Nemo)",
"Right, actually that was my idea when I opened this issues. That's what I suggested, taking my case as an exemple but you should think of a more general approach like adding a field to configure the viewer as you wish in the metadata (in the dataset card) or a config.yaml or json file. With a level of abstraction like the solution I proposed ot even higher abstraction, it would allow for more customizability :)"
] |
2,723,636,931
| 7,309
|
Faster parquet streaming + filters with predicate pushdown
|
closed
| 2024-12-06T18:01:54
| 2024-12-07T23:32:30
| 2024-12-07T23:32:28
|
https://github.com/huggingface/datasets/pull/7309
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7309",
"html_url": "https://github.com/huggingface/datasets/pull/7309",
"diff_url": "https://github.com/huggingface/datasets/pull/7309.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7309.patch",
"merged_at": "2024-12-07T23:32:28"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7309). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,720,244,889
| 7,307
|
refactor: remove unnecessary else
|
open
| 2024-12-05T12:11:09
| 2024-12-06T15:11:33
| null |
https://github.com/huggingface/datasets/pull/7307
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7307",
"html_url": "https://github.com/huggingface/datasets/pull/7307",
"diff_url": "https://github.com/huggingface/datasets/pull/7307.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7307.patch",
"merged_at": null
}
|
HarikrishnanBalagopal
| true
|
[] |
2,719,807,464
| 7,306
|
Creating new dataset from list loses information. (Audio Information Lost - either Datatype or Values).
|
open
| 2024-12-05T09:07:53
| 2024-12-05T09:09:38
| null |
https://github.com/huggingface/datasets/issues/7306
| null |
ai-nikolai
| false
|
[] |
2,715,907,267
| 7,305
|
Build Documentation Test Fails Due to "Bad Credentials" Error
|
open
| 2024-12-03T20:22:54
| 2025-01-08T22:38:14
| null |
https://github.com/huggingface/datasets/issues/7305
| null |
ruidazeng
| false
|
[
"how were you able to fix this please?",
"> how were you able to fix this please?\r\n\r\nI was not able to fix this."
] |
2,715,179,811
| 7,304
|
Update iterable_dataset.py
|
closed
| 2024-12-03T14:25:42
| 2024-12-03T14:28:10
| 2024-12-03T14:27:02
|
https://github.com/huggingface/datasets/pull/7304
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7304",
"html_url": "https://github.com/huggingface/datasets/pull/7304",
"diff_url": "https://github.com/huggingface/datasets/pull/7304.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7304.patch",
"merged_at": "2024-12-03T14:27:02"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7304). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,705,729,696
| 7,303
|
DataFilesNotFoundError for datasets LM1B
|
closed
| 2024-11-29T17:27:45
| 2024-12-11T13:22:47
| 2024-12-11T13:22:47
|
https://github.com/huggingface/datasets/issues/7303
| null |
hml1996-fight
| false
|
[
"Hi ! Can you try with a more recent version of `datasets` ? Also you might need to pass trust_remote_code=True since it's a script based dataset"
] |
2,702,626,386
| 7,302
|
Let server decide default repo visibility
|
closed
| 2024-11-28T16:01:13
| 2024-11-29T17:00:40
| 2024-11-29T17:00:38
|
https://github.com/huggingface/datasets/pull/7302
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7302",
"html_url": "https://github.com/huggingface/datasets/pull/7302",
"diff_url": "https://github.com/huggingface/datasets/pull/7302.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7302.patch",
"merged_at": "2024-11-29T17:00:38"
}
|
Wauplin
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7302). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"No need for a specific version of huggingface_hub to avoid a breaking change no (it's a server-side change)"
] |
2,701,813,922
| 7,301
|
update load_dataset doctring
|
closed
| 2024-11-28T11:19:20
| 2024-11-29T10:31:43
| 2024-11-29T10:31:40
|
https://github.com/huggingface/datasets/pull/7301
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7301",
"html_url": "https://github.com/huggingface/datasets/pull/7301",
"diff_url": "https://github.com/huggingface/datasets/pull/7301.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7301.patch",
"merged_at": "2024-11-29T10:31:40"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7301). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,701,424,320
| 7,300
|
fix: update elasticsearch version
|
closed
| 2024-11-28T09:14:21
| 2024-12-03T14:36:56
| 2024-12-03T14:24:42
|
https://github.com/huggingface/datasets/pull/7300
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7300",
"html_url": "https://github.com/huggingface/datasets/pull/7300",
"diff_url": "https://github.com/huggingface/datasets/pull/7300.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7300.patch",
"merged_at": "2024-12-03T14:24:42"
}
|
ruidazeng
| true
|
[
"May I request a review @lhoestq",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7300). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,695,378,251
| 7,299
|
Efficient Image Augmentation in Hugging Face Datasets
|
open
| 2024-11-26T16:50:32
| 2024-11-26T16:53:53
| null |
https://github.com/huggingface/datasets/issues/7299
| null |
fabiozappo
| false
|
[] |
2,694,196,968
| 7,298
|
loading dataset issue with load_dataset() when training controlnet
|
open
| 2024-11-26T10:50:18
| 2024-11-26T10:50:18
| null |
https://github.com/huggingface/datasets/issues/7298
| null |
sarahahtee
| false
|
[] |
2,683,977,430
| 7,297
|
wrong return type for `IterableDataset.shard()`
|
closed
| 2024-11-22T17:25:46
| 2024-12-03T14:27:27
| 2024-12-03T14:27:03
|
https://github.com/huggingface/datasets/issues/7297
| null |
ysngshn
| false
|
[
"Oops my bad ! thanks for reporting"
] |
2,675,573,974
| 7,296
|
Remove upper version limit of fsspec[http]
|
closed
| 2024-11-20T11:29:16
| 2025-03-06T04:47:04
| 2025-03-06T04:47:01
|
https://github.com/huggingface/datasets/pull/7296
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7296",
"html_url": "https://github.com/huggingface/datasets/pull/7296",
"diff_url": "https://github.com/huggingface/datasets/pull/7296.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7296.patch",
"merged_at": null
}
|
cyyever
| true
|
[] |
2,672,003,384
| 7,295
|
[BUG]: Streaming from S3 triggers `unexpected keyword argument 'requote_redirect_url'`
|
open
| 2024-11-19T12:23:36
| 2024-11-19T13:01:53
| null |
https://github.com/huggingface/datasets/issues/7295
| null |
casper-hansen
| false
|
[] |
2,668,663,130
| 7,294
|
Remove `aiohttp` from direct dependencies
|
closed
| 2024-11-18T14:00:59
| 2025-05-07T14:27:18
| 2025-05-07T14:27:17
|
https://github.com/huggingface/datasets/pull/7294
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7294",
"html_url": "https://github.com/huggingface/datasets/pull/7294",
"diff_url": "https://github.com/huggingface/datasets/pull/7294.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7294.patch",
"merged_at": "2025-05-07T14:27:17"
}
|
akx
| true
|
[] |
2,664,592,054
| 7,293
|
Updated inconsistent output in documentation examples for `ClassLabel`
|
closed
| 2024-11-16T16:20:57
| 2024-12-06T11:33:33
| 2024-12-06T11:32:01
|
https://github.com/huggingface/datasets/pull/7293
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7293",
"html_url": "https://github.com/huggingface/datasets/pull/7293",
"diff_url": "https://github.com/huggingface/datasets/pull/7293.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7293.patch",
"merged_at": "2024-12-06T11:32:01"
}
|
sergiopaniego
| true
|
[
"Updated! 😄 ",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7293). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"@lhoestq, can you help with this failing test please? 🙏 "
] |
2,664,250,855
| 7,292
|
DataFilesNotFoundError for datasets `OpenMol/PubChemSFT`
|
closed
| 2024-11-16T11:54:31
| 2024-11-19T00:53:00
| 2024-11-19T00:52:59
|
https://github.com/huggingface/datasets/issues/7292
| null |
xnuohz
| false
|
[
"Hi ! If the dataset owner uses `push_to_hub()` instead of `save_to_disk()` and upload the local files it will fix the issue.\r\nRight now `datasets` sees the train/test/valid pickle files but they are not supported file formats.",
"Alternatively you can load the arrow file instead:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('OpenMol/PubChemSFT', data_files='stage1/*.arrow')\r\n```",
"Thanks! I'll have a try."
] |
2,662,244,643
| 7,291
|
Why return_tensors='pt' doesn't work?
|
open
| 2024-11-15T15:01:23
| 2024-11-18T13:47:08
| null |
https://github.com/huggingface/datasets/issues/7291
| null |
bw-wang19
| false
|
[
"Hi ! `datasets` uses Arrow as storage backend which is agnostic to deep learning frameworks like torch. If you want to get torch tensors back, you need to do `dataset = dataset.with_format(\"torch\")`",
"> Hi ! `datasets` uses Arrow as storage backend which is agnostic to deep learning frameworks like torch. If you want to get torch tensors back, you need to do `dataset = dataset.with_format(\"torch\")`\r\n\r\nIt does work! Thanks for your suggestion!"
] |
2,657,620,816
| 7,290
|
`Dataset.save_to_disk` hangs when using num_proc > 1
|
open
| 2024-11-14T05:25:13
| 2025-06-27T00:56:47
| null |
https://github.com/huggingface/datasets/issues/7290
| null |
JohannesAck
| false
|
[
"I've met the same situations.\r\n\r\nHere's my logs:\r\nnum_proc = 64, I stop it early as it cost **too** much time.\r\n```\r\nSaving the dataset (1540/4775 shards): 32%|███▏ | 47752224/147853764 [15:32:54<132:28:34, 209.89 examples/s]\r\nSaving the dataset (1540/4775 shards): 32%|███▏ | 47754224/147853764 [15:32:54<78:14:37, 355.37 examples/s] \r\nSaving the dataset (1540/4775 shards): 32%|███▏ | 47755224/147853764 [15:32:59<93:43:45, 296.65 examples/s]\r\n```\r\n\r\nnum_proc = 1(Not set num_proc parameter)\r\n```\r\nSaving the dataset (1753/4775 shards): 37%|███▋ | 54301556/147853764 [24:46<38:33, 40440.93 examples/s]\r\nSaving the dataset (1753/4775 shards): 37%|███▋ | 54306556/147853764 [24:46<39:34, 39392.01 examples/s]\r\nSaving the dataset (1753/4775 shards): 37%|███▋ | 54311520/147853764 [24:46<38:56, 40030.53 examples/s]\r\n```\r\n\r\nI check the conditions of CPUs and Memory I/O, I found that disk I/O was blocked, but CPU and memory usage was high. There should be some bugs in the code.\r\n\r\n",
"Any new process on this issue? I'm encountering the same issue.",
"Not getting this issue. \n\nMy output;\n\n`Saving the dataset (0/87 shards): 7%|▎ | 588000/8557560 [01:29<21:02, 6314.04 examples/s]`\n\nAt setting `num_proc=64`, and \n\n`Saving the dataset (0/87 shards): 0%| | 28000/8557560 [03:20<16:59:06, 139.49 examples/s]` \n\nAt num_proc=1 (pass nothing)\n\nMy `pyproject.toml`; \n\n```\n[project]\nname = \"test\"\nversion = \"0.1.0\"\nrequires-python = \"==3.13.0\"\ndependencies = [\n \"absl-py==2.1.0\",\n \"accelerate==1.7.0\",\n \"aiohappyeyeballs==2.6.1\",\n \"aiohttp==3.12.11\",\n \"aiosignal==1.3.2\",\n \"annotated-types==0.7.0\",\n \"appdirs==1.4.4\",\n \"argcomplete>=1.8.1\",\n \"astunparse==1.6.3\",\n \"async-timeout==5.0.1\",\n \"attrs==21.2.0\",\n \"automat==20.2.0\",\n \"babel==2.8.0\",\n \"backcall==0.2.0\",\n \"bcrypt==3.2.0\",\n \"beautifulsoup4==4.10.0\",\n \"beniget==0.4.2\",\n \"bleach==4.1.0\",\n \"blinker==1.4\",\n \"blis==1.3.0\",\n \"bottle==0.12.19\",\n \"brotli==1.0.9\",\n \"catalogue==2.0.10\",\n \"certifi==2020.6.20\",\n \"cffi==1.15.0\",\n \"chardet==4.0.0\",\n \"charset-normalizer==3.4.2\",\n \"click==8.0.3\",\n \"cloudpathlib==0.21.1\",\n \"colorama>=0.4.4\",\n \"commonmark==0.9.1\",\n \"confection==0.1.5\",\n \"configobj==5.0.6\",\n \"constantly==15.1.0\",\n \"cryptography==3.4.8\",\n \"ctop==1.0.0\",\n \"cycler==0.11.0\",\n \"cymem==2.0.11\",\n \"datasets==3.6.0\",\n \"dbus-python==1.2.18\",\n \"decorator==4.4.2\",\n \"defusedxml==0.7.1\",\n \"dill==0.3.8\",\n \"distlib==0.3.4\",\n \"distro==1.7.0\",\n \"einops==0.8.1\",\n \"en-core-web-sm\",\n \"entrypoints==0.4\",\n \"evaluate==0.4.3\",\n \"filelock==3.6.0\",\n \"flake8==4.0.1\",\n \"flatbuffers==25.2.10\",\n \"fonttools==4.29.1\",\n \"frozenlist==1.7.0\",\n \"fs==2.4.12\",\n \"fsspec==2024.3.1\",\n \"future==0.18.2\",\n \"gast==0.6.0\",\n \"gitdb==4.0.12\",\n \"gitpython==3.1.44\",\n \"glances==3.2.4.2\",\n \"google-pasta==0.2.0\",\n \"grpcio==1.73.0\",\n \"h5py==3.14.0\",\n \"hf-xet==1.1.3\",\n \"html5lib==1.1\",\n \"httplib2==0.20.2\",\n \"huggingface-hub==0.32.4\",\n \"hyperlink==21.0.0\",\n \"icdiff==2.0.4\",\n \"idna==3.3\",\n \"importlib-metadata==4.6.4\",\n \"incremental==21.3.0\",\n \"influxdb==5.3.1\",\n \"ipykernel==6.7.0\",\n \"ipython==7.31.1\",\n \"ipython-genutils==0.2.0\",\n \"jax>=0.5.1\",\n \"jax-cuda12-pjrt>=0.5.1\",\n \"jax-cuda12-plugin>=0.5.1\",\n \"jaxlib>=0.5.1\",\n \"jedi==0.18.0\",\n \"jeepney==0.7.1\",\n \"jinja2==3.0.3\",\n \"joblib>=0.17.0\",\n \"jsonpatch==1.32\",\n \"jsonpointer==2.0\",\n \"jsonschema==3.2.0\",\n \"jupyter-client==7.1.2\",\n \"jupyter-core==4.9.1\",\n \"kaptan>=0.5.12\",\n \"keras==3.6.0\",\n \"keyring==23.5.0\",\n \"kiwisolver==1.3.2\",\n \"langcodes==3.5.0\",\n \"language-data==1.3.0\",\n \"launchpadlib==1.10.16\",\n \"lazr-restfulclient==0.14.4\",\n \"lazr-uri==1.0.6\",\n \"libclang==18.1.1\",\n \"libtmux==0.10.1\",\n \"livereload==2.6.3\",\n \"lxml==6.0.0\",\n \"lz4==3.1.3\",\n \"marisa-trie==1.2.1\",\n \"markdown==3.3.6\",\n \"markupsafe==2.0.1\",\n \"matplotlib==3.5.1\",\n \"matplotlib-inline==0.1.3\",\n \"mccabe==0.6.1\",\n \"mkdocs==1.1.2\",\n \"ml-dtypes==0.5.1\",\n \"more-itertools==8.10.0\",\n \"mpmath>=0.0.0\",\n \"msgpack==1.0.3\",\n \"multidict==6.4.4\",\n \"multiprocess==0.70.16\",\n \"murmurhash==1.0.13\",\n \"namex==0.0.8\",\n \"nest-asyncio==1.5.4\",\n \"netifaces==0.11.0\",\n \"networkx==2.4\",\n \"nltk==3.9.1\",\n \"numpy>=1.23.3\",\n \"nvidia-ml-py==12.555.43\",\n \"oauthlib==3.2.0\",\n \"olefile==0.46\",\n \"opt-einsum==3.3.0\",\n \"optree==0.13.1\",\n \"packaging==21.3\",\n \"pandas==2.3.0\",\n \"parso==0.8.1\",\n \"pexpect==4.8.0\",\n \"pickleshare==0.7.5\",\n \"pillow>=9.0.1\",\n \"pipx==1.0.0\",\n \"platformdirs==2.5.1\",\n \"ply==3.11\",\n \"portalocker==3.1.1\",\n \"preshed==3.0.10\",\n \"prompt-toolkit==3.0.28\",\n \"propcache==0.3.2\",\n \"protobuf==4.21.12\",\n \"psutil==5.9.0\",\n \"ptyprocess==0.7.0\",\n \"py==1.10.0\",\n \"pyarrow==20.0.0\",\n \"pyasn1==0.4.8\",\n \"pyasn1-modules==0.2.1\",\n \"pycodestyle==2.8.0\",\n \"pycparser==2.21\",\n \"pycryptodomex==3.11.0\",\n \"pydantic==2.11.5\",\n \"pydantic-core==2.33.2\",\n \"pyflakes==2.4.0\",\n \"pygments==2.11.2\",\n \"pyhamcrest==2.0.2\",\n \"pyinotify==0.9.6\",\n \"pyjwt==2.3.0\",\n \"pyopenssl==21.0.0\",\n \"pyparsing==2.4.7\",\n \"pyrsistent==0.18.1\",\n \"pyserial==3.5\",\n \"pysmi==0.3.2\",\n \"pysnmp==4.4.12\",\n \"pystache==0.6.0\",\n \"python-dateutil>=2.8.1\",\n \"python-magic==0.4.24\",\n \"pythran>=0.10.0\",\n \"pytz==2022.1\",\n \"pyyaml==6.0.2\",\n \"regex==2024.11.6\",\n \"requests==2.32.4\",\n \"rich==11.2.0\",\n \"rouge-score==0.1.2\",\n \"sacrebleu==2.5.1\",\n \"safetensors==0.5.3\",\n \"scikit-learn==1.7.0\",\n \"scipy>=1.8.0\",\n \"secretstorage==3.3.1\",\n \"sentence-transformers==4.1.0\",\n \"sentry-sdk==2.29.1\",\n \"service-identity==18.1.0\",\n \"setproctitle==1.3.6\",\n \"shellingham==1.5.4\",\n \"six==1.16.0\",\n \"smart-open==7.1.0\",\n \"smmap==5.0.2\",\n \"soupsieve==2.3.1\",\n \"spacy==3.8.7\",\n \"spacy-legacy==3.0.12\",\n \"spacy-loggers==1.0.5\",\n \"srsly==2.5.1\",\n \"ssh-import-id==5.11\",\n \"sympy>=1.12\",\n \"tabulate==0.9.0\",\n \"tensorboard==2.19.0\",\n \"tensorboard-data-server==0.7.2\",\n \"termcolor==1.1.0\",\n \"thinc==8.3.6\",\n \"threadpoolctl==3.1.0\",\n \"tmuxp==1.9.2\",\n \"tokenizers==0.21.1\",\n \"torch==2.6.0\",\n \"torchvision==0.21.0\",\n \"tornado==6.1\",\n \"tqdm==4.67.1\",\n \"traitlets==5.1.1\",\n \"transformers==4.52.4\",\n \"triton==3.2.0\",\n \"twisted==22.1.0\",\n \"typer==0.16.0\",\n \"typing-extensions==4.14.0\",\n \"typing-inspection==0.4.1\",\n \"ufolib2==0.13.1\",\n \"urllib3==2.4.0\",\n \"userpath==1.8.0\",\n \"virtualenv==20.13.0\",\n \"wadllib==1.3.6\",\n \"wandb==0.20.1\",\n \"wasabi==1.1.3\",\n \"wcwidth==0.2.5\",\n \"weasel==0.4.1\",\n \"webencodings==0.5.1\",\n \"websocket-client==1.2.3\",\n \"werkzeug==2.0.2\",\n \"xxhash==3.5.0\",\n \"yarl==1.20.1\",\n \"zipp==1.0.0\",\n \"zope-interface==5.4.0\",\n]\n\n[tool.uv.sources]\nen-core-web-sm = { url = \"https://github.com/explosion/spacy-models/releases/download/en_core_web_sm-3.8.0/en_core_web_sm-3.8.0-py3-none-any.whl\" }\n```"
] |
2,648,019,507
| 7,289
|
Dataset viewer displays wrong statists
|
closed
| 2024-11-11T03:29:27
| 2024-11-13T13:02:25
| 2024-11-13T13:02:25
|
https://github.com/huggingface/datasets/issues/7289
| null |
speedcell4
| false
|
[
"i think this issue is more for https://github.com/huggingface/dataset-viewer"
] |
2,647,052,280
| 7,288
|
Release v3.1.1
|
closed
| 2024-11-10T09:38:15
| 2024-11-10T09:38:48
| 2024-11-10T09:38:48
|
https://github.com/huggingface/datasets/pull/7288
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7288",
"html_url": "https://github.com/huggingface/datasets/pull/7288",
"diff_url": "https://github.com/huggingface/datasets/pull/7288.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7288.patch",
"merged_at": null
}
|
alex-hh
| true
|
[] |
2,646,958,393
| 7,287
|
Support for identifier-based automated split construction
|
open
| 2024-11-10T07:45:19
| 2024-11-19T14:37:02
| null |
https://github.com/huggingface/datasets/issues/7287
| null |
alex-hh
| false
|
[
"Hi ! You can already configure the README.md to have multiple sets of splits, e.g.\r\n\r\n```yaml\r\nconfigs:\r\n- config_name: my_first_set_of_split\r\n data_files:\r\n - split: train\r\n path: *.csv\r\n- config_name: my_second_set_of_split\r\n data_files:\r\n - split: train\r\n path: train-*.csv\r\n - split: test\r\n path: test-*.csv\r\n```",
"Hi - I had something slightly different in mind:\r\n\r\nCurrently the yaml splits specified like this only allow specifying which filenames to pass to each split.\r\nBut what if I have a situation where I know which individual *training examples* I want to put in each split.\r\n\r\nI could build split-specific files, however for large datasets with overlapping (e.g. multiple sets of) splits this could result in significant duplication of data.\r\n\r\nI can see that this could actually be very much intended (i.e. to discourage overlapping splits), but wondered whether some support for handling splits based on individual identifiers is something that could be considered. ",
"This is not supported right now :/ Though you can load the data in two steps like this\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nfull_dataset = load_dataset(\"username/dataset\", split=\"train\")\r\nmy_first_set_indices = load_dataset(\"username/dataset\", \"my_first_set_of_split\", split=\"train\")\r\n\r\nmy_first_set = full_dataset.select(my_first_set_indices[\"indices\"])\r\n```\r\n\r\nyou can create such a dataset by adapting this code for example\r\n```python\r\n# upload the full dataset\r\nfull_dataset.push_to_hub(\"username/dataset\")\r\n# then upload the indices for each set\r\nDatasetDict({\r\n \"train\": Dataset.from_dict({\"indices\": [0, 1, 2, 3]}),\r\n \"test\": Dataset.from_dict({\"indices\": [4, 5]}),\r\n}).push_to_hub(\"username/dataset\", \"my_first_set_of_split\")"
] |
2,645,350,151
| 7,286
|
Concurrent loading in `load_from_disk` - `num_proc` as a param
|
closed
| 2024-11-08T23:21:40
| 2024-11-09T16:14:37
| 2024-11-09T16:14:37
|
https://github.com/huggingface/datasets/issues/7286
| null |
unography
| false
|
[] |
2,644,488,598
| 7,285
|
Release v3.1.0
|
closed
| 2024-11-08T16:17:58
| 2024-11-08T16:18:05
| 2024-11-08T16:18:05
|
https://github.com/huggingface/datasets/pull/7285
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7285",
"html_url": "https://github.com/huggingface/datasets/pull/7285",
"diff_url": "https://github.com/huggingface/datasets/pull/7285.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7285.patch",
"merged_at": null
}
|
alex-hh
| true
|
[] |
2,644,302,386
| 7,284
|
support for custom feature encoding/decoding
|
closed
| 2024-11-08T15:04:08
| 2024-11-21T16:09:47
| 2024-11-21T16:09:47
|
https://github.com/huggingface/datasets/pull/7284
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7284",
"html_url": "https://github.com/huggingface/datasets/pull/7284",
"diff_url": "https://github.com/huggingface/datasets/pull/7284.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7284.patch",
"merged_at": "2024-11-21T16:09:47"
}
|
alex-hh
| true
|
[
"@lhoestq ",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7284). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,642,537,708
| 7,283
|
Allow for variation in metadata file names as per issue #7123
|
open
| 2024-11-08T00:44:47
| 2024-11-08T00:44:47
| null |
https://github.com/huggingface/datasets/pull/7283
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7283",
"html_url": "https://github.com/huggingface/datasets/pull/7283",
"diff_url": "https://github.com/huggingface/datasets/pull/7283.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7283.patch",
"merged_at": null
}
|
egrace479
| true
|
[] |
2,642,075,491
| 7,282
|
Faulty datasets.exceptions.ExpectedMoreSplitsError
|
open
| 2024-11-07T20:15:01
| 2024-11-07T20:15:42
| null |
https://github.com/huggingface/datasets/issues/7282
| null |
meg-huggingface
| false
|
[] |
2,640,346,339
| 7,281
|
File not found error
|
open
| 2024-11-07T09:04:49
| 2024-11-07T09:22:43
| null |
https://github.com/huggingface/datasets/issues/7281
| null |
MichielBontenbal
| false
|
[
"Link to the dataset: https://huggingface.co/datasets/MichielBontenbal/UrbanSounds "
] |
2,639,977,077
| 7,280
|
Add filename in error message when ReadError or similar occur
|
open
| 2024-11-07T06:00:53
| 2024-11-20T13:23:12
| null |
https://github.com/huggingface/datasets/issues/7280
| null |
elisa-aleman
| false
|
[
"Hi Elisa, please share the error traceback here, and if you manage to find the location in the `datasets` code where the error occurs, feel free to open a PR to add the necessary logging / improve the error message.",
"> please share the error traceback\n\nI don't have access to it but it should be during [this exception](https://github.com/huggingface/datasets/blob/2049c00921c59cdeb835137a1c49639cf175af07/src/datasets/builder.py#L1643) which happens during the loading of a dataset. If one of the downloaded files is corrupted, the for loop will not yield correctly, and the error will come from, say, in the case of tar files, [this iterable](https://github.com/huggingface/datasets/blob/2049c00921c59cdeb835137a1c49639cf175af07/src/datasets/utils/file_utils.py#L1293) which has no explicit error handling that leaves clues as to which file has failed.\n\nI only know the case for tar files, but I consider this issue could be happening across different file types too.",
"I think having a better error handling for this tar iterable would be useful already, maybe a simple try/except in `_iter_from_urlpath` that checks for `tarfile.ReadError` and raises an error with the `urlpath` mentioned in the error ?",
"I think not just from higher calls like the `_iter_from_urlpath` but directly wherever a file is attempted to be opened would be the best case, as the traceback would simply lead to that.",
"so maybe there should be better error messages in each dataset builder definition ? e.g. in https://github.com/huggingface/datasets/blob/main/src/datasets/packaged_modules/webdataset/webdataset.py for webdataset TAR archives"
] |
2,635,813,932
| 7,279
|
Feature proposal: Stacking, potentially heterogeneous, datasets
|
open
| 2024-11-05T15:40:50
| 2024-11-05T15:40:50
| null |
https://github.com/huggingface/datasets/pull/7279
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7279",
"html_url": "https://github.com/huggingface/datasets/pull/7279",
"diff_url": "https://github.com/huggingface/datasets/pull/7279.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7279.patch",
"merged_at": null
}
|
TimCares
| true
|
[] |
2,633,436,151
| 7,278
|
Let soundfile directly read local audio files
|
open
| 2024-11-04T17:41:13
| 2024-11-18T14:01:25
| null |
https://github.com/huggingface/datasets/pull/7278
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7278",
"html_url": "https://github.com/huggingface/datasets/pull/7278",
"diff_url": "https://github.com/huggingface/datasets/pull/7278.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7278.patch",
"merged_at": null
}
|
fawazahmed0
| true
|
[] |
2,632,459,184
| 7,277
|
Add link to video dataset
|
closed
| 2024-11-04T10:45:12
| 2024-11-04T17:05:06
| 2024-11-04T17:05:06
|
https://github.com/huggingface/datasets/pull/7277
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7277",
"html_url": "https://github.com/huggingface/datasets/pull/7277",
"diff_url": "https://github.com/huggingface/datasets/pull/7277.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7277.patch",
"merged_at": "2024-11-04T17:05:06"
}
|
NielsRogge
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7277). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,631,917,431
| 7,276
|
Accessing audio dataset value throws Format not recognised error
|
open
| 2024-11-04T05:59:13
| 2024-11-09T18:51:52
| null |
https://github.com/huggingface/datasets/issues/7276
| null |
fawazahmed0
| false
|
[
"Hi ! can you try if this works ?\r\n\r\n```python\r\nimport soundfile as sf\r\n\r\nwith open('C:\\\\Users\\\\Nawaz-Server\\\\.cache\\\\huggingface\\\\hub\\\\datasets--fawazahmed0--bug-audio\\\\snapshots\\\\fab1398431fed1c0a2a7bff0945465bab8b5daef\\\\data\\\\Ghamadi\\\\037136.mp3', 'rb') as f:\r\n print(sf.read(f))\r\n```",
"@lhoestq Same error, here is the output:\r\n\r\n```bash\r\n(mypy) C:\\Users\\Nawaz-Server\\Documents\\ml>python myest.py\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\Nawaz-Server\\Documents\\ml\\myest.py\", line 5, in <module>\r\n print(sf.read(f))\r\n ^^^^^^^^^^\r\n File \"C:\\Users\\Nawaz-Server\\.conda\\envs\\mypy\\Lib\\site-packages\\soundfile.py\", line 285, in read\r\n with SoundFile(file, 'r', samplerate, channels,\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\Users\\Nawaz-Server\\.conda\\envs\\mypy\\Lib\\site-packages\\soundfile.py\", line 658, in __init__\r\n self._file = self._open(file, mode_int, closefd)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\Users\\Nawaz-Server\\.conda\\envs\\mypy\\Lib\\site-packages\\soundfile.py\", line 1216, in _open\r\n raise LibsndfileError(err, prefix=\"Error opening {0!r}: \".format(self.name))\r\nsoundfile.LibsndfileError: Error opening <_io.BufferedReader name='C:\\\\Users\\\\Nawaz-Server\\\\.cache\\\\huggingface\\\\hub\\\\datasets--fawazahmed0--bug-audio\\\\snapshots\\\\fab1398431fed1c0a2a7bff0945465bab8b5daef\\\\data\\\\Ghamadi\\\\037136.mp3'>: Format not recognised.\r\n\r\n```",
"upstream bug: https://github.com/bastibe/python-soundfile/issues/439"
] |
2,631,713,397
| 7,275
|
load_dataset
|
open
| 2024-11-04T03:01:44
| 2024-11-04T03:01:44
| null |
https://github.com/huggingface/datasets/issues/7275
| null |
santiagobp99
| false
|
[] |
2,629,882,821
| 7,274
|
[MINOR:TYPO] Fix typo in exception text
|
closed
| 2024-11-01T21:15:29
| 2025-05-21T13:17:20
| 2025-05-21T13:17:20
|
https://github.com/huggingface/datasets/pull/7274
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7274",
"html_url": "https://github.com/huggingface/datasets/pull/7274",
"diff_url": "https://github.com/huggingface/datasets/pull/7274.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7274.patch",
"merged_at": null
}
|
cakiki
| true
|
[] |
2,628,896,492
| 7,273
|
Raise error for incorrect JSON serialization
|
closed
| 2024-11-01T11:54:35
| 2024-11-18T11:25:01
| 2024-11-18T11:25:01
|
https://github.com/huggingface/datasets/pull/7273
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7273",
"html_url": "https://github.com/huggingface/datasets/pull/7273",
"diff_url": "https://github.com/huggingface/datasets/pull/7273.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7273.patch",
"merged_at": "2024-11-18T11:25:01"
}
|
varadhbhatnagar
| true
|
[
"PTAL @lhoestq @albertvillanova ",
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7273). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,627,223,390
| 7,272
|
fix conda release worlflow
|
closed
| 2024-10-31T15:56:19
| 2024-10-31T15:58:35
| 2024-10-31T15:57:29
|
https://github.com/huggingface/datasets/pull/7272
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7272",
"html_url": "https://github.com/huggingface/datasets/pull/7272",
"diff_url": "https://github.com/huggingface/datasets/pull/7272.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7272.patch",
"merged_at": "2024-10-31T15:57:29"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7272). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,627,135,540
| 7,271
|
Set dev version
|
closed
| 2024-10-31T15:22:51
| 2024-10-31T15:25:27
| 2024-10-31T15:22:59
|
https://github.com/huggingface/datasets/pull/7271
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7271",
"html_url": "https://github.com/huggingface/datasets/pull/7271",
"diff_url": "https://github.com/huggingface/datasets/pull/7271.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7271.patch",
"merged_at": "2024-10-31T15:22:59"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7271). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,627,107,016
| 7,270
|
Release: 3.1.0
|
closed
| 2024-10-31T15:10:01
| 2024-10-31T15:14:23
| 2024-10-31T15:14:20
|
https://github.com/huggingface/datasets/pull/7270
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7270",
"html_url": "https://github.com/huggingface/datasets/pull/7270",
"diff_url": "https://github.com/huggingface/datasets/pull/7270.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7270.patch",
"merged_at": "2024-10-31T15:14:20"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7270). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,626,873,843
| 7,269
|
Memory leak when streaming
|
open
| 2024-10-31T13:33:52
| 2025-08-05T12:38:44
| null |
https://github.com/huggingface/datasets/issues/7269
| null |
Jourdelune
| false
|
[
"I seem to have encountered the same problem when loading non streaming datasets. load_from_disk. Causing hundreds of GB of memory, but the dataset actually only has 50GB",
"FYI when streaming parquet data, only one row group per worker is loaded in memory at a time.\r\n\r\nBtw for datasets of embeddings you can surely optimize your RAM by reading the data as torch tensors directly instead of the default python lists\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nfrom torch.utils.data import DataLoader\r\n\r\ndataset = load_dataset(\"WaveGenAI/dataset\", streaming=True).with_format(\"torch\")\r\n\r\ndataloader = DataLoader(dataset[\"train\"], num_workers=3)\r\n```",
"Im also, hitting this issue.....\n\n```python\n # This is what's causing the leak:\n batch_datasets = []\n for file_path in batch_files:\n dataset = load_dataset(..., streaming=True)\n shuffled_dataset = dataset.shuffle(seed=42, buffer_size=1000) # 1000-item buffer\n batch_datasets.append(shuffled_dataset) # Buffer persists\n\n interleaved_dataset = interleave_datasets(batch_datasets, seed=42) \n```\n\nAnd, nothing helps\n```python\n del batch_datasets, interleaved_dataset\n gc.collect() # This doesn't work for HuggingFace internal memory structures\n```\nso my guess is that they wrote this in RUST and forgot to clean up!!!\n\nNow, if i remove the interleaving and process files sequentially... like this it still leaks\n```python\n\n # Process files one by one - no batching, no interleaving\n for file_idx, file_path in enumerate(file_paths):\n dataset = load_dataset(\"parquet\", data_files=file_path, split=\"train\", streaming=True)\n shuffled_dataset = dataset.shuffle(seed=42, buffer_size=1000) \n \n for record in shuffled_dataset:\n # Process record immediately\n pass\n \n del dataset, shuffled_dataset\n gc.collect()\n```\n\n - File 1: 42.4% memory\n - File 2: 42.5% memory\n - File 3: 42.5% memory\n - File 4: 48.4% memory (+6%)\n - File 5: 52.7% memory (+4.3%)\n - File 6: 56.7% memory (+4%)\n - File 7: 59.6% memory (+2.9%)\n - File 8: 62.0% memory (+2.4%)\n\nI had to go back to sequential shuffling (NO Interleaving) and clean up like this\n```python\n dataset.cleanup_cache_files() \n del dataset, shuffled_dataset \n gc.collect() \n pa.default_memory_pool().release_unused() \n libc.malloc_trim(0) # when available \n```"
] |
2,626,664,687
| 7,268
|
load_from_disk
|
open
| 2024-10-31T11:51:56
| 2025-07-01T08:42:17
| null |
https://github.com/huggingface/datasets/issues/7268
| null |
ghaith-mq
| false
|
[
"Hello, It's an interesting issue here. I have the same problem, I have a local dataset and I want to push the dataset to the hub but huggingface does a copy of it.\r\n\r\n```py\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"webdataset\", data_files=\"/media/works/data/*.tar\") # copy here\r\ndataset.push_to_hub(\"WaveGenAI/audios2\")\r\n```\r\n\r\nEdit: I can use HfApi for my use case\r\n",
"Is there any update on this issue? I found the same behavior too.\nMy datasets version is `2.13.2`",
"Updating to the newest version of datasets lib resolved the issue. "
] |
2,626,490,029
| 7,267
|
Source installation fails on Macintosh with python 3.10
|
open
| 2024-10-31T10:18:45
| 2024-11-04T22:18:06
| null |
https://github.com/huggingface/datasets/issues/7267
| null |
mayankagarwals
| false
|
[
"I encountered the same problem on M1, a workaround I did was to simply comment out the dependency:\r\n\r\n```python\r\n...\r\n \"zstandard\",\r\n \"polars[timezone]>=0.20.0\",\r\n # \"decord==0.6.0\",\r\n]\r\n```\r\n\r\nThis worked for me as the adjustments I did to the code do not use the dependency, but I do not know if the same holds for you.\r\n\r\nI also do not think it is a good idea to rely on a dependency (I mean decord) that has not been maintained for 2 years, but I saw that even eva-decord hasn't been maintained since last year.\r\n\r\nDid you get it to work with eva-decord?"
] |
2,624,666,087
| 7,266
|
The dataset viewer should be available soon. Please retry later.
|
closed
| 2024-10-30T16:32:00
| 2024-10-31T03:48:11
| 2024-10-31T03:48:10
|
https://github.com/huggingface/datasets/issues/7266
| null |
viiika
| false
|
[
"Waiting is all you need. 10 hours later, it works."
] |
2,624,090,418
| 7,265
|
Disallow video push_to_hub
|
closed
| 2024-10-30T13:21:55
| 2024-10-30T13:36:05
| 2024-10-30T13:36:02
|
https://github.com/huggingface/datasets/pull/7265
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7265",
"html_url": "https://github.com/huggingface/datasets/pull/7265",
"diff_url": "https://github.com/huggingface/datasets/pull/7265.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7265.patch",
"merged_at": "2024-10-30T13:36:02"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7265). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,624,047,640
| 7,264
|
fix docs relative links
|
closed
| 2024-10-30T13:07:34
| 2024-10-30T13:10:13
| 2024-10-30T13:09:02
|
https://github.com/huggingface/datasets/pull/7264
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7264",
"html_url": "https://github.com/huggingface/datasets/pull/7264",
"diff_url": "https://github.com/huggingface/datasets/pull/7264.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7264.patch",
"merged_at": "2024-10-30T13:09:02"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7264). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,621,844,054
| 7,263
|
Small addition to video docs
|
closed
| 2024-10-29T16:58:37
| 2024-10-29T17:01:05
| 2024-10-29T16:59:10
|
https://github.com/huggingface/datasets/pull/7263
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7263",
"html_url": "https://github.com/huggingface/datasets/pull/7263",
"diff_url": "https://github.com/huggingface/datasets/pull/7263.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7263.patch",
"merged_at": "2024-10-29T16:59:10"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7263). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,620,879,059
| 7,262
|
Allow video with disabeld decoding without decord
|
closed
| 2024-10-29T10:54:04
| 2024-10-29T10:56:19
| 2024-10-29T10:55:37
|
https://github.com/huggingface/datasets/pull/7262
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7262",
"html_url": "https://github.com/huggingface/datasets/pull/7262",
"diff_url": "https://github.com/huggingface/datasets/pull/7262.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7262.patch",
"merged_at": "2024-10-29T10:55:37"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7262). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,620,510,840
| 7,261
|
Cannot load the cache when mapping the dataset
|
open
| 2024-10-29T08:29:40
| 2025-03-24T13:27:55
| null |
https://github.com/huggingface/datasets/issues/7261
| null |
zhangn77
| false
|
[
"@zhangn77 Hi ,have you solved this problem? I encountered the same issue during training. Could we discuss it?",
"I also encountered the same problem, why is that?"
] |
2,620,014,285
| 7,260
|
cache can't cleaned or disabled
|
open
| 2024-10-29T03:15:28
| 2024-12-11T09:04:52
| null |
https://github.com/huggingface/datasets/issues/7260
| null |
charliedream1
| false
|
[
"Hey I have a similar problem and found a workaround using [temporary directories](https://docs.python.org/3/library/tempfile.html):\r\n\r\n```python\r\nfrom tempfile import TemporaryDirectory\r\n\r\nwith TemporaryDirectory() as cache_dir:\r\n data = load_dataset('json', data_files=save_local_path, split='train', cache_dir=cache_dir)\r\n```\r\n\r\nBut I do agree that it would be more intuitive if `datasets` supported this directly. Especially `disable_caching` is confusing, since it basically doesn't disable caching."
] |
2,618,909,241
| 7,259
|
Don't embed videos
|
closed
| 2024-10-28T16:25:10
| 2024-10-28T16:27:34
| 2024-10-28T16:26:01
|
https://github.com/huggingface/datasets/pull/7259
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7259",
"html_url": "https://github.com/huggingface/datasets/pull/7259",
"diff_url": "https://github.com/huggingface/datasets/pull/7259.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7259.patch",
"merged_at": "2024-10-28T16:26:01"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7259). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,618,758,399
| 7,258
|
Always set non-null writer batch size
|
closed
| 2024-10-28T15:26:14
| 2024-10-28T15:28:41
| 2024-10-28T15:26:29
|
https://github.com/huggingface/datasets/pull/7258
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7258",
"html_url": "https://github.com/huggingface/datasets/pull/7258",
"diff_url": "https://github.com/huggingface/datasets/pull/7258.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7258.patch",
"merged_at": "2024-10-28T15:26:29"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7258). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,618,602,173
| 7,257
|
fix ci for pyarrow 18
|
closed
| 2024-10-28T14:31:34
| 2024-10-28T14:34:05
| 2024-10-28T14:31:44
|
https://github.com/huggingface/datasets/pull/7257
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7257",
"html_url": "https://github.com/huggingface/datasets/pull/7257",
"diff_url": "https://github.com/huggingface/datasets/pull/7257.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7257.patch",
"merged_at": "2024-10-28T14:31:44"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7257). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,618,580,188
| 7,256
|
Retry all requests timeouts
|
closed
| 2024-10-28T14:23:16
| 2024-10-28T14:56:28
| 2024-10-28T14:56:26
|
https://github.com/huggingface/datasets/pull/7256
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7256",
"html_url": "https://github.com/huggingface/datasets/pull/7256",
"diff_url": "https://github.com/huggingface/datasets/pull/7256.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7256.patch",
"merged_at": "2024-10-28T14:56:26"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7256). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,618,540,355
| 7,255
|
fix decord import
|
closed
| 2024-10-28T14:08:19
| 2024-10-28T14:10:43
| 2024-10-28T14:09:14
|
https://github.com/huggingface/datasets/pull/7255
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7255",
"html_url": "https://github.com/huggingface/datasets/pull/7255",
"diff_url": "https://github.com/huggingface/datasets/pull/7255.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7255.patch",
"merged_at": "2024-10-28T14:09:14"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7255). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,616,174,996
| 7,254
|
mismatch for datatypes when providing `Features` with `Array2D` and user specified `dtype` and using with_format("numpy")
|
open
| 2024-10-26T22:06:27
| 2024-10-26T22:07:37
| null |
https://github.com/huggingface/datasets/issues/7254
| null |
Akhil-CM
| false
|
[
"It seems that https://github.com/huggingface/datasets/issues/5517 is exactly the same issue.\r\n\r\nIt was mentioned there that this would be fixed in version 3.x"
] |
2,615,862,202
| 7,253
|
Unable to upload a large dataset zip either from command line or UI
|
open
| 2024-10-26T13:17:06
| 2024-10-26T13:17:06
| null |
https://github.com/huggingface/datasets/issues/7253
| null |
vakyansh
| false
|
[] |
2,613,795,544
| 7,252
|
Add IterableDataset.shard()
|
closed
| 2024-10-25T11:07:12
| 2025-03-21T03:58:43
| 2024-10-25T15:45:22
|
https://github.com/huggingface/datasets/pull/7252
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7252",
"html_url": "https://github.com/huggingface/datasets/pull/7252",
"diff_url": "https://github.com/huggingface/datasets/pull/7252.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7252.patch",
"merged_at": "2024-10-25T15:45:21"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7252). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"Is there some way to get this to work for pytorch dataloader workers?\r\n\r\neg. start with a single sharded IterableDataset.from_generator(), then reshard before calling map() to do expensive processing over multiple workers"
] |
2,612,097,435
| 7,251
|
Missing video docs
|
closed
| 2024-10-24T16:45:12
| 2024-10-24T16:48:29
| 2024-10-24T16:48:27
|
https://github.com/huggingface/datasets/pull/7251
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7251",
"html_url": "https://github.com/huggingface/datasets/pull/7251",
"diff_url": "https://github.com/huggingface/datasets/pull/7251.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7251.patch",
"merged_at": "2024-10-24T16:48:27"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7251). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,612,041,969
| 7,250
|
Basic XML support (mostly copy pasted from text)
|
closed
| 2024-10-24T16:14:50
| 2024-10-24T16:19:18
| 2024-10-24T16:19:16
|
https://github.com/huggingface/datasets/pull/7250
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7250",
"html_url": "https://github.com/huggingface/datasets/pull/7250",
"diff_url": "https://github.com/huggingface/datasets/pull/7250.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7250.patch",
"merged_at": "2024-10-24T16:19:16"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7250). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,610,136,636
| 7,249
|
How to debugging
|
open
| 2024-10-24T01:03:51
| 2024-10-24T01:03:51
| null |
https://github.com/huggingface/datasets/issues/7249
| null |
ShDdu
| false
|
[] |
2,609,926,089
| 7,248
|
ModuleNotFoundError: No module named 'datasets.tasks'
|
open
| 2024-10-23T21:58:25
| 2024-10-24T17:00:19
| null |
https://github.com/huggingface/datasets/issues/7248
| null |
shoowadoo
| false
|
[
"tasks was removed in v3: #6999 \r\n\r\nI also don't see why TextClassification is imported, since it's not used after. So the fix is simple: delete this line.",
"I opened https://huggingface.co/datasets/knowledgator/events_classification_biotech/discussions/7 to remove the line, hopefully the dataset owner will merge it soon"
] |
2,606,230,029
| 7,247
|
Adding column with dict struction when mapping lead to wrong order
|
open
| 2024-10-22T18:55:11
| 2024-10-22T18:55:23
| null |
https://github.com/huggingface/datasets/issues/7247
| null |
chchch0109
| false
|
[] |
2,605,734,447
| 7,246
|
Set dev version
|
closed
| 2024-10-22T15:04:47
| 2024-10-22T15:07:31
| 2024-10-22T15:04:58
|
https://github.com/huggingface/datasets/pull/7246
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7246",
"html_url": "https://github.com/huggingface/datasets/pull/7246",
"diff_url": "https://github.com/huggingface/datasets/pull/7246.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7246.patch",
"merged_at": "2024-10-22T15:04:58"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7246). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,605,701,235
| 7,245
|
Release: 3.0.2
|
closed
| 2024-10-22T14:53:34
| 2024-10-22T15:01:50
| 2024-10-22T15:01:47
|
https://github.com/huggingface/datasets/pull/7245
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7245",
"html_url": "https://github.com/huggingface/datasets/pull/7245",
"diff_url": "https://github.com/huggingface/datasets/pull/7245.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7245.patch",
"merged_at": "2024-10-22T15:01:47"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7245). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,605,461,515
| 7,244
|
use huggingface_hub offline mode
|
closed
| 2024-10-22T13:27:16
| 2024-10-22T14:10:45
| 2024-10-22T14:10:20
|
https://github.com/huggingface/datasets/pull/7244
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7244",
"html_url": "https://github.com/huggingface/datasets/pull/7244",
"diff_url": "https://github.com/huggingface/datasets/pull/7244.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7244.patch",
"merged_at": "2024-10-22T14:10:20"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7244). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,602,853,172
| 7,243
|
ArrayXD with None as leading dim incompatible with DatasetCardData
|
open
| 2024-10-21T15:08:13
| 2024-10-22T14:18:10
| null |
https://github.com/huggingface/datasets/issues/7243
| null |
alex-hh
| false
|
[
"It looks like `CardData` in `huggingface_hub` removes None values where it shouldn't. Indeed it calls `_remove_none` on the return of `to_dict()`:\r\n\r\n```python\r\n def to_dict(self) -> Dict[str, Any]:\r\n \"\"\"Converts CardData to a dict.\r\n\r\n Returns:\r\n `dict`: CardData represented as a dictionary ready to be dumped to a YAML\r\n block for inclusion in a README.md file.\r\n \"\"\"\r\n\r\n data_dict = copy.deepcopy(self.__dict__)\r\n self._to_dict(data_dict)\r\n return _remove_none(data_dict)\r\n```\r\n\r\nWould it be ok to remove `list()` from being scanned in `_remove_none` ? it could also be a specific behavior to DatasetCardData if necessary @Wauplin ",
"I have actually no idea why none values are removed in model and dataset card data... :see_no_evil:\r\nLooks like `_remove_none` has been introduced at the same time as the entire repocard module (see https://github.com/huggingface/huggingface_hub/pull/940). I would be tempted to remove `_remove_none` entirely actually and only remove \"top-level\" None values (i.e. if something like `pipeline_tag=None` due to a default value in kwargs => we remove it). Hard to tell what could be the side effects but I'm not against trying.\r\n\r\n\r\nHowever, I'm not really in favor in making an exception only for lists. It would mean that tuples, sets and dicts are filtered but not lists, which is pretty inconsistent.",
"let's do it for top level attributes yes",
"I opened https://github.com/huggingface/huggingface_hub/pull/2626 to address it :)",
"thanks !"
] |
2,599,899,156
| 7,241
|
`push_to_hub` overwrite argument
|
closed
| 2024-10-20T03:23:26
| 2024-10-24T17:39:08
| 2024-10-24T17:39:08
|
https://github.com/huggingface/datasets/issues/7241
| null |
ceferisbarov
| false
|
[
"Hi ! Do you mean deleting all the files ? or erasing the repository git history before push_to_hub ?",
"Hi! I meant the latter.",
"I don't think there is a `huggingface_hub` utility to erase the git history, cc @Wauplin maybe ?",
"What is the goal exactly of deleting all the git history without deleting the repo? ",
"You can use [`super_squash_commit`](https://huggingface.co/docs/huggingface_hub/main/en/package_reference/hf_api#huggingface_hub.HfApi.super_squash_history) to squash all the commits into a single one, hence deleting the git history. This is not exactly what you asked for since it squashes the commits for a specific revision (example: \"all commits on main\"). This means that if other branches exists, they are kept the same. Also if some PRs are already opened on the repo, they will become unmergeable since the commits will have diverted.",
"So the solution is:\r\n\r\n```python\r\nfrom huggingface_hub import HfApi\r\nrepo_id = \"username/dataset_name\"\r\nds.push_to_hub(repo_id)\r\nHfApi().super_squash_commit(repo_id)\r\n```\r\n\r\nThis way you erase previous git history to end up with only 1 commit containing your dataset.\r\nStill, I'd be curious why it's important in your case. Is it to save storage space ? or to disallow loading old versions of the data ?",
"Thanks, everyone! I am building a new dataset and playing around with column names, splits, etc. Sometimes I push to the hub to share it with other teammates, I don't want those variations to be part of the repo. Deleting the repo from the website takes a little time, but it also loses repo settings that I have set, since I always set it to public with manually approved requests.\r\n\r\nBTW, I had to write `HfApi().super_squash_history(repo_id, repo_type=\"dataset\")`, but otherwise it works.",
"@ceferisbarov just to let you know, recreating a gated repo + granting access to your teammates is something that you can automate with something like this (not fully tested but should work):\r\n\r\n```py\r\nfrom huggingface_hub import HfApi\r\n\r\napi = HfApi()\r\napi.delete_repo(repo_id, repo_type=\"dataset\", missing_ok=True)\r\napi.create_repo(repo_id, repo_type=\"dataset\", private=False)\r\napi.update_repo_settings(repo_id, repo_type=\"dataset\", gated=\"manual\")\r\nfor user in [\"user1\", \"user2\"] # list of teammates\r\n api.grant_access(repo_id, user, repo_type=\"dataset\")\r\n```\r\n\r\nI think it'd be a better solution than squashing commits (which is more of a hack), typically if you are using the dataset viewer.",
"This is great, @Wauplin. If we can achieve this with HfApi, then we probably don't need to add another parameter to push_to_hub. I am closing the issue."
] |
2,598,980,027
| 7,240
|
Feature Request: Add functionality to pass split types like train, test in DatasetDict.map
|
closed
| 2024-10-19T09:59:12
| 2025-01-06T08:04:08
| 2025-01-06T08:04:08
|
https://github.com/huggingface/datasets/pull/7240
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7240",
"html_url": "https://github.com/huggingface/datasets/pull/7240",
"diff_url": "https://github.com/huggingface/datasets/pull/7240.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7240.patch",
"merged_at": null
}
|
jp1924
| true
|
[] |
2,598,409,993
| 7,238
|
incompatibily issue when using load_dataset with datasets==3.0.1
|
open
| 2024-10-18T21:25:23
| 2024-12-09T09:49:32
| null |
https://github.com/huggingface/datasets/issues/7238
| null |
jupiterMJM
| false
|
[
"Hi! I'm also getting the same issue - have you been able to find a solution to this? ",
"From what I remember, I stayed at the \"downgraded\" version of dataset (2.21.0)"
] |
2,597,358,525
| 7,236
|
[MINOR:TYPO] Update arrow_dataset.py
|
closed
| 2024-10-18T12:10:03
| 2024-10-24T15:06:43
| 2024-10-24T15:06:43
|
https://github.com/huggingface/datasets/pull/7236
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7236",
"html_url": "https://github.com/huggingface/datasets/pull/7236",
"diff_url": "https://github.com/huggingface/datasets/pull/7236.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7236.patch",
"merged_at": "2024-10-24T15:06:43"
}
|
cakiki
| true
|
[] |
2,594,220,624
| 7,234
|
No need for dataset_info
|
closed
| 2024-10-17T09:54:03
| 2024-10-22T12:30:40
| 2024-10-21T16:44:34
|
https://github.com/huggingface/datasets/pull/7234
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7234",
"html_url": "https://github.com/huggingface/datasets/pull/7234",
"diff_url": "https://github.com/huggingface/datasets/pull/7234.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7234.patch",
"merged_at": "2024-10-21T16:44:34"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7234). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"merging this one for now, let me know if you'd like to see additional changes for error handling (I'll take care of them before doing a release)"
] |
2,593,903,113
| 7,233
|
数据集数量问题
|
open
| 2024-10-17T07:41:44
| 2024-10-17T07:41:44
| null |
https://github.com/huggingface/datasets/issues/7233
| null |
want-well
| false
|
[] |
2,593,720,548
| 7,232
|
(Super tiny doc update) Mention to_polars
|
closed
| 2024-10-17T06:08:53
| 2024-10-24T23:11:05
| 2024-10-24T15:06:16
|
https://github.com/huggingface/datasets/pull/7232
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7232",
"html_url": "https://github.com/huggingface/datasets/pull/7232",
"diff_url": "https://github.com/huggingface/datasets/pull/7232.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7232.patch",
"merged_at": "2024-10-24T15:06:16"
}
|
fzyzcjy
| true
|
[
"You are welcome!"
] |
2,592,011,737
| 7,231
|
Fix typo in image dataset docs
|
closed
| 2024-10-16T14:05:46
| 2024-10-16T17:06:21
| 2024-10-16T17:06:19
|
https://github.com/huggingface/datasets/pull/7231
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7231",
"html_url": "https://github.com/huggingface/datasets/pull/7231",
"diff_url": "https://github.com/huggingface/datasets/pull/7231.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7231.patch",
"merged_at": "2024-10-16T17:06:19"
}
|
albertvillanova
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7231). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,589,531,942
| 7,230
|
Video support
|
closed
| 2024-10-15T18:17:29
| 2024-10-24T16:39:51
| 2024-10-24T16:39:50
|
https://github.com/huggingface/datasets/pull/7230
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7230",
"html_url": "https://github.com/huggingface/datasets/pull/7230",
"diff_url": "https://github.com/huggingface/datasets/pull/7230.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7230.patch",
"merged_at": "2024-10-24T16:39:50"
}
|
lhoestq
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7230). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] |
2,588,847,398
| 7,229
|
handle config_name=None in push_to_hub
|
closed
| 2024-10-15T13:48:57
| 2024-10-24T17:51:52
| 2024-10-24T17:51:52
|
https://github.com/huggingface/datasets/pull/7229
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7229",
"html_url": "https://github.com/huggingface/datasets/pull/7229",
"diff_url": "https://github.com/huggingface/datasets/pull/7229.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7229.patch",
"merged_at": null
}
|
alex-hh
| true
|
[
"not sure it's a good idea, we always need a config name so better have the correct default and not support None (which could lead to think it doesn't have a config name, while it does)"
] |
2,587,310,094
| 7,228
|
Composite (multi-column) features
|
open
| 2024-10-14T23:59:19
| 2024-10-15T11:17:15
| null |
https://github.com/huggingface/datasets/issues/7228
| null |
alex-hh
| false
|
[] |
2,587,048,312
| 7,227
|
fast array extraction
|
open
| 2024-10-14T20:51:32
| 2025-01-28T09:39:26
| null |
https://github.com/huggingface/datasets/pull/7227
|
{
"url": "https://api.github.com/repos/huggingface/datasets/pulls/7227",
"html_url": "https://github.com/huggingface/datasets/pull/7227",
"diff_url": "https://github.com/huggingface/datasets/pull/7227.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/7227.patch",
"merged_at": null
}
|
alex-hh
| true
|
[
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7227). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.",
"I've updated the most straightforward failing test cases - lmk if you agree with those.\r\n\r\nMight need some help / pointers on the remaining new failing tests, which seem a little bit more subtle.",
"@lhoestq I've had a go at fixing a few more test cases but getting quite uncertain about the remaining ones (as well as about some of the array writing ones that I tried to fix in my last commit). There are still 27 failures vs 21 on main. I'm not completely sure in some cases what intended behaviour is and my understanding of the flow for typed writing is a bit vague.",
"@lhoestq do you have any thoughts on this? I wasn't able to resolve all the test issues but the basic functionality seemed useful?"
] |
2,586,920,351
| 7,226
|
Add R as a How to use from the Polars (R) Library as an option
|
open
| 2024-10-14T19:56:07
| 2024-10-14T19:57:13
| null |
https://github.com/huggingface/datasets/issues/7226
| null |
ran-codes
| false
|
[] |
2,586,229,216
| 7,225
|
Huggingface GIT returns null as Content-Type instead of application/x-git-receive-pack-result
|
open
| 2024-10-14T14:33:06
| 2024-10-14T14:33:06
| null |
https://github.com/huggingface/datasets/issues/7225
| null |
padmalcom
| false
|
[] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.