id
int64
1.9B
3.25B
title
stringlengths
2
244
state
stringclasses
2 values
body
stringlengths
3
58.6k
created_at
timestamp[s]date
2023-09-15 14:23:33
2025-07-22 09:33:54
updated_at
timestamp[s]date
2023-09-18 16:20:09
2025-07-22 10:44:03
closed_at
timestamp[s]date
2023-09-18 16:20:09
2025-07-19 22:45:08
html_url
stringlengths
49
51
pull_request
dict
number
int64
6.24k
7.7k
is_pull_request
bool
2 classes
comments
listlengths
0
24
2,631,917,431
Accessing audio dataset value throws Format not recognised error
open
### Describe the bug Accessing audio dataset value throws `Format not recognised error` ### Steps to reproduce the bug **code:** ```py from datasets import load_dataset dataset = load_dataset("fawazahmed0/bug-audio") for data in dataset["train"]: print(data) ``` **output:** ```bash (mypy) C:\Users\Nawaz-Server\Documents\ml>python myest.py [C:\vcpkg\buildtrees\mpg123\src\0d8db63f9b-3db975bc05.clean\src\libmpg123\layer3.c:INT123_do_layer3():1801] error: dequantization failed! {'audio': {'path': 'C:\\Users\\Nawaz-Server\\.cache\\huggingface\\hub\\datasets--fawazahmed0--bug-audio\\snapshots\\fab1398431fed1c0a2a7bff0945465bab8b5daef\\data\\Ghamadi\\037135.mp3', 'array': array([ 0.00000000e+00, -2.86519935e-22, -2.56504911e-21, ..., -1.94239747e-02, -2.42924765e-02, -2.99104657e-02]), 'sampling_rate': 22050}, 'reciter': 'Ghamadi', 'transcription': 'الا عجوز ا في الغبرين', 'line': 3923, 'chapter': 37, 'verse': 135, 'text': 'إِلَّا عَجُوزࣰ ا فِي ٱلۡغَٰبِرِينَ'} Traceback (most recent call last): File "C:\Users\Nawaz-Server\Documents\ml\myest.py", line 5, in <module> for data in dataset["train"]: ~~~~~~~^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\datasets\arrow_dataset.py", line 2372, in __iter__ formatted_output = format_table( ^^^^^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\datasets\formatting\formatting.py", line 639, in format_table return formatter(pa_table, query_type=query_type) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\datasets\formatting\formatting.py", line 403, in __call__ return self.format_row(pa_table) ^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\datasets\formatting\formatting.py", line 444, in format_row row = self.python_features_decoder.decode_row(row) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\datasets\formatting\formatting.py", line 222, in decode_row return self.features.decode_example(row) if self.features else row ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\datasets\features\features.py", line 2042, in decode_example column_name: decode_nested_example(feature, value, token_per_repo_id=token_per_repo_id) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\datasets\features\features.py", line 1403, in decode_nested_example return schema.decode_example(obj, token_per_repo_id=token_per_repo_id) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\datasets\features\audio.py", line 184, in decode_example array, sampling_rate = sf.read(f) ^^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\soundfile.py", line 285, in read with SoundFile(file, 'r', samplerate, channels, ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\soundfile.py", line 658, in __init__ self._file = self._open(file, mode_int, closefd) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\Nawaz-Server\.conda\envs\mypy\Lib\site-packages\soundfile.py", line 1216, in _open raise LibsndfileError(err, prefix="Error opening {0!r}: ".format(self.name)) soundfile.LibsndfileError: Error opening <_io.BufferedReader name='C:\\Users\\Nawaz-Server\\.cache\\huggingface\\hub\\datasets--fawazahmed0--bug-audio\\snapshots\\fab1398431fed1c0a2a7bff0945465bab8b5daef\\data\\Ghamadi\\037136.mp3'>: Format not recognised. ``` ### Expected behavior Everything should work fine, as loading the problematic audio file directly with soundfile package works fine **code:** ``` import soundfile as sf print(sf.read('C:\\Users\\Nawaz-Server\\.cache\\huggingface\\hub\\datasets--fawazahmed0--bug-audio\\snapshots\\fab1398431fed1c0a2a7bff0945465bab8b5daef\\data\\Ghamadi\\037136.mp3')) ``` **output:** ```bash (mypy) C:\Users\Nawaz-Server\Documents\ml>python myest.py [C:\vcpkg\buildtrees\mpg123\src\0d8db63f9b-3db975bc05.clean\src\libmpg123\layer3.c:INT123_do_layer3():1801] error: dequantization failed! (array([ 0.00000000e+00, -8.43723821e-22, -2.45370628e-22, ..., -7.71464454e-03, -6.90496899e-03, -8.63333419e-03]), 22050) ``` ### Environment info - `datasets` version: 3.0.2 - Platform: Windows-11-10.0.22621-SP0 - Python version: 3.12.7 - `huggingface_hub` version: 0.26.2 - PyArrow version: 17.0.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.10.0 - soundfile: 0.12.1
2024-11-04T05:59:13
2024-11-09T18:51:52
null
https://github.com/huggingface/datasets/issues/7276
null
7,276
false
[ "Hi ! can you try if this works ?\r\n\r\n```python\r\nimport soundfile as sf\r\n\r\nwith open('C:\\\\Users\\\\Nawaz-Server\\\\.cache\\\\huggingface\\\\hub\\\\datasets--fawazahmed0--bug-audio\\\\snapshots\\\\fab1398431fed1c0a2a7bff0945465bab8b5daef\\\\data\\\\Ghamadi\\\\037136.mp3', 'rb') as f:\r\n print(sf.read(f))\r\n```", "@lhoestq Same error, here is the output:\r\n\r\n```bash\r\n(mypy) C:\\Users\\Nawaz-Server\\Documents\\ml>python myest.py\r\nTraceback (most recent call last):\r\n File \"C:\\Users\\Nawaz-Server\\Documents\\ml\\myest.py\", line 5, in <module>\r\n print(sf.read(f))\r\n ^^^^^^^^^^\r\n File \"C:\\Users\\Nawaz-Server\\.conda\\envs\\mypy\\Lib\\site-packages\\soundfile.py\", line 285, in read\r\n with SoundFile(file, 'r', samplerate, channels,\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\Users\\Nawaz-Server\\.conda\\envs\\mypy\\Lib\\site-packages\\soundfile.py\", line 658, in __init__\r\n self._file = self._open(file, mode_int, closefd)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\Users\\Nawaz-Server\\.conda\\envs\\mypy\\Lib\\site-packages\\soundfile.py\", line 1216, in _open\r\n raise LibsndfileError(err, prefix=\"Error opening {0!r}: \".format(self.name))\r\nsoundfile.LibsndfileError: Error opening <_io.BufferedReader name='C:\\\\Users\\\\Nawaz-Server\\\\.cache\\\\huggingface\\\\hub\\\\datasets--fawazahmed0--bug-audio\\\\snapshots\\\\fab1398431fed1c0a2a7bff0945465bab8b5daef\\\\data\\\\Ghamadi\\\\037136.mp3'>: Format not recognised.\r\n\r\n```", "upstream bug: https://github.com/bastibe/python-soundfile/issues/439" ]
2,631,713,397
load_dataset
open
### Describe the bug I am performing two operations I see on a hugging face tutorial (Fine-tune a language model), and I am defining every aspect inside the mapped functions, also some imports of the library because it doesnt identify anything not defined outside that function where the dataset elements are being mapped: https://colab.research.google.com/github/huggingface/notebooks/blob/main/examples/language_modeling.ipynb#scrollTo=iaAJy5Hu3l_B `- lm_datasets = tokenized_datasets.map( group_texts, batched=True, batch_size=batch_size, num_proc=4, ) - tokenized_datasets = datasets.map(tokenize_function, batched=True, num_proc=4, remove_columns=["text"]) def tokenize_function(examples): model_checkpoint = 'gpt2' from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained(model_checkpoint, use_fast=True) return tokenizer(examples["text"])` ### Steps to reproduce the bug Currently handle all the imports inside the function ### Expected behavior The code must work es expected in the notebook, but currently this is not happening. https://colab.research.google.com/github/huggingface/notebooks/blob/main/examples/language_modeling.ipynb#scrollTo=iaAJy5Hu3l_B ### Environment info print(transformers.__version__) 4.46.1
2024-11-04T03:01:44
2024-11-04T03:01:44
null
https://github.com/huggingface/datasets/issues/7275
null
7,275
false
[]
2,629,882,821
[MINOR:TYPO] Fix typo in exception text
closed
null
2024-11-01T21:15:29
2025-05-21T13:17:20
2025-05-21T13:17:20
https://github.com/huggingface/datasets/pull/7274
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7274", "html_url": "https://github.com/huggingface/datasets/pull/7274", "diff_url": "https://github.com/huggingface/datasets/pull/7274.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7274.patch", "merged_at": null }
7,274
true
[]
2,628,896,492
Raise error for incorrect JSON serialization
closed
Raise error when `lines = False` and `batch_size < Dataset.num_rows` in `Dataset.to_json()`. Issue: #7037 Related PRs: #7039 #7181
2024-11-01T11:54:35
2024-11-18T11:25:01
2024-11-18T11:25:01
https://github.com/huggingface/datasets/pull/7273
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7273", "html_url": "https://github.com/huggingface/datasets/pull/7273", "diff_url": "https://github.com/huggingface/datasets/pull/7273.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7273.patch", "merged_at": "2024-11-18T11:25:01" }
7,273
true
[ "PTAL @lhoestq @albertvillanova ", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7273). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,627,223,390
fix conda release worlflow
closed
null
2024-10-31T15:56:19
2024-10-31T15:58:35
2024-10-31T15:57:29
https://github.com/huggingface/datasets/pull/7272
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7272", "html_url": "https://github.com/huggingface/datasets/pull/7272", "diff_url": "https://github.com/huggingface/datasets/pull/7272.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7272.patch", "merged_at": "2024-10-31T15:57:29" }
7,272
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7272). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,627,135,540
Set dev version
closed
null
2024-10-31T15:22:51
2024-10-31T15:25:27
2024-10-31T15:22:59
https://github.com/huggingface/datasets/pull/7271
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7271", "html_url": "https://github.com/huggingface/datasets/pull/7271", "diff_url": "https://github.com/huggingface/datasets/pull/7271.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7271.patch", "merged_at": "2024-10-31T15:22:59" }
7,271
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7271). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,627,107,016
Release: 3.1.0
closed
null
2024-10-31T15:10:01
2024-10-31T15:14:23
2024-10-31T15:14:20
https://github.com/huggingface/datasets/pull/7270
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7270", "html_url": "https://github.com/huggingface/datasets/pull/7270", "diff_url": "https://github.com/huggingface/datasets/pull/7270.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7270.patch", "merged_at": "2024-10-31T15:14:20" }
7,270
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7270). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,626,873,843
Memory leak when streaming
open
### Describe the bug I try to use a dataset with streaming=True, the issue I have is that the RAM usage becomes higher and higher until it is no longer sustainable. I understand that huggingface store data in ram during the streaming, and more worker in dataloader there are, more a lot of shard will be stored in ram, but the issue I have is that the ram usage is not constant. So after each new shard loaded, the ram usage will be higher and higher. ### Steps to reproduce the bug You can run this code and see you ram usage, after each shard of 255 examples, your ram usage will be extended. ```py from datasets import load_dataset from torch.utils.data import DataLoader dataset = load_dataset("WaveGenAI/dataset", streaming=True) dataloader = DataLoader(dataset["train"], num_workers=3) for i, data in enumerate(dataloader): print(i, end="\r") ``` ### Expected behavior The Ram usage should be always the same (just 3 shards loaded in the ram). ### Environment info - `datasets` version: 3.0.1 - Platform: Linux-6.10.5-arch1-1-x86_64-with-glibc2.40 - Python version: 3.12.4 - `huggingface_hub` version: 0.26.0 - PyArrow version: 17.0.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.6.1
2024-10-31T13:33:52
2024-11-18T11:46:07
null
https://github.com/huggingface/datasets/issues/7269
null
7,269
false
[ "I seem to have encountered the same problem when loading non streaming datasets. load_from_disk. Causing hundreds of GB of memory, but the dataset actually only has 50GB", "FYI when streaming parquet data, only one row group per worker is loaded in memory at a time.\r\n\r\nBtw for datasets of embeddings you can surely optimize your RAM by reading the data as torch tensors directly instead of the default python lists\r\n\r\n```python\r\nfrom datasets import load_dataset\r\nfrom torch.utils.data import DataLoader\r\n\r\ndataset = load_dataset(\"WaveGenAI/dataset\", streaming=True).with_format(\"torch\")\r\n\r\ndataloader = DataLoader(dataset[\"train\"], num_workers=3)\r\n```" ]
2,626,664,687
load_from_disk
open
### Describe the bug I have data saved with save_to_disk. The data is big (700Gb). When I try loading it, the only option is load_from_disk, and this function copies the data to a tmp directory, causing me to run out of disk space. Is there an alternative solution to that? ### Steps to reproduce the bug when trying to load data using load_From_disk after being saved using save_to_disk ### Expected behavior run out of disk space ### Environment info lateest version
2024-10-31T11:51:56
2025-07-01T08:42:17
null
https://github.com/huggingface/datasets/issues/7268
null
7,268
false
[ "Hello, It's an interesting issue here. I have the same problem, I have a local dataset and I want to push the dataset to the hub but huggingface does a copy of it.\r\n\r\n```py\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset(\"webdataset\", data_files=\"/media/works/data/*.tar\") # copy here\r\ndataset.push_to_hub(\"WaveGenAI/audios2\")\r\n```\r\n\r\nEdit: I can use HfApi for my use case\r\n", "Is there any update on this issue? I found the same behavior too.\nMy datasets version is `2.13.2`", "Updating to the newest version of datasets lib resolved the issue. " ]
2,626,490,029
Source installation fails on Macintosh with python 3.10
open
### Describe the bug Hi, Decord is a dev dependency not maintained since couple years. It does not have an ARM package available rendering it uninstallable on non-intel based macs Suggestion is to move to eva-decord (https://github.com/georgia-tech-db/eva-decord) which doesnt have this problem. Happy to raise a PR ### Steps to reproduce the bug Source installation as mentioned in contributinog.md ### Expected behavior Installation without decord failing to be installed. ### Environment info python=3.10, M3 Mac
2024-10-31T10:18:45
2024-11-04T22:18:06
null
https://github.com/huggingface/datasets/issues/7267
null
7,267
false
[ "I encountered the same problem on M1, a workaround I did was to simply comment out the dependency:\r\n\r\n```python\r\n...\r\n \"zstandard\",\r\n \"polars[timezone]>=0.20.0\",\r\n # \"decord==0.6.0\",\r\n]\r\n```\r\n\r\nThis worked for me as the adjustments I did to the code do not use the dependency, but I do not know if the same holds for you.\r\n\r\nI also do not think it is a good idea to rely on a dependency (I mean decord) that has not been maintained for 2 years, but I saw that even eva-decord hasn't been maintained since last year.\r\n\r\nDid you get it to work with eva-decord?" ]
2,624,666,087
The dataset viewer should be available soon. Please retry later.
closed
### Describe the bug After waiting for 2 hours, it still presents ``The dataset viewer should be available soon. Please retry later.'' ### Steps to reproduce the bug dataset link: https://huggingface.co/datasets/BryanW/HI_EDIT ### Expected behavior Present the dataset viewer. ### Environment info NA
2024-10-30T16:32:00
2024-10-31T03:48:11
2024-10-31T03:48:10
https://github.com/huggingface/datasets/issues/7266
null
7,266
false
[ "Waiting is all you need. 10 hours later, it works." ]
2,624,090,418
Disallow video push_to_hub
closed
null
2024-10-30T13:21:55
2024-10-30T13:36:05
2024-10-30T13:36:02
https://github.com/huggingface/datasets/pull/7265
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7265", "html_url": "https://github.com/huggingface/datasets/pull/7265", "diff_url": "https://github.com/huggingface/datasets/pull/7265.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7265.patch", "merged_at": "2024-10-30T13:36:02" }
7,265
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7265). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,624,047,640
fix docs relative links
closed
null
2024-10-30T13:07:34
2024-10-30T13:10:13
2024-10-30T13:09:02
https://github.com/huggingface/datasets/pull/7264
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7264", "html_url": "https://github.com/huggingface/datasets/pull/7264", "diff_url": "https://github.com/huggingface/datasets/pull/7264.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7264.patch", "merged_at": "2024-10-30T13:09:02" }
7,264
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7264). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,621,844,054
Small addition to video docs
closed
null
2024-10-29T16:58:37
2024-10-29T17:01:05
2024-10-29T16:59:10
https://github.com/huggingface/datasets/pull/7263
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7263", "html_url": "https://github.com/huggingface/datasets/pull/7263", "diff_url": "https://github.com/huggingface/datasets/pull/7263.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7263.patch", "merged_at": "2024-10-29T16:59:10" }
7,263
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7263). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,620,879,059
Allow video with disabeld decoding without decord
closed
for the viewer, this way it can use Video(decode=False) and doesn't need decord (which causes segfaults)
2024-10-29T10:54:04
2024-10-29T10:56:19
2024-10-29T10:55:37
https://github.com/huggingface/datasets/pull/7262
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7262", "html_url": "https://github.com/huggingface/datasets/pull/7262", "diff_url": "https://github.com/huggingface/datasets/pull/7262.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7262.patch", "merged_at": "2024-10-29T10:55:37" }
7,262
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7262). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,620,510,840
Cannot load the cache when mapping the dataset
open
### Describe the bug I'm training the flux controlnet. The train_dataset.map() takes long time to finish. However, when I killed one training process and want to restart a new training with the same dataset. I can't reuse the mapped result even I defined the cache dir for the dataset. with accelerator.main_process_first(): from datasets.fingerprint import Hasher # fingerprint used by the cache for the other processes to load the result # details: https://github.com/huggingface/diffusers/pull/4038#discussion_r1266078401 new_fingerprint = Hasher.hash(args) train_dataset = train_dataset.map( compute_embeddings_fn, batched=True, new_fingerprint=new_fingerprint, batch_size=10, ) ### Steps to reproduce the bug train flux controlnet and start again ### Expected behavior will not map again ### Environment info latest diffusers
2024-10-29T08:29:40
2025-03-24T13:27:55
null
https://github.com/huggingface/datasets/issues/7261
null
7,261
false
[ "@zhangn77 Hi ,have you solved this problem? I encountered the same issue during training. Could we discuss it?", "I also encountered the same problem, why is that?" ]
2,620,014,285
cache can't cleaned or disabled
open
### Describe the bug I tried following ways, the cache can't be disabled. I got 2T data, but I also got more than 2T cache file. I got pressure on storage. I need to diable cache or cleaned immediately after processed. Following ways are all not working, please give some help! ```python from datasets import disable_caching from transformers import AutoTokenizer disable_caching() tokenizer = AutoTokenizer.from_pretrained(args.tokenizer_path) def tokenization_fn(examples): column_name = 'text' if 'text' in examples else 'data' tokenized_inputs = tokenizer( examples[column_name], return_special_tokens_mask=True, truncation=False, max_length=tokenizer.model_max_length ) return tokenized_inputs data = load_dataset('json', data_files=save_local_path, split='train', cache_dir=None) data.cleanup_cache_files() updated_dataset = data.map(tokenization_fn, load_from_cache_file=False) updated_dataset .cleanup_cache_files() ``` ### Expected behavior no cache file generated ### Environment info Ubuntu 20.04.6 LTS datasets 3.0.2
2024-10-29T03:15:28
2024-12-11T09:04:52
null
https://github.com/huggingface/datasets/issues/7260
null
7,260
false
[ "Hey I have a similar problem and found a workaround using [temporary directories](https://docs.python.org/3/library/tempfile.html):\r\n\r\n```python\r\nfrom tempfile import TemporaryDirectory\r\n\r\nwith TemporaryDirectory() as cache_dir:\r\n data = load_dataset('json', data_files=save_local_path, split='train', cache_dir=cache_dir)\r\n```\r\n\r\nBut I do agree that it would be more intuitive if `datasets` supported this directly. Especially `disable_caching` is confusing, since it basically doesn't disable caching." ]
2,618,909,241
Don't embed videos
closed
don't include video bytes when running download_and_prepare(format="parquet") this also affects push_to_hub which will just upload the local paths of the videos though
2024-10-28T16:25:10
2024-10-28T16:27:34
2024-10-28T16:26:01
https://github.com/huggingface/datasets/pull/7259
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7259", "html_url": "https://github.com/huggingface/datasets/pull/7259", "diff_url": "https://github.com/huggingface/datasets/pull/7259.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7259.patch", "merged_at": "2024-10-28T16:26:01" }
7,259
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7259). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,618,758,399
Always set non-null writer batch size
closed
bug introduced in #7230, it was preventing the Viewer limit writes to work
2024-10-28T15:26:14
2024-10-28T15:28:41
2024-10-28T15:26:29
https://github.com/huggingface/datasets/pull/7258
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7258", "html_url": "https://github.com/huggingface/datasets/pull/7258", "diff_url": "https://github.com/huggingface/datasets/pull/7258.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7258.patch", "merged_at": "2024-10-28T15:26:29" }
7,258
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7258). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,618,602,173
fix ci for pyarrow 18
closed
null
2024-10-28T14:31:34
2024-10-28T14:34:05
2024-10-28T14:31:44
https://github.com/huggingface/datasets/pull/7257
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7257", "html_url": "https://github.com/huggingface/datasets/pull/7257", "diff_url": "https://github.com/huggingface/datasets/pull/7257.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7257.patch", "merged_at": "2024-10-28T14:31:44" }
7,257
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7257). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,618,580,188
Retry all requests timeouts
closed
as reported in https://github.com/huggingface/datasets/issues/6843
2024-10-28T14:23:16
2024-10-28T14:56:28
2024-10-28T14:56:26
https://github.com/huggingface/datasets/pull/7256
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7256", "html_url": "https://github.com/huggingface/datasets/pull/7256", "diff_url": "https://github.com/huggingface/datasets/pull/7256.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7256.patch", "merged_at": "2024-10-28T14:56:26" }
7,256
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7256). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,618,540,355
fix decord import
closed
delay the import until Video() is instantiated + also import duckdb first (otherwise importing duckdb later causes a segfault)
2024-10-28T14:08:19
2024-10-28T14:10:43
2024-10-28T14:09:14
https://github.com/huggingface/datasets/pull/7255
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7255", "html_url": "https://github.com/huggingface/datasets/pull/7255", "diff_url": "https://github.com/huggingface/datasets/pull/7255.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7255.patch", "merged_at": "2024-10-28T14:09:14" }
7,255
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7255). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,616,174,996
mismatch for datatypes when providing `Features` with `Array2D` and user specified `dtype` and using with_format("numpy")
open
### Describe the bug If the user provides a `Features` type value to `datasets.Dataset` with members having `Array2D` with a value for `dtype`, it is not respected during `with_format("numpy")` which should return a `np.array` with `dtype` that the user provided for `Array2D`. It seems for floats, it will be set to `float32` and for ints it will be set to `int64` ### Steps to reproduce the bug ```python import numpy as np import datasets from datasets import Dataset, Features, Array2D print(f"datasets version: {datasets.__version__}") data_info = { "arr_float" : "float64", "arr_int" : "int32" } sample = {key : [np.zeros([4, 5], dtype=dtype)] for key, dtype in data_info.items()} features = {key : Array2D(shape=(None, 5), dtype=dtype) for key, dtype in data_info.items()} features = Features(features) dataset = Dataset.from_dict(sample, features=features) ds = dataset.with_format("numpy") for key in features: print(f"{key} feature dtype: ", ds.features[key].dtype) print(f"{key} dtype:", ds[key].dtype) ``` Output: ```bash datasets version: 3.0.2 arr_float feature dtype: float64 arr_float dtype: float32 arr_int feature dtype: int32 arr_int dtype: int64 ``` ### Expected behavior It should return a `np.array` with `dtype` that the user provided for the corresponding member in the `Features` type value ### Environment info - `datasets` version: 3.0.2 - Platform: Linux-6.11.5-arch1-1-x86_64-with-glibc2.40 - Python version: 3.12.7 - `huggingface_hub` version: 0.26.1 - PyArrow version: 16.1.0 - Pandas version: 2.2.2 - `fsspec` version: 2024.5.0
2024-10-26T22:06:27
2024-10-26T22:07:37
null
https://github.com/huggingface/datasets/issues/7254
null
7,254
false
[ "It seems that https://github.com/huggingface/datasets/issues/5517 is exactly the same issue.\r\n\r\nIt was mentioned there that this would be fixed in version 3.x" ]
2,615,862,202
Unable to upload a large dataset zip either from command line or UI
open
### Describe the bug Unable to upload a large dataset zip from command line or UI. UI simply says error. I am trying to a upload a tar.gz file of 17GB. <img width="550" alt="image" src="https://github.com/user-attachments/assets/f9d29024-06c8-49c4-a109-0492cff79d34"> <img width="755" alt="image" src="https://github.com/user-attachments/assets/a8d4acda-7f02-4279-9c2d-b2e0282b4faa"> ### Steps to reproduce the bug Upload a large file ### Expected behavior The file should upload without any issue. ### Environment info None
2024-10-26T13:17:06
2024-10-26T13:17:06
null
https://github.com/huggingface/datasets/issues/7253
null
7,253
false
[]
2,613,795,544
Add IterableDataset.shard()
closed
Will be useful to distribute a dataset across workers (other than pytorch) like spark I also renamed `.n_shards` -> `.num_shards` for consistency and kept the old name for backward compatibility. And a few changes in internal functions for consistency as well (rank, world_size -> num_shards, index) Breaking change: the new default for `contiguous` in `Dataset.shard()` is `True`, but imo not a big deal since I couldn't find any usage of `contiguous=False` internally (we always do contiguous=True for map-style datasets since its more optimized) or in the wild
2024-10-25T11:07:12
2025-03-21T03:58:43
2024-10-25T15:45:22
https://github.com/huggingface/datasets/pull/7252
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7252", "html_url": "https://github.com/huggingface/datasets/pull/7252", "diff_url": "https://github.com/huggingface/datasets/pull/7252.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7252.patch", "merged_at": "2024-10-25T15:45:21" }
7,252
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7252). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "Is there some way to get this to work for pytorch dataloader workers?\r\n\r\neg. start with a single sharded IterableDataset.from_generator(), then reshard before calling map() to do expensive processing over multiple workers" ]
2,612,097,435
Missing video docs
closed
null
2024-10-24T16:45:12
2024-10-24T16:48:29
2024-10-24T16:48:27
https://github.com/huggingface/datasets/pull/7251
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7251", "html_url": "https://github.com/huggingface/datasets/pull/7251", "diff_url": "https://github.com/huggingface/datasets/pull/7251.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7251.patch", "merged_at": "2024-10-24T16:48:27" }
7,251
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7251). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,612,041,969
Basic XML support (mostly copy pasted from text)
closed
enable the viewer for datasets like https://huggingface.co/datasets/FrancophonIA/e-calm (there will be more and more apparently)
2024-10-24T16:14:50
2024-10-24T16:19:18
2024-10-24T16:19:16
https://github.com/huggingface/datasets/pull/7250
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7250", "html_url": "https://github.com/huggingface/datasets/pull/7250", "diff_url": "https://github.com/huggingface/datasets/pull/7250.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7250.patch", "merged_at": "2024-10-24T16:19:16" }
7,250
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7250). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,610,136,636
How to debugging
open
### Describe the bug I wanted to use my own script to handle the processing, and followed the tutorial documentation by rewriting the MyDatasetConfig and MyDatasetBuilder (which contains the _info,_split_generators and _generate_examples methods) classes. Testing with simple data was able to output the results of the processing, but when I wished to do more complex processing, I found that I was unable to debug (even the simple samples were inaccessible). There are no errors reported, and I am able to print the _info,_split_generators and _generate_examples messages, but I am unable to access the breakpoints. ### Steps to reproduce the bug # my_dataset.py import json import datasets class MyDatasetConfig(datasets.BuilderConfig): def __init__(self, **kwargs): super(MyDatasetConfig, self).__init__(**kwargs) class MyDataset(datasets.GeneratorBasedBuilder): VERSION = datasets.Version("1.0.0") BUILDER_CONFIGS = [ MyDatasetConfig( name="default", version=VERSION, description="myDATASET" ), ] def _info(self): print("info") # breakpoints return datasets.DatasetInfo( description="myDATASET", features=datasets.Features( { "id": datasets.Value("int32"), "text": datasets.Value("string"), "label": datasets.ClassLabel(names=["negative", "positive"]), } ), supervised_keys=("text", "label"), ) def _split_generators(self, dl_manager): print("generate") # breakpoints data_file = "data.json" return [ datasets.SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={"filepath": data_file} ), ] def _generate_examples(self, filepath): print("example") # breakpoints with open(filepath, encoding="utf-8") as f: data = json.load(f) for idx, sample in enumerate(data): yield idx, { "id": sample["id"], "text": sample["text"], "label": sample["label"], } #main.py import os os.environ["TRANSFORMERS_NO_MULTIPROCESSING"] = "1" from datasets import load_dataset dataset = load_dataset("my_dataset.py", split="train", cache_dir=None) print(dataset[:5]) ### Expected behavior Pause at breakpoints while running debugging ### Environment info pycharm
2024-10-24T01:03:51
2024-10-24T01:03:51
null
https://github.com/huggingface/datasets/issues/7249
null
7,249
false
[]
2,609,926,089
ModuleNotFoundError: No module named 'datasets.tasks'
open
### Describe the bug --------------------------------------------------------------------------- ModuleNotFoundError Traceback (most recent call last) [<ipython-input-9-13b5f31bd391>](https://bcb6shpazyu-496ff2e9c6d22116-0-colab.googleusercontent.com/outputframe.html?vrz=colab_20241022-060119_RC00_688494744#) in <cell line: 1>() ----> 1 dataset = load_dataset('knowledgator/events_classification_biotech') 11 frames [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://bcb6shpazyu-496ff2e9c6d22116-0-colab.googleusercontent.com/outputframe.html?vrz=colab_20241022-060119_RC00_688494744#) in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, keep_in_memory, save_infos, revision, token, streaming, num_proc, storage_options, trust_remote_code, **config_kwargs) 2130 2131 # Create a dataset builder -> 2132 builder_instance = load_dataset_builder( 2133 path=path, 2134 name=name, [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://bcb6shpazyu-496ff2e9c6d22116-0-colab.googleusercontent.com/outputframe.html?vrz=colab_20241022-060119_RC00_688494744#) in load_dataset_builder(path, name, data_dir, data_files, cache_dir, features, download_config, download_mode, revision, token, storage_options, trust_remote_code, _require_default_config_name, **config_kwargs) 1886 raise ValueError(error_msg) 1887 -> 1888 builder_cls = get_dataset_builder_class(dataset_module, dataset_name=dataset_name) 1889 # Instantiate the dataset builder 1890 builder_instance: DatasetBuilder = builder_cls( [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://bcb6shpazyu-496ff2e9c6d22116-0-colab.googleusercontent.com/outputframe.html?vrz=colab_20241022-060119_RC00_688494744#) in get_dataset_builder_class(dataset_module, dataset_name) 246 dataset_module.importable_file_path 247 ) if dataset_module.importable_file_path else nullcontext(): --> 248 builder_cls = import_main_class(dataset_module.module_path) 249 if dataset_module.builder_configs_parameters.builder_configs: 250 dataset_name = dataset_name or dataset_module.builder_kwargs.get("dataset_name") [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://bcb6shpazyu-496ff2e9c6d22116-0-colab.googleusercontent.com/outputframe.html?vrz=colab_20241022-060119_RC00_688494744#) in import_main_class(module_path) 167 def import_main_class(module_path) -> Optional[Type[DatasetBuilder]]: 168 """Import a module at module_path and return its main class: a DatasetBuilder""" --> 169 module = importlib.import_module(module_path) 170 # Find the main class in our imported module 171 module_main_cls = None [/usr/lib/python3.10/importlib/__init__.py](https://bcb6shpazyu-496ff2e9c6d22116-0-colab.googleusercontent.com/outputframe.html?vrz=colab_20241022-060119_RC00_688494744#) in import_module(name, package) 124 break 125 level += 1 --> 126 return _bootstrap._gcd_import(name[level:], package, level) 127 128 /usr/lib/python3.10/importlib/_bootstrap.py in _gcd_import(name, package, level) /usr/lib/python3.10/importlib/_bootstrap.py in _find_and_load(name, import_) /usr/lib/python3.10/importlib/_bootstrap.py in _find_and_load_unlocked(name, import_) /usr/lib/python3.10/importlib/_bootstrap.py in _load_unlocked(spec) /usr/lib/python3.10/importlib/_bootstrap_external.py in exec_module(self, module) /usr/lib/python3.10/importlib/_bootstrap.py in _call_with_frames_removed(f, *args, **kwds) [~/.cache/huggingface/modules/datasets_modules/datasets/knowledgator--events_classification_biotech/9c8086d498c3104de3a3c5b6640837e18ccd829dcaca49f1cdffe3eb5c4a6361/events_classification_biotech.py](https://bcb6shpazyu-496ff2e9c6d22116-0-colab.googleusercontent.com/outputframe.html?vrz=colab_20241022-060119_RC00_688494744#) in <module> 1 import datasets 2 from datasets import load_dataset ----> 3 from datasets.tasks import TextClassification 4 5 DESCRIPTION = """ ModuleNotFoundError: No module named 'datasets.tasks' --------------------------------------------------------------------------- NOTE: If your import is failing due to a missing package, you can manually install dependencies using either !pip or !apt. To view examples of installing some common dependencies, click the "Open Examples" button below. --------------------------------------------------------------------------- ### Steps to reproduce the bug !pip install datasets from datasets import load_dataset dataset = load_dataset('knowledgator/events_classification_biotech') ### Expected behavior no ModuleNotFoundError ### Environment info google colab
2024-10-23T21:58:25
2024-10-24T17:00:19
null
https://github.com/huggingface/datasets/issues/7248
null
7,248
false
[ "tasks was removed in v3: #6999 \r\n\r\nI also don't see why TextClassification is imported, since it's not used after. So the fix is simple: delete this line.", "I opened https://huggingface.co/datasets/knowledgator/events_classification_biotech/discussions/7 to remove the line, hopefully the dataset owner will merge it soon" ]
2,606,230,029
Adding column with dict struction when mapping lead to wrong order
open
### Describe the bug in `map()` function, I want to add a new column with a dict structure. ``` def map_fn(example): example['text'] = {'user': ..., 'assistant': ...} return example ``` However this leads to a wrong order `{'assistant':..., 'user':...}` in the dataset. Thus I can't concatenate two datasets due to the different feature structures. [Here](https://colab.research.google.com/drive/1zeaWq9Ith4DKWP_EiBNyLfc8S8I68LyY?usp=sharing) is a minimal reproducible example This seems an issue in low level pyarrow library instead of datasets, however, I think datasets should allow concatenate two datasets actually in the same structure. ### Steps to reproduce the bug [Here](https://colab.research.google.com/drive/1zeaWq9Ith4DKWP_EiBNyLfc8S8I68LyY?usp=sharing) is a minimal reproducible example ### Expected behavior two datasets could be concatenated. ### Environment info N/A
2024-10-22T18:55:11
2024-10-22T18:55:23
null
https://github.com/huggingface/datasets/issues/7247
null
7,247
false
[]
2,605,734,447
Set dev version
closed
null
2024-10-22T15:04:47
2024-10-22T15:07:31
2024-10-22T15:04:58
https://github.com/huggingface/datasets/pull/7246
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7246", "html_url": "https://github.com/huggingface/datasets/pull/7246", "diff_url": "https://github.com/huggingface/datasets/pull/7246.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7246.patch", "merged_at": "2024-10-22T15:04:58" }
7,246
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7246). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,605,701,235
Release: 3.0.2
closed
null
2024-10-22T14:53:34
2024-10-22T15:01:50
2024-10-22T15:01:47
https://github.com/huggingface/datasets/pull/7245
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7245", "html_url": "https://github.com/huggingface/datasets/pull/7245", "diff_url": "https://github.com/huggingface/datasets/pull/7245.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7245.patch", "merged_at": "2024-10-22T15:01:47" }
7,245
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7245). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,605,461,515
use huggingface_hub offline mode
closed
and better handling of LocalEntryNotfoundError cc @Wauplin follow up to #7234
2024-10-22T13:27:16
2024-10-22T14:10:45
2024-10-22T14:10:20
https://github.com/huggingface/datasets/pull/7244
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7244", "html_url": "https://github.com/huggingface/datasets/pull/7244", "diff_url": "https://github.com/huggingface/datasets/pull/7244.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7244.patch", "merged_at": "2024-10-22T14:10:20" }
7,244
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7244). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,602,853,172
ArrayXD with None as leading dim incompatible with DatasetCardData
open
### Describe the bug Creating a dataset with ArrayXD features leads to errors when downloading from hub due to DatasetCardData removing the Nones @lhoestq ### Steps to reproduce the bug ```python import numpy as np from datasets import Array2D, Dataset, Features, load_dataset def examples_generator(): for i in range(4): yield { "array_1d": np.zeros((10,1), dtype="uint16"), "array_2d": np.zeros((10, 1), dtype="uint16"), } features = Features(array_1d=Array2D((None,1), "uint16"), array_2d=Array2D((None, 1), "uint16")) dataset = Dataset.from_generator(examples_generator, features=features) dataset.push_to_hub("alex-hh/test_array_1d2d") ds = load_dataset("alex-hh/test_array_1d2d") ``` Source of error appears to be DatasetCardData.to_dict invoking DatasetCardData._remove_none ```python from huggingface_hub import DatasetCardData from datasets.info import DatasetInfosDict dataset_card_data = DatasetCardData() DatasetInfosDict({"default": dataset.info.copy()}).to_dataset_card_data(dataset_card_data) print(dataset_card_data.to_dict()) # removes Nones in shape ``` ### Expected behavior Should be possible to load datasets saved with shape None in leading dimension ### Environment info 3.0.2 and latest huggingface_hub
2024-10-21T15:08:13
2024-10-22T14:18:10
null
https://github.com/huggingface/datasets/issues/7243
null
7,243
false
[ "It looks like `CardData` in `huggingface_hub` removes None values where it shouldn't. Indeed it calls `_remove_none` on the return of `to_dict()`:\r\n\r\n```python\r\n def to_dict(self) -> Dict[str, Any]:\r\n \"\"\"Converts CardData to a dict.\r\n\r\n Returns:\r\n `dict`: CardData represented as a dictionary ready to be dumped to a YAML\r\n block for inclusion in a README.md file.\r\n \"\"\"\r\n\r\n data_dict = copy.deepcopy(self.__dict__)\r\n self._to_dict(data_dict)\r\n return _remove_none(data_dict)\r\n```\r\n\r\nWould it be ok to remove `list()` from being scanned in `_remove_none` ? it could also be a specific behavior to DatasetCardData if necessary @Wauplin ", "I have actually no idea why none values are removed in model and dataset card data... :see_no_evil:\r\nLooks like `_remove_none` has been introduced at the same time as the entire repocard module (see https://github.com/huggingface/huggingface_hub/pull/940). I would be tempted to remove `_remove_none` entirely actually and only remove \"top-level\" None values (i.e. if something like `pipeline_tag=None` due to a default value in kwargs => we remove it). Hard to tell what could be the side effects but I'm not against trying.\r\n\r\n\r\nHowever, I'm not really in favor in making an exception only for lists. It would mean that tuples, sets and dicts are filtered but not lists, which is pretty inconsistent.", "let's do it for top level attributes yes", "I opened https://github.com/huggingface/huggingface_hub/pull/2626 to address it :)", "thanks !" ]
2,599,899,156
`push_to_hub` overwrite argument
closed
### Feature request Add an `overwrite` argument to the `push_to_hub` method. ### Motivation I want to overwrite a repo without deleting it on Hugging Face. Is this possible? I couldn't find anything in the documentation or tutorials. ### Your contribution I can create a PR.
2024-10-20T03:23:26
2024-10-24T17:39:08
2024-10-24T17:39:08
https://github.com/huggingface/datasets/issues/7241
null
7,241
false
[ "Hi ! Do you mean deleting all the files ? or erasing the repository git history before push_to_hub ?", "Hi! I meant the latter.", "I don't think there is a `huggingface_hub` utility to erase the git history, cc @Wauplin maybe ?", "What is the goal exactly of deleting all the git history without deleting the repo? ", "You can use [`super_squash_commit`](https://huggingface.co/docs/huggingface_hub/main/en/package_reference/hf_api#huggingface_hub.HfApi.super_squash_history) to squash all the commits into a single one, hence deleting the git history. This is not exactly what you asked for since it squashes the commits for a specific revision (example: \"all commits on main\"). This means that if other branches exists, they are kept the same. Also if some PRs are already opened on the repo, they will become unmergeable since the commits will have diverted.", "So the solution is:\r\n\r\n```python\r\nfrom huggingface_hub import HfApi\r\nrepo_id = \"username/dataset_name\"\r\nds.push_to_hub(repo_id)\r\nHfApi().super_squash_commit(repo_id)\r\n```\r\n\r\nThis way you erase previous git history to end up with only 1 commit containing your dataset.\r\nStill, I'd be curious why it's important in your case. Is it to save storage space ? or to disallow loading old versions of the data ?", "Thanks, everyone! I am building a new dataset and playing around with column names, splits, etc. Sometimes I push to the hub to share it with other teammates, I don't want those variations to be part of the repo. Deleting the repo from the website takes a little time, but it also loses repo settings that I have set, since I always set it to public with manually approved requests.\r\n\r\nBTW, I had to write `HfApi().super_squash_history(repo_id, repo_type=\"dataset\")`, but otherwise it works.", "@ceferisbarov just to let you know, recreating a gated repo + granting access to your teammates is something that you can automate with something like this (not fully tested but should work):\r\n\r\n```py\r\nfrom huggingface_hub import HfApi\r\n\r\napi = HfApi()\r\napi.delete_repo(repo_id, repo_type=\"dataset\", missing_ok=True)\r\napi.create_repo(repo_id, repo_type=\"dataset\", private=False)\r\napi.update_repo_settings(repo_id, repo_type=\"dataset\", gated=\"manual\")\r\nfor user in [\"user1\", \"user2\"] # list of teammates\r\n api.grant_access(repo_id, user, repo_type=\"dataset\")\r\n```\r\n\r\nI think it'd be a better solution than squashing commits (which is more of a hack), typically if you are using the dataset viewer.", "This is great, @Wauplin. If we can achieve this with HfApi, then we probably don't need to add another parameter to push_to_hub. I am closing the issue." ]
2,598,980,027
Feature Request: Add functionality to pass split types like train, test in DatasetDict.map
closed
Hello datasets! We often encounter situations where we need to preprocess data differently depending on split types such as train, valid, and test. However, while DatasetDict.map has features to pass rank or index, there's no functionality to pass split types. Therefore, I propose adding a 'with_splits' parameter to DatasetDict, which would allow passing the split type through fn_kwargs.
2024-10-19T09:59:12
2025-01-06T08:04:08
2025-01-06T08:04:08
https://github.com/huggingface/datasets/pull/7240
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7240", "html_url": "https://github.com/huggingface/datasets/pull/7240", "diff_url": "https://github.com/huggingface/datasets/pull/7240.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7240.patch", "merged_at": null }
7,240
true
[]
2,598,409,993
incompatibily issue when using load_dataset with datasets==3.0.1
open
### Describe the bug There is a bug when using load_dataset with dataset version at 3.0.1 . Please see below in the "steps to reproduce the bug". To resolve the bug, I had to downgrade to version 2.21.0 OS: Ubuntu 24 (AWS instance) Python: same bug under 3.12 and 3.10 The error I had was: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/ubuntu/miniconda3/envs/maxence_env/lib/python3.10/site-packages/datasets/load.py", line 2096, in load_dataset builder_instance.download_and_prepare( File "/home/ubuntu/miniconda3/envs/maxence_env/lib/python3.10/site-packages/datasets/builder.py", line 924, in download_and_prepare self._download_and_prepare( File "/home/ubuntu/miniconda3/envs/maxence_env/lib/python3.10/site-packages/datasets/builder.py", line 1647, in _download_and_prepare super()._download_and_prepare( File "/home/ubuntu/miniconda3/envs/maxence_env/lib/python3.10/site-packages/datasets/builder.py", line 977, in _download_and_prepare split_generators = self._split_generators(dl_manager, **split_generators_kwargs) File "/home/ubuntu/.cache/huggingface/modules/datasets_modules/datasets/mozilla-foundation--common_voice_6_0/cb17afd34f5799f97e8f48398748f83006335b702bd785f9880797838d541b81/common_voice_6_0.py", line 159, in _split_generators archive_path = dl_manager.download(self._get_bundle_url(self.config.name, bundle_url_template)) File "/home/ubuntu/miniconda3/envs/maxence_env/lib/python3.10/site-packages/datasets/download/download_manager.py", line 150, in download download_config = self.download_config.copy() File "/home/ubuntu/miniconda3/envs/maxence_env/lib/python3.10/site-packages/datasets/download/download_config.py", line 73, in copy return self.__class__(**{k: copy.deepcopy(v) for k, v in self.__dict__.items()}) TypeError: DownloadConfig.__init__() got an unexpected keyword argument 'ignore_url_params' ### Steps to reproduce the bug 1. install dataset with ```pip install datasets --upgrade``` 2. launch python; from datasets import loaad_dataset 3. run load_dataset("mozilla-foundation/common_voice_6_0") 4. exit python 5. uninstall datasets; then ```pip install datasets==2.21.0``` 6. launch python; from datasets import loaad_dataset 7. run load_dataset("mozilla-foundation/common_voice_6_0") 8. Everything runs great now ### Expected behavior Be able to download a dataset without error ### Environment info Copy-and-paste the text below in your GitHub issue. - `datasets` version: 3.0.1 - Platform: Linux-6.8.0-1017-aws-x86_64-with-glibc2.39 - Python version: 3.12.4 - `huggingface_hub` version: 0.26.0 - PyArrow version: 17.0.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.6.1
2024-10-18T21:25:23
2024-12-09T09:49:32
null
https://github.com/huggingface/datasets/issues/7238
null
7,238
false
[ "Hi! I'm also getting the same issue - have you been able to find a solution to this? ", "From what I remember, I stayed at the \"downgraded\" version of dataset (2.21.0)" ]
2,597,358,525
[MINOR:TYPO] Update arrow_dataset.py
closed
Fix wrong link. csv kwargs docstring link was pointing to pandas json docs.
2024-10-18T12:10:03
2024-10-24T15:06:43
2024-10-24T15:06:43
https://github.com/huggingface/datasets/pull/7236
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7236", "html_url": "https://github.com/huggingface/datasets/pull/7236", "diff_url": "https://github.com/huggingface/datasets/pull/7236.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7236.patch", "merged_at": "2024-10-24T15:06:43" }
7,236
true
[]
2,594,220,624
No need for dataset_info
closed
save a useless call to /api/datasets/repo_id
2024-10-17T09:54:03
2024-10-22T12:30:40
2024-10-21T16:44:34
https://github.com/huggingface/datasets/pull/7234
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7234", "html_url": "https://github.com/huggingface/datasets/pull/7234", "diff_url": "https://github.com/huggingface/datasets/pull/7234.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7234.patch", "merged_at": "2024-10-21T16:44:34" }
7,234
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7234). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "merging this one for now, let me know if you'd like to see additional changes for error handling (I'll take care of them before doing a release)" ]
2,593,903,113
数据集数量问题
open
### Describe the bug 这里我进行大模型微调,当数据集数量为718时,模型可以正常微调,但是当我添加一个在前718个数据集中的数据或者新增一个数据就会报错 ### Steps to reproduce the bug 1. 这里我的数据集可以微调的最后两个数据集是: { "messages": [ { "role": "user", "content": "完成校正装置设计后需要进行哪些工作?" }, { "role": "assistant", "content": "一旦完成校正装置设计后,需要进行系统实际调校工作,或者进行MATLAB仿真以检查系统的时间响应特性。这时,需将系统建模时省略的部分尽可能加入系统,以保证仿真结果的逼真度。如果由于系统各种固有非线性因素影响,或者由于系统噪声和负载效应等因素的影响,使已校正系统不能满足全部性能指标要求,则需要适当调整校正装置的形式或参数,直到已校正系统满足全部性能指标为止。" } ] } { "messages": [ { "role": "user", "content": "为什么闭环带宽要求会影响串联超前校正的效果?" }, { "role": "assistant", "content": "如果待校正系统不稳定,为了满足相角裕度的要求,超前网络需要提供很大的相角超前量。这会导致超前网络的a值必须选得很大,从而使已校正系统的带宽过大。带宽过大会使通过系统的高频噪声电平变高,可能导致系统失控。因此,闭环带宽的要求会影响串联超前校正的效果。" } ] } 2. 当我把倒数第二个数据集再次添加到数据集中时,此时我的最后三个数据集是 { "messages": [ { "role": "user", "content": "完成校正装置设计后需要进行哪些工作?" }, { "role": "assistant", "content": "一旦完成校正装置设计后,需要进行系统实际调校工作,或者进行MATLAB仿真以检查系统的时间响应特性。这时,需将系统建模时省略的部分尽可能加入系统,以保证仿真结果的逼真度。如果由于系统各种固有非线性因素影响,或者由于系统噪声和负载效应等因素的影响,使已校正系统不能满足全部性能指标要求,则需要适当调整校正装置的形式或参数,直到已校正系统满足全部性能指标为止。" } ] } { "messages": [ { "role": "user", "content": "为什么闭环带宽要求会影响串联超前校正的效果?" }, { "role": "assistant", "content": "如果待校正系统不稳定,为了满足相角裕度的要求,超前网络需要提供很大的相角超前量。这会导致超前网络的a值必须选得很大,从而使已校正系统的带宽过大。带宽过大会使通过系统的高频噪声电平变高,可能导致系统失控。因此,闭环带宽的要求会影响串联超前校正的效果。" } ] } { "messages": [ { "role": "user", "content": "完成校正装置设计后需要进行哪些工作?" }, { "role": "assistant", "content": "一旦完成校正装置设计后,需要进行系统实际调校工作,或者进行MATLAB仿真以检查系统的时间响应特性。这时,需将系统建模时省略的部分尽可能加入系统,以保证仿真结果的逼真度。如果由于系统各种固有非线性因素影响,或者由于系统噪声和负载效应等因素的影响,使已校正系统不能满足全部性能指标要求,则需要适当调整校正装置的形式或参数,直到已校正系统满足全部性能指标为止。" } ] } 这时系统会显示bug: root@autodl-container-027f4cad3d-6baf4e64:~/autodl-tmp# python GLM-4/finetune_demo/finetune.py datasets/ ZhipuAI/glm-4-9b-chat GLM-4/finetune_demo/configs/lora.yaml Loading checkpoint shards: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 10/10 [00:02<00:00, 4.04it/s] The installed version of bitsandbytes was compiled without GPU support. 8-bit optimizers, 8-bit multiplication, and GPU quantization are unavailable. trainable params: 2,785,280 || all params: 9,402,736,640 || trainable%: 0.0296 Generating train split: 0 examples [00:00, ? examples/s]Failed to load JSON from file '/root/autodl-tmp/datasets/train.jsonl' with error <class 'pyarrow.lib.ArrowInvalid'>: JSON parse error: Missing a name for object member. in row 718 Generating train split: 0 examples [00:00, ? examples/s] ╭──────────────────────────────────────────────────────────────────────────────────────────────────────── Traceback (most recent call last) ─────────────────────────────────────────────────────────────────────────────────────────────────────────╮ │ /root/miniconda3/lib/python3.10/site-packages/datasets/packaged_modules/json/json.py:153 in _generate_tables │ │ │ │ 150 │ │ │ │ │ │ │ │ with open( │ │ 151 │ │ │ │ │ │ │ │ │ file, encoding=self.config.encoding, errors=self.con │ │ 152 │ │ │ │ │ │ │ │ ) as f: │ │ ❱ 153 │ │ │ │ │ │ │ │ │ df = pd.read_json(f, dtype_backend="pyarrow") │ │ 154 │ │ │ │ │ │ │ except ValueError: │ │ 155 │ │ │ │ │ │ │ │ logger.error(f"Failed to load JSON from file '{file}' wi │ │ 156 │ │ │ │ │ │ │ │ raise e │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/pandas/io/json/_json.py:815 in read_json │ │ │ │ 812 │ if chunksize: │ │ 813 │ │ return json_reader │ │ 814 │ else: │ │ ❱ 815 │ │ return json_reader.read() │ │ 816 │ │ 817 │ │ 818 class JsonReader(abc.Iterator, Generic[FrameSeriesStrT]): │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/pandas/io/json/_json.py:1025 in read │ │ │ │ 1022 │ │ │ │ │ │ data_lines = data.split("\n") │ │ 1023 │ │ │ │ │ │ obj = self._get_object_parser(self._combine_lines(data_lines)) │ │ 1024 │ │ │ │ else: │ │ ❱ 1025 │ │ │ │ │ obj = self._get_object_parser(self.data) │ │ 1026 │ │ │ │ if self.dtype_backend is not lib.no_default: │ │ 1027 │ │ │ │ │ return obj.convert_dtypes( │ │ 1028 │ │ │ │ │ │ infer_objects=False, dtype_backend=self.dtype_backend │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/pandas/io/json/_json.py:1051 in _get_object_parser │ │ │ │ 1048 │ │ } │ │ 1049 │ │ obj = None │ │ 1050 │ │ if typ == "frame": │ │ ❱ 1051 │ │ │ obj = FrameParser(json, **kwargs).parse() │ │ 1052 │ │ │ │ 1053 │ │ if typ == "series" or obj is None: │ │ 1054 │ │ │ if not isinstance(dtype, bool): │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/pandas/io/json/_json.py:1187 in parse │ │ │ │ 1184 │ │ │ 1185 │ @final │ │ 1186 │ def parse(self): │ │ ❱ 1187 │ │ self._parse() │ │ 1188 │ │ │ │ 1189 │ │ if self.obj is None: │ │ 1190 │ │ │ return None │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/pandas/io/json/_json.py:1403 in _parse │ │ │ │ 1400 │ │ │ │ 1401 │ │ if orient == "columns": │ │ 1402 │ │ │ self.obj = DataFrame( │ │ ❱ 1403 │ │ │ │ ujson_loads(json, precise_float=self.precise_float), dtype=None │ │ 1404 │ │ │ ) │ │ 1405 │ │ elif orient == "split": │ │ 1406 │ │ │ decoded = { │ ╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ ValueError: Trailing data During handling of the above exception, another exception occurred: ╭──────────────────────────────────────────────────────────────────────────────────────────────────────── Traceback (most recent call last) ─────────────────────────────────────────────────────────────────────────────────────────────────────────╮ │ /root/miniconda3/lib/python3.10/site-packages/datasets/builder.py:1997 in _prepare_split_single │ │ │ │ 1994 │ │ │ ) │ │ 1995 │ │ │ try: │ │ 1996 │ │ │ │ _time = time.time() │ │ ❱ 1997 │ │ │ │ for _, table in generator: │ │ 1998 │ │ │ │ │ if max_shard_size is not None and writer._num_bytes > max_shard_size │ │ 1999 │ │ │ │ │ │ num_examples, num_bytes = writer.finalize() │ │ 2000 │ │ │ │ │ │ writer.close() │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/datasets/packaged_modules/json/json.py:156 in _generate_tables │ │ │ │ 153 │ │ │ │ │ │ │ │ │ df = pd.read_json(f, dtype_backend="pyarrow") │ │ 154 │ │ │ │ │ │ │ except ValueError: │ │ 155 │ │ │ │ │ │ │ │ logger.error(f"Failed to load JSON from file '{file}' wi │ │ ❱ 156 │ │ │ │ │ │ │ │ raise e │ │ 157 │ │ │ │ │ │ │ if df.columns.tolist() == [0]: │ │ 158 │ │ │ │ │ │ │ │ df.columns = list(self.config.features) if self.config.f │ │ 159 │ │ │ │ │ │ │ try: │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/datasets/packaged_modules/json/json.py:130 in _generate_tables │ │ │ │ 127 │ │ │ │ │ │ try: │ │ 128 │ │ │ │ │ │ │ while True: │ │ 129 │ │ │ │ │ │ │ │ try: │ │ ❱ 130 │ │ │ │ │ │ │ │ │ pa_table = paj.read_json( │ │ 131 │ │ │ │ │ │ │ │ │ │ io.BytesIO(batch), read_options=paj.ReadOptions( │ │ 132 │ │ │ │ │ │ │ │ │ ) │ │ 133 │ │ │ │ │ │ │ │ │ break │ │ │ │ in pyarrow._json.read_json:308 │ │ │ │ in pyarrow.lib.pyarrow_internal_check_status:154 │ │ │ │ in pyarrow.lib.check_status:91 │ ╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ ArrowInvalid: JSON parse error: Missing a name for object member. in row 718 The above exception was the direct cause of the following exception: ╭──────────────────────────────────────────────────────────────────────────────────────────────────────── Traceback (most recent call last) ─────────────────────────────────────────────────────────────────────────────────────────────────────────╮ │ /root/autodl-tmp/GLM-4/finetune_demo/finetune.py:406 in main │ │ │ │ 403 ): │ │ 404 │ ft_config = FinetuningConfig.from_file(config_file) │ │ 405 │ tokenizer, model = load_tokenizer_and_model(model_dir, peft_config=ft_config.peft_co │ │ ❱ 406 │ data_manager = DataManager(data_dir, ft_config.data_config) │ │ 407 │ │ │ 408 │ train_dataset = data_manager.get_dataset( │ │ 409 │ │ Split.TRAIN, │ │ │ │ /root/autodl-tmp/GLM-4/finetune_demo/finetune.py:204 in __init__ │ │ │ │ 201 │ def __init__(self, data_dir: str, data_config: DataConfig): │ │ 202 │ │ self._num_proc = data_config.num_proc │ │ 203 │ │ │ │ ❱ 204 │ │ self._dataset_dct = _load_datasets( │ │ 205 │ │ │ data_dir, │ │ 206 │ │ │ data_config.data_format, │ │ 207 │ │ │ data_config.data_files, │ │ │ │ /root/autodl-tmp/GLM-4/finetune_demo/finetune.py:189 in _load_datasets │ │ │ │ 186 │ │ num_proc: Optional[int], │ │ 187 ) -> DatasetDict: │ │ 188 │ if data_format == '.jsonl': │ │ ❱ 189 │ │ dataset_dct = load_dataset( │ │ 190 │ │ │ data_dir, │ │ 191 │ │ │ data_files=data_files, │ │ 192 │ │ │ split=None, │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/datasets/load.py:2616 in load_dataset │ │ │ │ 2613 │ │ return builder_instance.as_streaming_dataset(split=split) │ │ 2614 │ │ │ 2615 │ # Download and prepare data │ │ ❱ 2616 │ builder_instance.download_and_prepare( │ │ 2617 │ │ download_config=download_config, │ │ 2618 │ │ download_mode=download_mode, │ │ 2619 │ │ verification_mode=verification_mode, │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/datasets/builder.py:1029 in download_and_prepare │ │ │ │ 1026 │ │ │ │ │ │ │ prepare_split_kwargs["max_shard_size"] = max_shard_size │ │ 1027 │ │ │ │ │ │ if num_proc is not None: │ │ 1028 │ │ │ │ │ │ │ prepare_split_kwargs["num_proc"] = num_proc │ │ ❱ 1029 │ │ │ │ │ │ self._download_and_prepare( │ │ 1030 │ │ │ │ │ │ │ dl_manager=dl_manager, │ │ 1031 │ │ │ │ │ │ │ verification_mode=verification_mode, │ │ 1032 │ │ │ │ │ │ │ **prepare_split_kwargs, │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/datasets/builder.py:1124 in _download_and_prepare │ │ │ │ 1121 │ │ │ │ │ 1122 │ │ │ try: │ │ 1123 │ │ │ │ # Prepare split will record examples associated to the split │ │ ❱ 1124 │ │ │ │ self._prepare_split(split_generator, **prepare_split_kwargs) │ │ 1125 │ │ │ except OSError as e: │ │ 1126 │ │ │ │ raise OSError( │ │ 1127 │ │ │ │ │ "Cannot find data file. " │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/datasets/builder.py:1884 in _prepare_split │ │ │ │ 1881 │ │ │ gen_kwargs = split_generator.gen_kwargs │ │ 1882 │ │ │ job_id = 0 │ │ 1883 │ │ │ with pbar: │ │ ❱ 1884 │ │ │ │ for job_id, done, content in self._prepare_split_single( │ │ 1885 │ │ │ │ │ gen_kwargs=gen_kwargs, job_id=job_id, **_prepare_split_args │ │ 1886 │ │ │ │ ): │ │ 1887 │ │ │ │ │ if done: │ │ │ │ /root/miniconda3/lib/python3.10/site-packages/datasets/builder.py:2040 in _prepare_split_single │ │ │ │ 2037 │ │ │ │ e = e.__context__ │ │ 2038 │ │ │ if isinstance(e, DatasetGenerationError): │ │ 2039 │ │ │ │ raise │ │ ❱ 2040 │ │ │ raise DatasetGenerationError("An error occurred while generating the dataset │ │ 2041 │ │ │ │ 2042 │ │ yield job_id, True, (total_num_examples, total_num_bytes, writer._features, num_ │ │ 2043 │ ╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ DatasetGenerationError: An error occurred while generating the dataset 3.请问是否可以帮我解决 ### Expected behavior 希望问题可以得到解决 ### Environment info Copy-and-paste the text below in your GitHub issue. - `datasets` version: 2.20.0 - Platform: Linux-4.19.90-2107.6.0.0192.8.oe1.bclinux.x86_64-x86_64-with-glibc2.35 - Python version: 3.10.8 - `huggingface_hub` version: 0.24.6 - PyArrow version: 16.1.0 - Pandas version: 2.2.2 - `fsspec` version: 2023.12.2
2024-10-17T07:41:44
2024-10-17T07:41:44
null
https://github.com/huggingface/datasets/issues/7233
null
7,233
false
[]
2,593,720,548
(Super tiny doc update) Mention to_polars
closed
polars is also quite popular now, thus this tiny update can tell users polars is supported
2024-10-17T06:08:53
2024-10-24T23:11:05
2024-10-24T15:06:16
https://github.com/huggingface/datasets/pull/7232
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7232", "html_url": "https://github.com/huggingface/datasets/pull/7232", "diff_url": "https://github.com/huggingface/datasets/pull/7232.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7232.patch", "merged_at": "2024-10-24T15:06:16" }
7,232
true
[ "You are welcome!" ]
2,592,011,737
Fix typo in image dataset docs
closed
Fix typo in image dataset docs. Typo reported by @datavistics.
2024-10-16T14:05:46
2024-10-16T17:06:21
2024-10-16T17:06:19
https://github.com/huggingface/datasets/pull/7231
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7231", "html_url": "https://github.com/huggingface/datasets/pull/7231", "diff_url": "https://github.com/huggingface/datasets/pull/7231.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7231.patch", "merged_at": "2024-10-16T17:06:19" }
7,231
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7231). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,589,531,942
Video support
closed
(wip and experimental) adding the `Video` type based on `VideoReader` from `decord` ```python >>>from datasets import load_dataset >>> ds = load_dataset("path/to/videos", split="train").with_format("torch") >>> print(ds[0]["video"]) <decord.video_reader.VideoReader object at 0x337a47910> >>> print(ds[0]["video"][0]) tensor([[[73, 73, 73], [73, 73, 73], [73, 73, 73], ..., [23, 23, 23], [23, 23, 23], [23, 23, 23]]], dtype=torch.uint8) ``` the storage is the same as for audio and images: `{"path": pa.string(), "bytes": pa.binary()}` and I did a small to keep the hf:// URL in the "path" field if possible, this way the viewer can link to fiels on the hub if possible
2024-10-15T18:17:29
2024-10-24T16:39:51
2024-10-24T16:39:50
https://github.com/huggingface/datasets/pull/7230
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7230", "html_url": "https://github.com/huggingface/datasets/pull/7230", "diff_url": "https://github.com/huggingface/datasets/pull/7230.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7230.patch", "merged_at": "2024-10-24T16:39:50" }
7,230
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7230). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,588,847,398
handle config_name=None in push_to_hub
closed
This caught me out - thought it might be better to explicitly handle None?
2024-10-15T13:48:57
2024-10-24T17:51:52
2024-10-24T17:51:52
https://github.com/huggingface/datasets/pull/7229
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7229", "html_url": "https://github.com/huggingface/datasets/pull/7229", "diff_url": "https://github.com/huggingface/datasets/pull/7229.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7229.patch", "merged_at": null }
7,229
true
[ "not sure it's a good idea, we always need a config name so better have the correct default and not support None (which could lead to think it doesn't have a config name, while it does)" ]
2,587,310,094
Composite (multi-column) features
open
### Feature request Structured data types (graphs etc.) might often be most efficiently stored as multiple columns, which then need to be combined during feature decoding Although it is currently possible to nest features as structs, my impression is that in particular when dealing with e.g. a feature composed of multiple numpy array / ArrayXD's, it would be more efficient to store each ArrayXD as a separate column (though I'm not sure by how much) Perhaps specification / implementation could be supported by something like: ``` features=Features(**{("feature0", "feature1")=Features(feature0=Array2D((None,10), dtype="float32"), feature1=Array2D((None,10), dtype="float32")) ``` ### Motivation Defining efficient composite feature types based on numpy arrays for representing data such as graphs with multiple node and edge attributes is currently challenging. ### Your contribution Possibly able to contribute
2024-10-14T23:59:19
2024-10-15T11:17:15
null
https://github.com/huggingface/datasets/issues/7228
null
7,228
false
[]
2,587,048,312
fast array extraction
open
Implements #7210 using method suggested in https://github.com/huggingface/datasets/pull/7207#issuecomment-2411789307 ```python import numpy as np from datasets import Dataset, Features, Array3D features=Features(**{"array0": Array3D((None, 10, 10), dtype="float32"), "array1": Array3D((None,10,10), dtype="float32")}) dataset = Dataset.from_dict({f"array{i}": [np.zeros((x,10,10), dtype=np.float32) for x in [2000,1000]*25] for i in range(2)}, features=features) ``` ~0.02 s vs 0.9s on main ```python ds = dataset.to_iterable_dataset() t0 = time.time() for ex in ds: pass t1 = time.time() ``` < 0.01 s vs 1.3 s on main @lhoestq I can see this breaks a bunch of array-related tests but can update the test cases if you would support making this change? I also added an Array1D feature which will always be decoded into a numpy array and likewise improves extraction performance: ```python from datasets import Dataset, Features, Array1D, Sequence, Value array_features=Features(**{"array0": Array1D((None,), dtype="float32"), "array1": Array1D((None,), dtype="float32")}) sequence_features=Features(**{"array0": Sequence(feature=Value("float32"), length=-1), "array1": Sequence(feature=Value("float32"), length=-1)}) array_dataset = Dataset.from_dict({f"array{i}": [np.zeros((x,), dtype=np.float32) for x in [20000,10000]*25] for i in range(2)}, features=array_features) sequence_dataset = Dataset.from_dict({f"array{i}": [np.zeros((x,), dtype=np.float32) for x in [20000,10000]*25] for i in range(2)}, features=sequence_features) ```python t0 = time.time() for ex in array_dataset.to_iterable_dataset(): pass t1 = time.time() ``` < 0.01 s ```python t0 = time.time() for ex in sequence_dataset.to_iterable_dataset(): pass t1 = time.time() ``` ~1.1s And also added support for extracting structs of arrays as dicts of numpy arrays: ```python import numpy as np from datasets import Dataset, Features, Array3D, Sequence features=Features(struct={"array0": Array3D((None,10,10), dtype="float32"), "array1": Array3D((None,10,10), dtype="float32")}, _list=Sequence(feature=Array3D((None,10,10), dtype="float32"))) dataset = Dataset.from_dict({"struct": [{f"array{i}": np.zeros((x,10,10), dtype=np.float32) for i in range(2)} for x in [2000,1000]*25], "_list": [[np.zeros((x,10,10), dtype=np.float32) for i in range(2)] for x in [2000,1000]*25]}, features=features) ``` ```python t0 = time.time() for ex in dataset.to_iterable_dataset(): pass t1 = time.time() assert isinstance(ex["struct"]["array0"], np.ndarray) and ex["struct"]["array0"].ndim == 3 ``` ~0.02 s and no exception vs ~7s with an exception on main
2024-10-14T20:51:32
2025-01-28T09:39:26
null
https://github.com/huggingface/datasets/pull/7227
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7227", "html_url": "https://github.com/huggingface/datasets/pull/7227", "diff_url": "https://github.com/huggingface/datasets/pull/7227.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7227.patch", "merged_at": null }
7,227
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7227). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "I've updated the most straightforward failing test cases - lmk if you agree with those.\r\n\r\nMight need some help / pointers on the remaining new failing tests, which seem a little bit more subtle.", "@lhoestq I've had a go at fixing a few more test cases but getting quite uncertain about the remaining ones (as well as about some of the array writing ones that I tried to fix in my last commit). There are still 27 failures vs 21 on main. I'm not completely sure in some cases what intended behaviour is and my understanding of the flow for typed writing is a bit vague.", "@lhoestq do you have any thoughts on this? I wasn't able to resolve all the test issues but the basic functionality seemed useful?" ]
2,586,920,351
Add R as a How to use from the Polars (R) Library as an option
open
### Feature request The boiler plate code to access a dataset via the hugging face file system is very useful. Please addd ## Add Polars (R) option The equivailent code works, because the [Polars-R](https://github.com/pola-rs/r-polars) wrapper has hugging faces funcitonaliy as well. ```r library(polars) df <- pl$read_parquet("hf://datasets/SALURBAL/core__admin_cube_public/core__admin_cube_public.parquet") ``` ## Polars (python) option ![image](https://github.com/user-attachments/assets/8f1bcd19-e578-4b18-b324-7cc00b80ac0a) ## Libraries Currently ![image](https://github.com/user-attachments/assets/0cf50063-f9db-443c-97b4-3ef0664b6e6e) ### Motivation There are many data/analysis/research/statistics teams (particularly in academia and pharma) that use R as the default language. R has great integration with most of the newer data techs (arrow, parquet, polars) and having this included could really help in bringing this community into the hugging faces ecosystem. **This is a small/low-hanging-fruit front end change but would make a big impact expanding the community** ### Your contribution I am not sure which repositroy this should be in, but I have experience in R, Python and JS and happy to submit a PR in the appropriate repository.
2024-10-14T19:56:07
2024-10-14T19:57:13
null
https://github.com/huggingface/datasets/issues/7226
null
7,226
false
[]
2,586,229,216
Huggingface GIT returns null as Content-Type instead of application/x-git-receive-pack-result
open
### Describe the bug We push changes to our datasets programmatically. Our git client jGit reports that the hf git server returns null as Content-Type after a push. ### Steps to reproduce the bug A basic kotlin application: ``` val person = PersonIdent( "padmalcom", "padmalcom@sth.com" ) val cp = UsernamePasswordCredentialsProvider( "padmalcom", "mysecrettoken" ) val git = KGit.cloneRepository { setURI("https://huggingface.co/datasets/sth/images") setTimeout(60) setProgressMonitor(TextProgressMonitor()) setCredentialsProvider(cp) } FileOutputStream("./images/images.csv").apply { writeCsv(images) } git.add { addFilepattern("images.csv") } for (i in images) { FileUtils.copyFile( File("./files/${i.id}"), File("./images/${i.id + File(i.fileName).extension }") ) git.add { addFilepattern("${i.id + File(i.fileName).extension }") } } val revCommit = git.commit { author = person message = "Uploading images at " + LocalDateTime.now() .format(DateTimeFormatter.ISO_DATE_TIME) setCredentialsProvider(cp) } val push = git.push { setCredentialsProvider(cp) } ``` ### Expected behavior The git server is expected to return the Content-Type _application/x-git-receive-pack-result_. ### Environment info It is independent from the datasets library.
2024-10-14T14:33:06
2024-10-14T14:33:06
null
https://github.com/huggingface/datasets/issues/7225
null
7,225
false
[]
2,583,233,980
fallback to default feature casting in case custom features not available during dataset loading
open
a fix for #7223 in case datasets is happy to support this kind of extensibility! seems cool / powerful for allowing sharing of datasets with potentially different feature types
2024-10-12T16:13:56
2024-10-12T16:13:56
null
https://github.com/huggingface/datasets/pull/7224
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7224", "html_url": "https://github.com/huggingface/datasets/pull/7224", "diff_url": "https://github.com/huggingface/datasets/pull/7224.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7224.patch", "merged_at": null }
7,224
true
[]
2,583,231,590
Fallback to arrow defaults when loading dataset with custom features that aren't registered locally
open
### Describe the bug Datasets allows users to create and register custom features. However if datasets are then pushed to the hub, this means that anyone calling load_dataset without registering the custom Features in the same way as the dataset creator will get an error message. It would be nice to offer a fallback in this case. ### Steps to reproduce the bug ```python load_dataset("alex-hh/custom-features-example") ``` (Dataset creation process - must be run in separate session so that NewFeature isn't registered in session in which download is attempted:) ```python from dataclasses import dataclass, field import pyarrow as pa from datasets.features.features import register_feature from datasets import Dataset, Features, Value, load_dataset from datasets import Feature @dataclass class NewFeature(Feature): _type: str = field(default="NewFeature", init=False, repr=False) def __call__(self): return pa.int32() def examples_generator(): for i in range(5): yield {"feature": i} ds = Dataset.from_generator(examples_generator, features=Features(feature=NewFeature())) ds.push_to_hub("alex-hh/custom-features-example") register_feature(NewFeature, "NewFeature") ``` ### Expected behavior It would be nice, and offer greater extensibility, if there was some kind of graceful fallback mechanism in place for cases where user-defined features are stored in the dataset but not available locally. ### Environment info 3.0.2
2024-10-12T16:08:20
2024-10-12T16:08:20
null
https://github.com/huggingface/datasets/issues/7223
null
7,223
false
[]
2,582,678,033
TypeError: Couldn't cast array of type string to null in long json
open
### Describe the bug In general, changing the type from string to null is allowed within a dataset — there are even examples of this in the documentation. However, if the dataset is large and unevenly distributed, this allowance stops working. The schema gets locked in after reading a chunk. Consequently, if all values in the first chunk of a field are, for example, null, the field will be locked as type null, and if a string appears in that field in the second chunk, it will trigger this error: <details> <summary>Traceback </summary> ``` TypeError Traceback (most recent call last) [/usr/local/lib/python3.10/dist-packages/datasets/builder.py](https://localhost:8080/#) in _prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id) 1868 try: -> 1869 writer.write_table(table) 1870 except CastError as cast_error: 14 frames [/usr/local/lib/python3.10/dist-packages/datasets/arrow_writer.py](https://localhost:8080/#) in write_table(self, pa_table, writer_batch_size) 579 pa_table = pa_table.combine_chunks() --> 580 pa_table = table_cast(pa_table, self._schema) 581 if self.embed_local_files: [/usr/local/lib/python3.10/dist-packages/datasets/table.py](https://localhost:8080/#) in table_cast(table, schema) 2291 if table.schema != schema: -> 2292 return cast_table_to_schema(table, schema) 2293 elif table.schema.metadata != schema.metadata: [/usr/local/lib/python3.10/dist-packages/datasets/table.py](https://localhost:8080/#) in cast_table_to_schema(table, schema) 2244 ) -> 2245 arrays = [ 2246 cast_array_to_feature( [/usr/local/lib/python3.10/dist-packages/datasets/table.py](https://localhost:8080/#) in <listcomp>(.0) 2245 arrays = [ -> 2246 cast_array_to_feature( 2247 table[name] if name in table_column_names else pa.array([None] * len(table), type=schema.field(name).type), [/usr/local/lib/python3.10/dist-packages/datasets/table.py](https://localhost:8080/#) in wrapper(array, *args, **kwargs) 1794 if isinstance(array, pa.ChunkedArray): -> 1795 return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) 1796 else: [/usr/local/lib/python3.10/dist-packages/datasets/table.py](https://localhost:8080/#) in <listcomp>(.0) 1794 if isinstance(array, pa.ChunkedArray): -> 1795 return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) 1796 else: [/usr/local/lib/python3.10/dist-packages/datasets/table.py](https://localhost:8080/#) in cast_array_to_feature(array, feature, allow_primitive_to_str, allow_decimal_to_str) 2101 elif not isinstance(feature, (Sequence, dict, list, tuple)): -> 2102 return array_cast( 2103 array, [/usr/local/lib/python3.10/dist-packages/datasets/table.py](https://localhost:8080/#) in wrapper(array, *args, **kwargs) 1796 else: -> 1797 return func(array, *args, **kwargs) 1798 [/usr/local/lib/python3.10/dist-packages/datasets/table.py](https://localhost:8080/#) in array_cast(array, pa_type, allow_primitive_to_str, allow_decimal_to_str) 1947 if pa.types.is_null(pa_type) and not pa.types.is_null(array.type): -> 1948 raise TypeError(f"Couldn't cast array of type {_short_str(array.type)} to {_short_str(pa_type)}") 1949 return array.cast(pa_type) TypeError: Couldn't cast array of type string to null The above exception was the direct cause of the following exception: DatasetGenerationError Traceback (most recent call last) [<ipython-input-353-e02f83980611>](https://localhost:8080/#) in <cell line: 1>() ----> 1 dd = load_dataset("json", data_files=["TEST.json"]) [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://localhost:8080/#) in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, keep_in_memory, save_infos, revision, token, streaming, num_proc, storage_options, trust_remote_code, **config_kwargs) 2094 2095 # Download and prepare data -> 2096 builder_instance.download_and_prepare( 2097 download_config=download_config, 2098 download_mode=download_mode, [/usr/local/lib/python3.10/dist-packages/datasets/builder.py](https://localhost:8080/#) in download_and_prepare(self, output_dir, download_config, download_mode, verification_mode, dl_manager, base_path, file_format, max_shard_size, num_proc, storage_options, **download_and_prepare_kwargs) 922 if num_proc is not None: 923 prepare_split_kwargs["num_proc"] = num_proc --> 924 self._download_and_prepare( 925 dl_manager=dl_manager, 926 verification_mode=verification_mode, [/usr/local/lib/python3.10/dist-packages/datasets/builder.py](https://localhost:8080/#) in _download_and_prepare(self, dl_manager, verification_mode, **prepare_split_kwargs) 997 try: 998 # Prepare split will record examples associated to the split --> 999 self._prepare_split(split_generator, **prepare_split_kwargs) 1000 except OSError as e: 1001 raise OSError( [/usr/local/lib/python3.10/dist-packages/datasets/builder.py](https://localhost:8080/#) in _prepare_split(self, split_generator, file_format, num_proc, max_shard_size) 1738 job_id = 0 1739 with pbar: -> 1740 for job_id, done, content in self._prepare_split_single( 1741 gen_kwargs=gen_kwargs, job_id=job_id, **_prepare_split_args 1742 ): [/usr/local/lib/python3.10/dist-packages/datasets/builder.py](https://localhost:8080/#) in _prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id) 1894 if isinstance(e, DatasetGenerationError): 1895 raise -> 1896 raise DatasetGenerationError("An error occurred while generating the dataset") from e 1897 1898 yield job_id, True, (total_num_examples, total_num_bytes, writer._features, num_shards, shard_lengths) DatasetGenerationError: An error occurred while generating the dataset ``` </details> ### Steps to reproduce the bug ```python import json from datasets import load_dataset with open("TEST.json", "w") as f: row = {"ballast": "qwerty" * 1000, "b": None} row_str = json.dumps(row) + "\n" line_size = len(row_str) chunk_size = 10 << 20 lines_in_chunk = chunk_size // line_size + 1 print(f"Writing {lines_in_chunk} lines") for i in range(lines_in_chunk): f.write(row_str) null_row = {"ballast": "Gotcha", "b": "Not Null"} f.write(json.dumps(null_row) + "\n") load_dataset("json", data_files=["TEST.json"]) ``` ### Expected behavior Concatenation of the chunks without errors ### Environment info - `datasets` version: 3.0.1 - Platform: Linux-6.1.85+-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.24.7 - PyArrow version: 16.1.0 - Pandas version: 2.2.2 - `fsspec` version: 2024.6.1
2024-10-12T08:14:59
2025-07-21T03:07:32
null
https://github.com/huggingface/datasets/issues/7222
null
7,222
false
[ "I am encountering this same issue. It seems that the library manages to recognise an optional column (but not **exclusively** null) if there is at least one non-null instance within the same file. For example, given a `test_0.jsonl` file:\r\n```json\r\n{\"a\": \"a1\", \"b\": \"b1\", \"c\": null, \"d\": null}\r\n{\"a\": \"a2\", \"b\": null, \"c\": \"c2\", \"d\": null}\r\n```\r\nthe data is correctly loaded, recognising that columns `b` & `c` are optional, while `d` is null.\r\n```python\r\n{'a': ['a1', 'a2'], 'b': ['b1', None], 'c': [None, 'c2'], 'd': [None, None]}\r\n```\r\n\r\nBut if the `config` has another file, say `test_1.jsonl` where `d` now has some non-null values:\r\n```json\r\n{\"a\": null, \"b\": \"b3\", \"c\": \"c3\", \"d\": \"d3\"}\r\n{\"a\": \"a4\", \"b\": \"b4\", \"c\": null, \"d\": null}\r\n```\r\nthen, an error is raised:\r\n```\r\nTypeError Traceback (most recent call last)\r\n\r\n[/usr/local/lib/python3.10/dist-packages/datasets/builder.py](https://localhost:8080/#) in _prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id)\r\n 1869 try:\r\n-> 1870 writer.write_table(table)\r\n 1871 except CastError as cast_error:\r\n\r\n14 frames\r\n\r\nTypeError: Couldn't cast array of type string to null\r\n\r\n\r\nThe above exception was the direct cause of the following exception:\r\n\r\nDatasetGenerationError Traceback (most recent call last)\r\n\r\n[/usr/local/lib/python3.10/dist-packages/datasets/builder.py](https://localhost:8080/#) in _prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id)\r\n 1895 if isinstance(e, DatasetGenerationError):\r\n 1896 raise\r\n-> 1897 raise DatasetGenerationError(\"An error occurred while generating the dataset\") from e\r\n 1898 \r\n 1899 yield job_id, True, (total_num_examples, total_num_bytes, writer._features, num_shards, shard_lengths)\r\n\r\nDatasetGenerationError: An error occurred while generating the dataset\r\n```\r\n\r\n---\r\n\r\nI have created a [sample repository](https://huggingface.co/datasets/KurtMica/optional_columns_mutiple_files) if that helps. Interestingly, the dataset viewer correctly shows the data across files, although it still indicates the above error.", " Managed to find a workaround, by [specifying the features explicitly](https://huggingface.co/docs/datasets/main/en/loading#specify-features), which is also possible to do directly using the [YAML file configuration](https://discuss.huggingface.co/t/appropriate-yaml-for-dataset-info-list-float/74418).", "I hit the same issue for `datasets 3.2.0`. Given the two jsonl files with the same content but different ordering, `load_dataset` worked for one but did not work for the other.\n\n```\nfrom datasets import load_dataset\n\nissues_dataset = load_dataset(\n \"json\", data_files=\"NeMo-issues-fixed.jsonl\", split=\"train\"\n)\nissues_dataset\n```\n\nFor [NeMo-issues.jsonl](https://github.com/renweizhukov/jupyter-lab-notebook/blob/main/hugging-face-nlp-course/NeMo-issues.jsonl), I got an exception:\n\n```\n---------------------------------------------------------------------------\nTypeError Traceback (most recent call last)\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/builder.py:1870](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/builder.py#line=1869), in ArrowBasedBuilder._prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id)\n 1869 try:\n-> 1870 writer.write_table(table)\n 1871 except CastError as cast_error:\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/arrow_writer.py:622](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/arrow_writer.py#line=621), in ArrowWriter.write_table(self, pa_table, writer_batch_size)\n 621 pa_table = pa_table.combine_chunks()\n--> 622 pa_table = table_cast(pa_table, self._schema)\n 623 if self.embed_local_files:\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py:2292](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py#line=2291), in table_cast(table, schema)\n 2291 if table.schema != schema:\n-> 2292 return cast_table_to_schema(table, schema)\n 2293 elif table.schema.metadata != schema.metadata:\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py:2246](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py#line=2245), in cast_table_to_schema(table, schema)\n 2240 raise CastError(\n 2241 f\"Couldn't cast\\n{_short_str(table.schema)}\\nto\\n{_short_str(features)}\\nbecause column names don't match\",\n 2242 table_column_names=table.column_names,\n 2243 requested_column_names=list(features),\n 2244 )\n 2245 arrays = [\n-> 2246 cast_array_to_feature(\n 2247 table[name] if name in table_column_names else pa.array([None] * len(table), type=schema.field(name).type),\n 2248 feature,\n 2249 )\n 2250 for name, feature in features.items()\n 2251 ]\n 2252 return pa.Table.from_arrays(arrays, schema=schema)\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py:1795](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py#line=1794), in _wrap_for_chunked_arrays.<locals>.wrapper(array, *args, **kwargs)\n 1794 if isinstance(array, pa.ChunkedArray):\n-> 1795 return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])\n 1796 else:\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py:2102](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py#line=2101), in cast_array_to_feature(array, feature, allow_primitive_to_str, allow_decimal_to_str)\n 2101 elif not isinstance(feature, (Sequence, dict, list, tuple)):\n-> 2102 return array_cast(\n 2103 array,\n 2104 feature(),\n 2105 allow_primitive_to_str=allow_primitive_to_str,\n 2106 allow_decimal_to_str=allow_decimal_to_str,\n 2107 )\n 2108 raise TypeError(f\"Couldn't cast array of type\\n{_short_str(array.type)}\\nto\\n{_short_str(feature)}\")\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py:1797](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py#line=1796), in _wrap_for_chunked_arrays.<locals>.wrapper(array, *args, **kwargs)\n 1796 else:\n-> 1797 return func(array, *args, **kwargs)\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py:1948](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/table.py#line=1947), in array_cast(array, pa_type, allow_primitive_to_str, allow_decimal_to_str)\n 1947 if pa.types.is_null(pa_type) and not pa.types.is_null(array.type):\n-> 1948 raise TypeError(f\"Couldn't cast array of type {_short_str(array.type)} to {_short_str(pa_type)}\")\n 1949 return array.cast(pa_type)\n\nTypeError: Couldn't cast array of type string to null\n\nThe above exception was the direct cause of the following exception:\n\nDatasetGenerationError Traceback (most recent call last)\nCell In[73], line 3\n 1 from datasets import load_dataset\n----> 3 issues_dataset = load_dataset(\n 4 \"json\", data_files=\"NeMo-issues.jsonl\", split=\"train\"\n 5 )\n 6 issues_dataset\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/load.py:2151](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/load.py#line=2150), in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, keep_in_memory, save_infos, revision, token, streaming, num_proc, storage_options, trust_remote_code, **config_kwargs)\n 2148 return builder_instance.as_streaming_dataset(split=split)\n 2150 # Download and prepare data\n-> 2151 builder_instance.download_and_prepare(\n 2152 download_config=download_config,\n 2153 download_mode=download_mode,\n 2154 verification_mode=verification_mode,\n 2155 num_proc=num_proc,\n 2156 storage_options=storage_options,\n 2157 )\n 2159 # Build dataset for splits\n 2160 keep_in_memory = (\n 2161 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size)\n 2162 )\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/builder.py:924](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/builder.py#line=923), in DatasetBuilder.download_and_prepare(self, output_dir, download_config, download_mode, verification_mode, dl_manager, base_path, file_format, max_shard_size, num_proc, storage_options, **download_and_prepare_kwargs)\n 922 if num_proc is not None:\n 923 prepare_split_kwargs[\"num_proc\"] = num_proc\n--> 924 self._download_and_prepare(\n 925 dl_manager=dl_manager,\n 926 verification_mode=verification_mode,\n 927 **prepare_split_kwargs,\n 928 **download_and_prepare_kwargs,\n 929 )\n 930 # Sync info\n 931 self.info.dataset_size = sum(split.num_bytes for split in self.info.splits.values())\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/builder.py:1000](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/builder.py#line=999), in DatasetBuilder._download_and_prepare(self, dl_manager, verification_mode, **prepare_split_kwargs)\n 996 split_dict.add(split_generator.split_info)\n 998 try:\n 999 # Prepare split will record examples associated to the split\n-> 1000 self._prepare_split(split_generator, **prepare_split_kwargs)\n 1001 except OSError as e:\n 1002 raise OSError(\n 1003 \"Cannot find data file. \"\n 1004 + (self.manual_download_instructions or \"\")\n 1005 + \"\\nOriginal erro[r:\\n](file:///R:/n)\"\n 1006 + str(e)\n 1007 ) from None\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/builder.py:1741](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/builder.py#line=1740), in ArrowBasedBuilder._prepare_split(self, split_generator, file_format, num_proc, max_shard_size)\n 1739 job_id = 0\n 1740 with pbar:\n-> 1741 for job_id, done, content in self._prepare_split_single(\n 1742 gen_kwargs=gen_kwargs, job_id=job_id, **_prepare_split_args\n 1743 ):\n 1744 if done:\n 1745 result = content\n\nFile [~/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/builder.py:1897](http://localhost:8888/home/renwei/anaconda3/envs/llm/lib/python3.12/site-packages/datasets/builder.py#line=1896), in ArrowBasedBuilder._prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id)\n 1895 if isinstance(e, DatasetGenerationError):\n 1896 raise\n-> 1897 raise DatasetGenerationError(\"An error occurred while generating the dataset\") from e\n 1899 yield job_id, True, (total_num_examples, total_num_bytes, writer._features, num_shards, shard_lengths)\n\nDatasetGenerationError: An error occurred while generating the dataset\n```\n\nFor [NeMo-issues-fixed.json](https://github.com/renweizhukov/jupyter-lab-notebook/blob/main/hugging-face-nlp-course/NeMo-issues-fixed.jsonl) which consists of the last 1000 lines and then the first 9000 lines of NeMo-issues.jsonl, I could load the data:\n\n```\nDataset({\n features: ['url', 'repository_url', 'labels_url', 'comments_url', 'events_url', 'html_url', 'id', 'node_id', 'number', 'title', 'user', 'labels', 'state', 'locked', 'assignee', 'assignees', 'milestone', 'comments', 'created_at', 'updated_at', 'closed_at', 'author_association', 'sub_issues_summary', 'active_lock_reason', 'draft', 'pull_request', 'body', 'closed_by', 'reactions', 'timeline_url', 'performed_via_github_app', 'state_reason'],\n num_rows: 10000\n})\n```", "having the same issue as well!", "Is this fixed in the latest version?", "@DronHazra @renweizhukov Is this fixed in the latest version?" ]
2,582,114,631
add CustomFeature base class to support user-defined features with encoding/decoding logic
closed
intended as fix for #7220 if this kind of extensibility is something that datasets is willing to support! ```python from datasets.features.features import CustomFeature class ListOfStrs(CustomFeature): requires_encoding = True def _encode_example(self, value): if isinstance(value, str): return [str] else: return value feats = Features(strlist=ListOfStrs()) feats.encode_example({"strlist": "a"})["strlist"] == feats["strlist"].encode_example("a") ```
2024-10-11T20:10:27
2025-01-28T09:40:29
2025-01-28T09:40:29
https://github.com/huggingface/datasets/pull/7221
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7221", "html_url": "https://github.com/huggingface/datasets/pull/7221", "diff_url": "https://github.com/huggingface/datasets/pull/7221.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7221.patch", "merged_at": null }
7,221
true
[ "@lhoestq would you be open to supporting this kind of extensibility?", "I suggested a fix in https://github.com/huggingface/datasets/issues/7220 that would not necessarily require a parent class for custom features, lmk what you think" ]
2,582,036,110
Custom features not compatible with special encoding/decoding logic
open
### Describe the bug It is possible to register custom features using datasets.features.features.register_feature (https://github.com/huggingface/datasets/pull/6727) However such features are not compatible with Features.encode_example/decode_example if they require special encoding / decoding logic because encode_nested_example / decode_nested_example checks whether the feature is in a fixed list of encodable types: https://github.com/huggingface/datasets/blob/16a121d7821a7691815a966270f577e2c503473f/src/datasets/features/features.py#L1349 This prevents the extensibility of features to complex cases ### Steps to reproduce the bug ```python class ListOfStrs: def encode_example(self, value): if isinstance(value, str): return [str] else: return value feats = Features(strlist=ListOfStrs()) assert feats.encode_example({"strlist": "a"})["strlist"] = feats["strlist"].encode_example("a")} ``` ### Expected behavior Registered feature types should be encoded based on some property of the feature (e.g. requires_encoding)? ### Environment info 3.0.2
2024-10-11T19:20:11
2024-11-08T15:10:58
null
https://github.com/huggingface/datasets/issues/7220
null
7,220
false
[ "I think you can fix this simply by replacing the line with hardcoded features with `hastattr(schema, \"encode_example\")` actually", "#7284 " ]
2,581,708,084
bump fsspec
closed
null
2024-10-11T15:56:36
2024-10-14T08:21:56
2024-10-14T08:21:55
https://github.com/huggingface/datasets/pull/7219
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7219", "html_url": "https://github.com/huggingface/datasets/pull/7219", "diff_url": "https://github.com/huggingface/datasets/pull/7219.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7219.patch", "merged_at": "2024-10-14T08:21:55" }
7,219
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7219). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,581,095,098
ds.map(f, num_proc=10) is slower than df.apply
open
### Describe the bug pandas columns: song_id, song_name ds = Dataset.from_pandas(df) def has_cover(song_name): if song_name is None or pd.isna(song_name): return False return 'cover' in song_name.lower() df['has_cover'] = df.song_name.progress_apply(has_cover) ds = ds.map(lambda x: {'has_cover': has_cover(x['song_name'])}, num_proc=10) time cost: 1. df.apply: 100%|██████████| 12500592/12500592 [00:13<00:00, 959825.47it/s] 2. ds.map: Map (num_proc=10):  31%  3899028/12500592 [00:28<00:38, 222532.89 examples/s] ### Steps to reproduce the bug pandas columns: song_id, song_name ds = Dataset.from_pandas(df) def has_cover(song_name): if song_name is None or pd.isna(song_name): return False return 'cover' in song_name.lower() df['has_cover'] = df.song_name.progress_apply(has_cover) ds = ds.map(lambda x: {'has_cover': has_cover(x['song_name'])}, num_proc=10) ### Expected behavior ds.map is ~num_proc faster than df.apply ### Environment info pandas: 2.2.2 datasets: 2.19.1
2024-10-11T11:04:05
2025-02-28T21:21:01
null
https://github.com/huggingface/datasets/issues/7217
null
7,217
false
[ "Hi ! `map()` reads all the columns and writes the resulting dataset with all the columns as well, while df.column_name.apply only reads and writes one column and does it in memory. So this is speed difference is actually expected.\r\n\r\nMoreover using multiprocessing on a dataset that lives in memory (from_pandas uses the same in-memory data as the pandas DataFrame while load_dataset or from_generator load from disk) requires to copy the data to each subprocess which can also be slow. Data loaded from disk don't need to be copied though since they work as a form of shared memory thanks to memory mapping.\r\n\r\nHowever you can make you map() call much faster by making it read and write only the column you want:\r\n\r\n```python\r\nhas_cover_ds = ds.map(lambda song_name: {'has_cover': has_cover(song_name)}, input_columns=[\"song_name\"], remove_columns=ds.column_names) # outputs a dataset with 1 column\r\nds = ds.concatenate_datasets([ds, has_cover_ds], axis=1)\r\n```\r\n\r\nand if your dataset is loaded from disk you can pass num_proc=10 and get a nice speed up as well (no need to copy the data to subprocesses)", "Isn't there a way to do memory mapping with the in-memory dataset without saving it to disk?", "Maybe saving it to a memory-mapped filesystem? It'd be like a trick to make datasets save to \"disk\" but actually it's memory. But it feels like there should be a better \"automatic\" way provided by `datasets`." ]
2,579,942,939
Iterable dataset map with explicit features causes slowdown for Sequence features
open
### Describe the bug When performing map, it's nice to be able to pass the new feature type, and indeed required by interleave and concatenate datasets. However, this can cause a major slowdown for certain types of array features due to the features being re-encoded. This is separate to the slowdown reported in #7206 ### Steps to reproduce the bug ``` from datasets import Dataset, Features, Array3D, Sequence, Value import numpy as np import time features=Features(**{"array0": Sequence(feature=Value("float32"), length=-1), "array1": Sequence(feature=Value("float32"), length=-1)}) dataset = Dataset.from_dict({f"array{i}": [np.zeros((x,), dtype=np.float32) for x in [5000,10000]*25] for i in range(2)}, features=features) ``` ``` ds = dataset.to_iterable_dataset() ds = ds.with_format("numpy").map(lambda x: x) t0 = time.time() for ex in ds: pass t1 = time.time() ``` ~1.5 s on main ``` ds = dataset.to_iterable_dataset() ds = ds.with_format("numpy").map(lambda x: x, features=features) t0 = time.time() for ex in ds: pass t1 = time.time() ``` ~ 3 s on main ### Expected behavior I'm not 100% sure whether passing new feature types to formatted outputs of map should be supported or not, but assuming it should, then there should be a cost-free way to specify the new feature type - knowing feature type is required by interleave_datasets and concatenate_datasets for example ### Environment info 3.0.2
2024-10-10T22:08:20
2024-10-10T22:10:32
null
https://github.com/huggingface/datasets/issues/7215
null
7,215
false
[]
2,578,743,713
Formatted map + with_format(None) changes array dtype for iterable datasets
open
### Describe the bug When applying with_format -> map -> with_format(None), array dtypes seem to change, even if features are passed ### Steps to reproduce the bug ```python features=Features(**{"array0": Array3D((None, 10, 10), dtype="float32")}) dataset = Dataset.from_dict({f"array0": [np.zeros((100,10,10), dtype=np.float32)]*25}, features=features) ds = dataset.to_iterable_dataset().with_format("numpy").map(lambda x: x, features=features) ex_0 = next(iter(ds)) ds = dataset.to_iterable_dataset().with_format("numpy").map(lambda x: x, features=features).with_format(None) ex_1 = next(iter(ds)) assert ex_1["array0"].dtype == ex_0["array0"].dtype, f"{ex_1['array0'].dtype} {ex_0['array0'].dtype}" ``` ### Expected behavior Dtypes should be preserved. ### Environment info 3.0.2
2024-10-10T12:45:16
2024-10-12T16:55:57
null
https://github.com/huggingface/datasets/issues/7214
null
7,214
false
[ "possibly due to this logic:\r\n\r\n```python\r\n def _arrow_array_to_numpy(self, pa_array: pa.Array) -> np.ndarray:\r\n if isinstance(pa_array, pa.ChunkedArray):\r\n if isinstance(pa_array.type, _ArrayXDExtensionType):\r\n # don't call to_pylist() to preserve dtype of the fixed-size array\r\n zero_copy_only = _is_zero_copy_only(pa_array.type.storage_dtype, unnest=True)\r\n array: List = [\r\n row for chunk in pa_array.chunks for row in chunk.to_numpy(zero_copy_only=zero_copy_only)\r\n ]\r\n else:\r\n zero_copy_only = _is_zero_copy_only(pa_array.type) and all(\r\n not _is_array_with_nulls(chunk) for chunk in pa_array.chunks\r\n )\r\n array: List = [\r\n row for chunk in pa_array.chunks for row in chunk.to_numpy(zero_copy_only=zero_copy_only)\r\n ]\r\n else:\r\n if isinstance(pa_array.type, _ArrayXDExtensionType):\r\n # don't call to_pylist() to preserve dtype of the fixed-size array\r\n zero_copy_only = _is_zero_copy_only(pa_array.type.storage_dtype, unnest=True)\r\n array: List = pa_array.to_numpy(zero_copy_only=zero_copy_only)\r\n else:\r\n zero_copy_only = _is_zero_copy_only(pa_array.type) and not _is_array_with_nulls(pa_array)\r\n array: List = pa_array.to_numpy(zero_copy_only=zero_copy_only).tolist()\r\n```" ]
2,578,675,565
Add with_rank to Dataset.from_generator
open
### Feature request Add `with_rank` to `Dataset.from_generator` similar to `Dataset.map` and `Dataset.filter`. ### Motivation As for `Dataset.map` and `Dataset.filter`, this is useful when creating cache files using multi-GPU, where the rank can be used to select GPU IDs. For now, rank can be added in the `gen_kwars` argument; however, this, in turn, includes the rank when computing the fingerprint. ### Your contribution Added #7199 which passes rank based on the `job_id` set by `num_proc`.
2024-10-10T12:15:29
2024-10-10T12:17:11
null
https://github.com/huggingface/datasets/issues/7213
null
7,213
false
[]
2,578,641,259
Windows do not supprot signal.alarm and singal.signal
open
### Describe the bug signal.alarm and signal.signal are used in the load.py module, but these are not supported by Windows. ### Steps to reproduce the bug lighteval accelerate --model_args "pretrained=gpt2,trust_remote_code=True" --tasks "community|kinit_sts" --custom_tasks "community_tasks/kinit_evals.py" --output_dir "./evals" ### Expected behavior proceed with input(..) method ### Environment info Windows 11
2024-10-10T12:00:19
2024-10-10T12:00:19
null
https://github.com/huggingface/datasets/issues/7212
null
7,212
false
[]
2,576,400,502
Describe only selected fields in README
open
### Feature request Hi Datasets team! Is it possible to add the ability to describe only selected fields of the dataset files in `README.md`? For example, I have this open dataset ([open-llm-leaderboard/results](https://huggingface.co/datasets/open-llm-leaderboard/results?row=0)) and I want to describe only some fields in order not to overcomplicate the Dataset Preview and filter out some fields ### Motivation The `Results` dataset for the Open LLM Leaderboard contains json files with a complex nested structure. I would like to add `README.md` there to use the SQL console, for example. But if I describe the structure of this dataset completely, it will overcomplicate the use of Dataset Preview and the total number of columns will exceed 50 ### Your contribution I'm afraid I'm not familiar with the project structure, so I won't be able to open a PR, but I'll try to help with something else if possible
2024-10-09T16:25:47
2024-10-09T16:25:47
null
https://github.com/huggingface/datasets/issues/7211
null
7,211
false
[]
2,575,883,939
Convert Array features to numpy arrays rather than lists by default
open
### Feature request It is currently quite easy to cause massive slowdowns when using datasets and not familiar with the underlying data conversions by e.g. making bad choices of formatting. Would it be more user-friendly to set defaults that avoid this as much as possible? e.g. format Array features as numpy arrays rather than python lists ### Motivation Default array formatting leads to slow performance: e.g. ```python import numpy as np from datasets import Dataset, Features, Array3D features=Features(**{"array0": Array3D((None, 10, 10), dtype="float32"), "array1": Array3D((None,10,10), dtype="float32")}) dataset = Dataset.from_dict({f"array{i}": [np.zeros((x,10,10), dtype=np.float32) for x in [2000,1000]*25] for i in range(2)}, features=features) ``` ```python t0 = time.time() for ex in ds: pass t1 = time.time() ``` ~1.4 s ```python ds = dataset.to_iterable_dataset() t0 = time.time() for ex in ds: pass t1 = time.time() ``` ~10s ```python ds = dataset.with_format("numpy") t0 = time.time() for ex in ds: pass t1 = time.time() ``` ~0.04s ```python ds = dataset.to_iterable_dataset().with_format("numpy") t0 = time.time() for ex in ds: pass t1 = time.time() ``` ~0.04s ### Your contribution May be able to contribute
2024-10-09T13:05:21
2024-10-09T13:05:21
null
https://github.com/huggingface/datasets/issues/7210
null
7,210
false
[]
2,575,526,651
Preserve features in iterable dataset.filter
closed
Fixes example in #7208 - I'm not sure what other checks I should do? @lhoestq I also haven't thought hard about the concatenate / interleaving example iterables but think this might work assuming that features are either all identical or None?
2024-10-09T10:42:05
2024-10-16T11:27:22
2024-10-09T16:04:07
https://github.com/huggingface/datasets/pull/7209
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7209", "html_url": "https://github.com/huggingface/datasets/pull/7209", "diff_url": "https://github.com/huggingface/datasets/pull/7209.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7209.patch", "merged_at": "2024-10-09T16:04:07" }
7,209
true
[ "Yes your assumption on concatenate/interleave is ok imo.\r\n\r\nIt seems the TypedExamplesIterable can slow down things, it should take formatting into account to not convert numpy arrays to python lists\r\n\r\nright now it's slow (unrelatedly to your PR):\r\n\r\n```python\r\n>>> ds = Dataset.from_dict({\"a\": np.zeros((1000, 32, 32))}).to_iterable_dataset().with_format(\"np\")\r\n>>> filtered_ds = ds.filter(lambda x: True)\r\n>>> %time sum(1 for _ in ds)\r\nCPU times: user 175 ms, sys: 8.1 ms, total: 183 ms\r\nWall time: 184 ms\r\n1000\r\n>>> %time sum(1 for _ in filtered_ds)\r\nCPU times: user 4.1 s, sys: 8.41 ms, total: 4.1 s\r\nWall time: 4.12 s\r\n1000\r\n```", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7209). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "> It seems the TypedExamplesIterable can slow down things, it should take formatting into account to not convert numpy arrays to python lists\r\n\r\nShould be fixed by updated #7207 I hope!" ]
2,575,484,256
Iterable dataset.filter should not override features
closed
### Describe the bug When calling filter on an iterable dataset, the features get set to None ### Steps to reproduce the bug import numpy as np import time from datasets import Dataset, Features, Array3D ```python features=Features(**{"array0": Array3D((None, 10, 10), dtype="float32"), "array1": Array3D((None,10,10), dtype="float32")}) dataset = Dataset.from_dict({f"array{i}": [np.zeros((x,10,10), dtype=np.float32) for x in [2000,1000]*25] for i in range(2)}, features=features) ds = dataset.to_iterable_dataset() orig_column_names = ds.column_names ds = ds.filter(lambda x: True) assert ds.column_names == orig_column_names ``` ### Expected behavior Filter should preserve features information ### Environment info 3.0.2
2024-10-09T10:23:45
2024-10-09T16:08:46
2024-10-09T16:08:45
https://github.com/huggingface/datasets/issues/7208
null
7,208
false
[ "closed by https://github.com/huggingface/datasets/pull/7209, thanks @alex-hh !" ]
2,573,582,335
apply formatting after iter_arrow to speed up format -> map, filter for iterable datasets
closed
I got to this by hacking around a bit but it seems to solve #7206 I have no idea if this approach makes sense or would break something else? Could maybe work on a full pr if this looks reasonable @lhoestq ? I imagine the same issue might affect other iterable dataset methods?
2024-10-08T15:44:53
2025-01-14T18:36:03
2025-01-14T16:59:30
https://github.com/huggingface/datasets/pull/7207
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7207", "html_url": "https://github.com/huggingface/datasets/pull/7207", "diff_url": "https://github.com/huggingface/datasets/pull/7207.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7207.patch", "merged_at": "2025-01-14T16:59:30" }
7,207
true
[ "I think the problem is that the underlying ex_iterable will not use iter_arrow unless the formatting type is arrow, which leads to conversion from arrow -> python -> numpy in this case rather than arrow -> numpy.\r\n\r\nIdea of updated fix is to use the ex_iterable's iter_arrow in any case where it's available and any formatting is specified. The formatter then works directly on arrow tables; the outputs of the formatter get passed to the function to be mapped.\r\n\r\nWith updated version:\r\n\r\n```python\r\nimport numpy as np\r\nimport time\r\nfrom datasets import Dataset, Features, Array3D\r\n\r\nfeatures=Features(**{\"array0\": Array3D((None, 10, 10), dtype=\"float32\"), \"array1\": Array3D((None,10,10), dtype=\"float32\")})\r\ndataset = Dataset.from_dict({f\"array{i}\": [np.zeros((x,10,10), dtype=np.float32) for x in [2000,1000]*25] for i in range(2)}, features=features)\r\n```\r\n\r\n```python\r\nds = dataset.to_iterable_dataset()\r\nds = ds.with_format(\"numpy\").map(lambda x: x, batched=True, batch_size=10)\r\nt0 = time.time()\r\nfor ex in ds:\r\n pass\r\nt1 = time.time()\r\n```\r\nTotal time: < 0.01s (~30s on main)\r\n\r\n```python\r\nds = dataset.to_iterable_dataset()\r\nds = ds.with_format(\"numpy\").map(lambda x: x, batched=False)\r\nt0 = time.time()\r\nfor ex in ds:\r\n pass\r\nt1 = time.time()\r\n```\r\nTime: ~0.02 s (~30s on main)\r\n\r\n```python\r\nds = dataset.to_iterable_dataset()\r\nds = ds.with_format(\"numpy\")\r\nt0 = time.time()\r\nfor ex in ds:\r\n pass\r\nt1 = time.time()\r\n```\r\nTime: ~0.02s", "also now working for filter with similar performance improvements:\r\n\r\n```python\r\nfiltered_examples = []\r\nds = dataset.to_iterable_dataset()\r\nds = ds.with_format(\"numpy\").filter(lambda x: [arr.shape[0]==2000 for arr in x[\"array0\"]], batch_size=10, batched=True)\r\nt0 = time.time()\r\nfor ex in ds:\r\n filtered_examples.append(ex)\r\nt1 = time.time()\r\nassert len(filtered_examples) == 25\r\n```\r\n0.01s vs 50s on main\r\n\r\n\r\n```python\r\nfiltered_examples = []\r\nds = dataset.to_iterable_dataset()\r\nds = ds.with_format(\"numpy\").filter(lambda x: x[\"array0\"].shape[0]==2000, batched=False)\r\nt0 = time.time()\r\nfor ex in ds:\r\n filtered_examples.append(ex)\r\nt1 = time.time()\r\nassert len(filtered_examples) == 25\r\n```\r\n0.04s vs 50s on main\r\n", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7207). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "(the distributed tests failing in the CI are unrelated)", "There also appears to be a separate? issue with chaining filter and map bc filter iter_arrow only returns _iter_arrow if arrow formatting is applied (and vv presumably)\r\n\r\nI don't have a good minimal example atm", "> issue with chaining filter and map bc filter iter_arrow only returns _iter_arrow if arrow formatting is applied (and vv presumably)\r\n\r\nMaybe related to this issue ?\r\n\r\n```python\r\nds = Dataset.from_dict({\"a\": range(10)}).to_iterable_dataset()\r\nds = ds.with_format(\"arrow\").map(lambda x: x, features=Features({\"a\": Value(\"string\")})).with_format(None)\r\nprint(list(ds)) # yields integers instead of strings\r\n```", "I feel like we could get rid of TypedExampleIterable altogether and apply formatting with feature conversion with `formatted_python_examples_iterator ` and `formatted_arrow_examples_iterator`\r\n\r\nbtw you can pass `features=` in `get_formatter()` to get a formatter that does the feature conversion at the same time as formatting\r\n\r\n(edit:\r\n\r\nexcept maybe the arrow formatter doesn't use `features` yet, we can fix it like this if it's really needed\r\n```diff\r\nclass ArrowFormatter(Formatter[pa.Table, pa.Array, pa.Table]):\r\n def format_row(self, pa_table: pa.Table) -> pa.Table:\r\n- return self.simple_arrow_extractor().extract_row(pa_table)\r\n+ pa_table = self.simple_arrow_extractor().extract_row(pa_table)\r\n+. return cast_table_to_features(pa_table, self.features) if self.features else pa_table\r\n \r\n```\r\n\r\n\r\n)", "> I feel like we could get rid of TypedExampleIterable altogether and apply formatting with feature conversion with formatted_python_examples_iterator and formatted_arrow_examples_iterator\r\n\r\nOh nice didn't know about the feature support in get_formatter. Haven't thought through whether this works but would a FormattedExampleIterable (with feature conversion) be able to solve this and fit the API better?", "> Oh nice didn't know about the feature support in get_formatter. Haven't thought through whether this works but would a FormattedExampleIterable (with feature conversion) be able to solve this and fit the API better?\r\n\r\nYes this is surely the way to go actually !", "ok i've fixed the chaining issue with my last two commits.\r\n\r\nWill see if I can refactor into a FormattedExampleIterable\r\n\r\nThe other issue you posted seems to be unrelated (maybe something to do with feature decoding?)", "updated with FormattedExamplesIterable.\r\n\r\nthere might be a few unnecessary format calls once the data is already formatted - doesn't seem like a big performance bottleneck but could maybe be fixed with e.g. an is_formatted property\r\n\r\nIt also might be possible to do a wider refactor and use FormattedExamplesIterable elsewhere. But I'd personally prefer not to try that rn.", "Thinking about this in the context of #7210 - am wondering if it would make sense for Features to define their own extraction arrow->object logic? e.g. Arrays should *always* be extracted with NumpyArrowExtractor, not only in case with_format is set to numpy (which a user can easily forget or not know to do)\r\n", "> Thinking about this in the context of https://github.com/huggingface/datasets/issues/7210 - am wondering if it would make sense for Features to define their own extraction arrow->object logic? e.g. Arrays should always be extracted with NumpyArrowExtractor, not only in case with_format is set to numpy (which a user can easily forget or not know to do)\r\n\r\nFor `ArrayND` they already implement `to_pylist` to decode arrow data and it can be updated to return a numpy array (see the `ArrayExtensionArray` class for more details)", "@lhoestq im no longer sure my specific concern about with_format(None) was well-founded - I didn't appreciate that the python formatter tries to do nothing to python objects including numpy arrays, so the existing with_format(None) should I *think* do what I want. Do you think with_format(None) is ok as is after all? If so think this is hopefully ready for final review!", "@lhoestq I've updated to make compatible with latest changes on main, and think the current with_format None behaviour is probably fine - please let me know if there's anything else I can do!", "Hi Alex, I will be less available from today and for a week. I'll review your PR and play with it once I come back if you don't mind !", "thanks for the reviews and extensions, happy to see this merged :)" ]
2,573,567,467
Slow iteration for iterable dataset with numpy formatting for array data
open
### Describe the bug When working with large arrays, setting with_format to e.g. numpy then applying map causes a significant slowdown for iterable datasets. ### Steps to reproduce the bug ```python import numpy as np import time from datasets import Dataset, Features, Array3D features=Features(**{"array0": Array3D((None, 10, 10), dtype="float32"), "array1": Array3D((None,10,10), dtype="float32")}) dataset = Dataset.from_dict({f"array{i}": [np.zeros((x,10,10), dtype=np.float32) for x in [2000,1000]*25] for i in range(2)}, features=features) ``` Then ```python ds = dataset.to_iterable_dataset() ds = ds.with_format("numpy").map(lambda x: x) t0 = time.time() for ex in ds: pass t1 = time.time() print(t1-t0) ``` takes 27 s, whereas ```python ds = dataset.to_iterable_dataset() ds = ds.with_format("numpy") ds = dataset.to_iterable_dataset() t0 = time.time() for ex in ds: pass t1 = time.time() print(t1 - t0) ``` takes ~1s ### Expected behavior Map should not introduce a slowdown when formatting is enabled. ### Environment info 3.0.2
2024-10-08T15:38:11
2024-10-17T17:14:52
null
https://github.com/huggingface/datasets/issues/7206
null
7,206
false
[ "The below easily eats up 32G of RAM. Leaving it for a while bricked the laptop with 16GB.\r\n\r\n```\r\ndataset = load_dataset(\"Voxel51/OxfordFlowers102\", data_dir=\"data\").with_format(\"numpy\")\r\nprocessed_dataset = dataset.map(lambda x: x)\r\n```\r\n\r\n![image](https://github.com/user-attachments/assets/c1863a69-b18f-4014-89dc-98994336df96)\r\n\r\nSimilar problems occur if using a real transform function in `.map()`." ]
2,573,490,859
fix ci benchmark
closed
we're not using the benchmarks anymore + they were not working anyway due to token permissions I keep the code in case we ever want to re-run the benchmark manually
2024-10-08T15:06:18
2024-10-08T15:25:28
2024-10-08T15:25:25
https://github.com/huggingface/datasets/pull/7205
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7205", "html_url": "https://github.com/huggingface/datasets/pull/7205", "diff_url": "https://github.com/huggingface/datasets/pull/7205.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7205.patch", "merged_at": "2024-10-08T15:25:25" }
7,205
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7205). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,573,289,063
fix unbatched arrow map for iterable datasets
closed
Fixes the bug when applying map to an arrow-formatted iterable dataset described here: https://github.com/huggingface/datasets/issues/6833#issuecomment-2399903885 ```python from datasets import load_dataset ds = load_dataset("rotten_tomatoes", split="train", streaming=True) ds = ds.with_format("arrow").map(lambda x: x) for ex in ds: pass ``` @lhoestq
2024-10-08T13:54:09
2024-10-08T14:19:47
2024-10-08T14:19:47
https://github.com/huggingface/datasets/pull/7204
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7204", "html_url": "https://github.com/huggingface/datasets/pull/7204", "diff_url": "https://github.com/huggingface/datasets/pull/7204.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7204.patch", "merged_at": "2024-10-08T14:19:46" }
7,204
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7204). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,573,154,222
with_format docstring
closed
reported at https://github.com/huggingface/datasets/issues/3444
2024-10-08T13:05:19
2024-10-08T13:13:12
2024-10-08T13:13:05
https://github.com/huggingface/datasets/pull/7203
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7203", "html_url": "https://github.com/huggingface/datasets/pull/7203", "diff_url": "https://github.com/huggingface/datasets/pull/7203.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7203.patch", "merged_at": "2024-10-08T13:13:05" }
7,203
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7203). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,572,583,798
`from_parquet` return type annotation
open
### Describe the bug As already posted in https://github.com/microsoft/pylance-release/issues/6534, the correct type hinting fails when building a dataset using the `from_parquet` constructor. Their suggestion is to comprehensively annotate the method's return type to better align with the docstring information. ### Steps to reproduce the bug ```python from datasets import Dataset dataset = Dataset.from_parquet(path_or_paths="file") dataset.map(lambda x: {"new": x["old"]}, batched=True) ``` ### Expected behavior map is a [valid](https://huggingface.co/docs/datasets/v3.0.1/en/package_reference/main_classes#datasets.Dataset.map), no error should be thrown. ### Environment info - `datasets` version: 3.0.1 - Platform: macOS-15.0.1-arm64-arm-64bit - Python version: 3.12.6 - `huggingface_hub` version: 0.25.1 - PyArrow version: 17.0.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.6.1
2024-10-08T09:08:10
2024-10-08T09:08:10
null
https://github.com/huggingface/datasets/issues/7202
null
7,202
false
[]
2,569,837,015
`load_dataset()` of images from a single directory where `train.png` image exists
open
### Describe the bug Hey! Firstly, thanks for maintaining such framework! I had a small issue, where I wanted to load a custom dataset of image+text captioning. I had all of my images in a single directory, and one of the images had the name `train.png`. Then, the loaded dataset had only this image. I guess it's related to "train" as a split name, but it's definitely an unexpected behavior :) Unfortunately I don't have time to submit a proper PR. I'm attaching a toy example to reproduce the issue. Thanks, Sagi ### Steps to reproduce the bug All of the steps I'm attaching are in a fresh env :) ``` (base) sagipolaczek@Sagis-MacBook-Pro ~ % conda activate hf_issue_env (hf_issue_env) sagipolaczek@Sagis-MacBook-Pro ~ % python --version Python 3.10.15 (hf_issue_env) sagipolaczek@Sagis-MacBook-Pro ~ % pip list | grep datasets datasets 3.0.1 (hf_issue_env) sagipolaczek@Sagis-MacBook-Pro ~ % ls -la Documents/hf_datasets_issue total 352 drwxr-xr-x 6 sagipolaczek staff 192 Oct 7 11:59 . drwx------@ 23 sagipolaczek staff 736 Oct 7 11:46 .. -rw-r--r--@ 1 sagipolaczek staff 72 Oct 7 11:59 metadata.csv -rw-r--r--@ 1 sagipolaczek staff 160154 Oct 6 18:00 pika.png -rw-r--r--@ 1 sagipolaczek staff 5495 Oct 6 12:02 pika_pika.png -rw-r--r--@ 1 sagipolaczek staff 1753 Oct 6 11:50 train.png (hf_issue_env) sagipolaczek@Sagis-MacBook-Pro ~ % cat Documents/hf_datasets_issue/metadata.csv file_name,text train.png,A train pika.png,Pika pika_pika.png,Pika Pika! (hf_issue_env) sagipolaczek@Sagis-MacBook-Pro ~ % python Python 3.10.15 (main, Oct 3 2024, 02:33:33) [Clang 14.0.6 ] on darwin Type "help", "copyright", "credits" or "license" for more information. >>> from datasets import load_dataset >>> dataset = load_dataset("imagefolder", data_dir="Documents/hf_datasets_issue/") >>> dataset DatasetDict({ train: Dataset({ features: ['image', 'text'], num_rows: 1 }) }) >>> dataset["train"][0] {'image': <PIL.PngImagePlugin.PngImageFile image mode=RGBA size=354x84 at 0x10B50FD90>, 'text': 'A train'} ### DELETING `train.png` sample ### (hf_issue_env) sagipolaczek@Sagis-MacBook-Pro ~ % vim Documents/hf_datasets_issue/metadata.csv (hf_issue_env) sagipolaczek@Sagis-MacBook-Pro ~ % rm Documents/hf_datasets_issue/train.png (hf_issue_env) sagipolaczek@Sagis-MacBook-Pro ~ % python Python 3.10.15 (main, Oct 3 2024, 02:33:33) [Clang 14.0.6 ] on darwin Type "help", "copyright", "credits" or "license" for more information. >>> from datasets import load_dataset >>> dataset = load_dataset("imagefolder", data_dir="Documents/hf_datasets_issue/") Generating train split: 2 examples [00:00, 65.99 examples/s] >>> dataset DatasetDict({ train: Dataset({ features: ['image', 'text'], num_rows: 2 }) }) >>> dataset["train"] Dataset({ features: ['image', 'text'], num_rows: 2 }) >>> dataset["train"][0],dataset["train"][1] ({'image': <PIL.PngImagePlugin.PngImageFile image mode=RGBA size=2356x1054 at 0x10DD11E70>, 'text': 'Pika'}, {'image': <PIL.PngImagePlugin.PngImageFile image mode=RGBA size=343x154 at 0x10E258C70>, 'text': 'Pika Pika!'}) ``` ### Expected behavior My expected behavior would be to get a dataset with the sample `train.png` in it (along with the others data points). ### Environment info I've attached it in the example: Python 3.10.15 datasets 3.0.1
2024-10-07T09:14:17
2024-10-07T09:14:17
null
https://github.com/huggingface/datasets/issues/7201
null
7,201
false
[]
2,567,921,694
Fix the environment variable for huggingface cache
closed
Resolve #6256. As far as I tested, `HF_DATASETS_CACHE` was ignored and I could not specify the cache directory at all except for the default one by this environment variable. `HF_HOME` has worked. Perhaps the recent change on file downloading by `huggingface_hub` could affect this bug. In my testing, I could not specify the cache directory even by `load_dataset("dataset_name" cache_dir="...")`. It might be another issue. I also welcome any advice to solve this issue.
2024-10-05T11:54:35
2024-10-30T23:10:27
2024-10-08T15:45:18
https://github.com/huggingface/datasets/pull/7200
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7200", "html_url": "https://github.com/huggingface/datasets/pull/7200", "diff_url": "https://github.com/huggingface/datasets/pull/7200.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7200.patch", "merged_at": "2024-10-08T15:45:17" }
7,200
true
[ "Hi ! yes now `datasets` uses `huggingface_hub` to download and cache files from the HF Hub so you need to use `HF_HOME` (or manually `HF_HUB_CACHE` and `HF_DATASETS_CACHE` if you want to separate HF Hub cached files and cached datasets Arrow files)\r\n\r\nSo in your change I guess it needs to be `HF_HOME` instead of `HF_CACHE` ?", "Thank you for your comment. You are right. I am sorry for my mistake, I fixed it.", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7200). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "I just had this issue, and needed to move the setting the env code in the python file to top, before the import of the lib \r\nie. \r\n```python\r\nimport os\r\nLOCAL_DISK_MOUNT = '/mnt/data'\r\n\r\nos.environ['HF_HOME'] = f'{LOCAL_DISK_MOUNT}/hf_cache/'\r\nos.environ['HF_DATASETS_CACHE'] = f'{LOCAL_DISK_MOUNT}/datasets/'\r\n\r\nfrom datasets import load_dataset\r\nfrom datasets import load_dataset_builder\r\nfrom psutil._common import bytes2human\r\n\r\n\r\n```" ]
2,566,788,225
Add with_rank to Dataset.from_generator
open
Adds `with_rank` to `Dataset.from_generator`. As for `Dataset.map` and `Dataset.filter`, this is useful when creating cache files using multi-GPU.
2024-10-04T16:51:53
2024-10-04T16:51:53
null
https://github.com/huggingface/datasets/pull/7199
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7199", "html_url": "https://github.com/huggingface/datasets/pull/7199", "diff_url": "https://github.com/huggingface/datasets/pull/7199.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7199.patch", "merged_at": null }
7,199
true
[]
2,566,064,849
Add repeat method to datasets
closed
Following up on discussion in #6623 and #7198 I thought this would be pretty useful for my case so had a go at implementing. My main motivation is to be able to call iterable_dataset.repeat(None).take(samples_per_epoch) to safely avoid timeout issues in a distributed training setting. This would provide a straightforward workaround for several open issues related to this situation: https://github.com/huggingface/datasets/issues/6437, https://github.com/huggingface/datasets/issues/6594, https://github.com/huggingface/datasets/issues/6623, https://github.com/huggingface/datasets/issues/6719. @lhoestq let me know if this looks on the right track!
2024-10-04T10:45:16
2025-02-05T16:32:31
2025-02-05T16:32:31
https://github.com/huggingface/datasets/pull/7198
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7198", "html_url": "https://github.com/huggingface/datasets/pull/7198", "diff_url": "https://github.com/huggingface/datasets/pull/7198.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7198.patch", "merged_at": "2025-02-05T16:32:31" }
7,198
true
[ "@lhoestq does this look reasonable?", "updated and added test cases!", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7198). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "thanks for the fixes!" ]
2,565,924,788
ConnectionError: Couldn't reach 'allenai/c4' on the Hub (ConnectionError)数据集下不下来,怎么回事
open
### Describe the bug from datasets import load_dataset print("11") traindata = load_dataset('ptb_text_only', 'penn_treebank', split='train') print("22") valdata = load_dataset('ptb_text_only', 'penn_treebank', split='validation') ### Steps to reproduce the bug 1 ### Expected behavior 1 ### Environment info 1
2024-10-04T09:33:25
2025-02-26T02:26:16
null
https://github.com/huggingface/datasets/issues/7197
null
7,197
false
[ "Also cant download \"allenai/c4\", but with different error reported:\r\n```\r\nTraceback (most recent call last): \r\n File \"/***/lib/python3.10/site-packages/datasets/load.py\", line 2074, in load_dataset \r\n builder_instance = load_dataset_builder( \r\n File \"/***/lib/python3.10/site-packages/datasets/load.py\", line 1795, in load_dataset_builder \r\n dataset_module = dataset_module_factory( \r\n File \"/***/lib/python3.10/site-packages/datasets/load.py\", line 1659, in dataset_module_factory \r\n raise e1 from None \r\n File \"/***/lib/python3.10/site-packages/datasets/load.py\", line 1647, in dataset_module_factory \r\n ).get_module() \r\n File \"/***/lib/python3.10/site-packages/datasets/load.py\", line 1069, in get_module \r\n module_name, default_builder_kwargs = infer_module_for_data_files( \r\n File \"/***/lib/python3.10/site-packages/datasets/load.py\", line 594, in infer_module_for_data_files \r\n raise DataFilesNotFoundError(\"No (supported) data files found\" + (f\" in {path}\" if path else \"\")) \r\ndatasets.exceptions.DataFilesNotFoundError: No (supported) data files found in allenai/c4 \r\n```\r\n\r\n## Code to reproduce\r\n```\r\ndataset = load_dataset(\"allenai/c4\", \"en\", split=\"train\", streaming=True,trust_remote_code=True,\r\n cache_dir=\"dataset/en\",\r\n download_mode=\"force_redownload\")\r\n```\r\n\r\n## Environment\r\ndatasets 3.0.1 \r\nhuggingface_hub 0.25.1", "应该是网络问题,无法访问外网?" ]
2,564,218,566
concatenate_datasets does not preserve shuffling state
open
### Describe the bug After concatenate datasets on an iterable dataset, the shuffling state is destroyed, similar to #7156 This means concatenation cant be used for resolving uneven numbers of samples across devices when using iterable datasets in a distributed setting as discussed in #6623 I also noticed that the number of shards is the same after concatenation, which I found surprising, but I don't understand the internals well enough to know whether this is actually surprising or not ### Steps to reproduce the bug ```python import datasets import torch.utils.data def gen(shards): yield {"shards": shards} def main(): dataset1 = datasets.IterableDataset.from_generator( gen, gen_kwargs={"shards": list(range(25))} # TODO: how to understand this? ) dataset2 = datasets.IterableDataset.from_generator( gen, gen_kwargs={"shards": list(range(25, 50))} # TODO: how to understand this? ) dataset1 = dataset1.shuffle(buffer_size=1) dataset2 = dataset2.shuffle(buffer_size=1) print(dataset1.n_shards) print(dataset2.n_shards) dataset = datasets.concatenate_datasets( [dataset1, dataset2] ) print(dataset.n_shards) # dataset = dataset1 dataloader = torch.utils.data.DataLoader( dataset, batch_size=8, num_workers=0, ) for i, batch in enumerate(dataloader): print(batch) print("\nNew epoch") dataset = dataset.set_epoch(1) for i, batch in enumerate(dataloader): print(batch) if __name__ == "__main__": main() ``` ### Expected behavior Shuffling state should be preserved ### Environment info Latest datasets
2024-10-03T14:30:38
2025-03-18T10:56:47
null
https://github.com/huggingface/datasets/issues/7196
null
7,196
false
[ "It also does preserve `split_by_node`, so in the meantime you should call `shuffle` or `split_by_node` AFTER `interleave_datasets` or `concatenate_datasets`" ]
2,564,070,809
Add support for 3D datasets
open
See https://huggingface.co/datasets/allenai/objaverse for example
2024-10-03T13:27:44
2024-10-04T09:23:36
null
https://github.com/huggingface/datasets/issues/7195
null
7,195
false
[ "maybe related: https://github.com/huggingface/datasets/issues/6388", "Also look at https://github.com/huggingface/dataset-viewer/blob/f5fd117ceded990a7766e705bba1203fa907d6ad/services/worker/src/worker/job_runners/dataset/modalities.py#L241 which lists the 3D file formats that will assign the 3D modality to a dataset.", "~~we can brainstorm about the UX maybe (i don't expect we should load all models on the page at once – IMO there should be a manual action from user to load + maybe load first couple of row by default) cc @gary149 @cfahlgren1~~\r\n\r\nit's more for the viewer issue (https://github.com/huggingface/dataset-viewer/issues/1003)" ]
2,563,364,199
datasets.exceptions.DatasetNotFoundError for private dataset
closed
### Describe the bug The following Python code tries to download a private dataset and fails with the error `datasets.exceptions.DatasetNotFoundError: Dataset 'ClimatePolicyRadar/all-document-text-data-weekly' doesn't exist on the Hub or cannot be accessed.`. Downloading a public dataset doesn't work. ``` py from datasets import load_dataset _ = load_dataset("ClimatePolicyRadar/all-document-text-data-weekly") ``` This seems to be just an issue with my machine config as the code above works with a colleague's machine. So far I have tried: - logging back out and in from the Huggingface CLI using `huggingface-cli logout` - manually removing the token cache at `/Users/kalyan/.cache/huggingface/token` (found using `huggingface-cli env`) - manually passing a token in `load_dataset` My output of `huggingface-cli whoami`: ``` kdutia orgs: ClimatePolicyRadar ``` ### Steps to reproduce the bug ``` python Python 3.12.2 (main, Feb 6 2024, 20:19:44) [Clang 15.0.0 (clang-1500.1.0.2.5)] on darwin Type "help", "copyright", "credits" or "license" for more information. >>> from datasets import load_dataset >>> _ = load_dataset("ClimatePolicyRadar/all-document-text-data-weekly") Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/Users/kalyan/Library/Caches/pypoetry/virtualenvs/open-data-cnKQNmjn-py3.12/lib/python3.12/site-packages/datasets/load.py", line 2074, in load_dataset builder_instance = load_dataset_builder( ^^^^^^^^^^^^^^^^^^^^^ File "/Users/kalyan/Library/Caches/pypoetry/virtualenvs/open-data-cnKQNmjn-py3.12/lib/python3.12/site-packages/datasets/load.py", line 1795, in load_dataset_builder dataset_module = dataset_module_factory( ^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/kalyan/Library/Caches/pypoetry/virtualenvs/open-data-cnKQNmjn-py3.12/lib/python3.12/site-packages/datasets/load.py", line 1659, in dataset_module_factory raise e1 from None File "/Users/kalyan/Library/Caches/pypoetry/virtualenvs/open-data-cnKQNmjn-py3.12/lib/python3.12/site-packages/datasets/load.py", line 1597, in dataset_module_factory raise DatasetNotFoundError(f"Dataset '{path}' doesn't exist on the Hub or cannot be accessed.") from e datasets.exceptions.DatasetNotFoundError: Dataset 'ClimatePolicyRadar/all-document-text-data-weekly' doesn't exist on the Hub or cannot be accessed. >>> ``` ### Expected behavior The dataset downloads successfully. ### Environment info From `huggingface-cli env`: ``` - huggingface_hub version: 0.25.1 - Platform: macOS-14.2.1-arm64-arm-64bit - Python version: 3.12.2 - Running in iPython ?: No - Running in notebook ?: No - Running in Google Colab ?: No - Running in Google Colab Enterprise ?: No - Token path ?: /Users/kalyan/.cache/huggingface/token - Has saved token ?: True - Who am I ?: kdutia - Configured git credential helpers: osxkeychain - FastAI: N/A - Tensorflow: N/A - Torch: N/A - Jinja2: 3.1.4 - Graphviz: N/A - keras: N/A - Pydot: N/A - Pillow: N/A - hf_transfer: N/A - gradio: N/A - tensorboard: N/A - numpy: 2.1.1 - pydantic: N/A - aiohttp: 3.10.8 - ENDPOINT: https://huggingface.co - HF_HUB_CACHE: /Users/kalyan/.cache/huggingface/hub - HF_ASSETS_CACHE: /Users/kalyan/.cache/huggingface/assets - HF_TOKEN_PATH: /Users/kalyan/.cache/huggingface/token - HF_HUB_OFFLINE: False - HF_HUB_DISABLE_TELEMETRY: False - HF_HUB_DISABLE_PROGRESS_BARS: None - HF_HUB_DISABLE_SYMLINKS_WARNING: False - HF_HUB_DISABLE_EXPERIMENTAL_WARNING: False - HF_HUB_DISABLE_IMPLICIT_TOKEN: False - HF_HUB_ENABLE_HF_TRANSFER: False - HF_HUB_ETAG_TIMEOUT: 10 - HF_HUB_DOWNLOAD_TIMEOUT: 10 ``` from `datasets-cli env`: ``` - `datasets` version: 3.0.1 - Platform: macOS-14.2.1-arm64-arm-64bit - Python version: 3.12.2 - `huggingface_hub` version: 0.25.1 - PyArrow version: 17.0.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.6.1 ```
2024-10-03T07:49:36
2024-10-03T10:09:28
2024-10-03T10:09:28
https://github.com/huggingface/datasets/issues/7194
null
7,194
false
[ "Actually there is no such dataset available, that is why you are getting that error.", "Fixed with @kdutia in Slack chat. Generating a new token fixed this issue. " ]
2,562,392,887
Support of num_workers (multiprocessing) in map for IterableDataset
open
### Feature request Currently, IterableDataset doesn't support setting num_worker in .map(), which results in slow processing here. Could we add support for it? As .map() can be run in the batch fashion (e.g., batch_size is default to 1000 in datasets), it seems to be doable for IterableDataset as the regular Dataset. ### Motivation Improving data processing efficiency ### Your contribution Testing
2024-10-02T18:34:04
2024-10-03T09:54:15
null
https://github.com/huggingface/datasets/issues/7193
null
7,193
false
[ "I was curious about the same - since map is applied on the fly I was assuming that setting num_workers>1 in DataLoader would effectively do the map in parallel, have you tried that?" ]
2,562,289,642
Add repeat() for iterable datasets
closed
### Feature request It would be useful to be able to straightforwardly repeat iterable datasets indefinitely, to provide complete control over starting and ending of iteration to the user. An IterableDataset.repeat(n) function could do this automatically ### Motivation This feature was discussed in this issue https://github.com/huggingface/datasets/issues/7147, and would resolve the need to use the hack of interleave datasets with probability 0 as a simple way to achieve this functionality. An additional benefit might be the simplification of the use of iterable datasets in a distributed setting: If the user can assume that datasets will repeat indefinitely, then issues around different numbers of samples appearing on different devices (e.g. https://github.com/huggingface/datasets/issues/6437, https://github.com/huggingface/datasets/issues/6594, https://github.com/huggingface/datasets/issues/6623, https://github.com/huggingface/datasets/issues/6719) can potentially be straightforwardly resolved by simply doing: ids.repeat(None).take(n_samples_per_epoch) ### Your contribution I'm not familiar enough with the codebase to assess how straightforward this would be to implement. If it might be very straightforward, I could possibly have a go.
2024-10-02T17:48:13
2025-03-18T10:48:33
2025-03-18T10:48:32
https://github.com/huggingface/datasets/issues/7192
null
7,192
false
[ "perhaps concatenate_datasets can already be used to achieve almost the same effect? ", "`concatenate_datasets` does the job when there is a finite number of repetitions, but in case of `.repeat()` forever we need a new logic in `iterable_dataset.py`", "done in https://github.com/huggingface/datasets/pull/7198" ]
2,562,206,949
Solution to issue: #7080 Modified load_dataset function, so that it prompts the user to select a dataset when subdatasets or splits (train, test) are available
closed
# Feel free to give suggestions please.. ### This PR is raised because of issue: https://github.com/huggingface/datasets/issues/7080 ![image](https://github.com/user-attachments/assets/8fbc604f-f0a5-4a59-a63e-aa4c26442c83) ### This PR gives solution to https://github.com/huggingface/datasets/issues/7080 1. Checking whether the dataset has splits or subdatasets. 2. Printing the available splits/subdatasets. 3. Asking the user to choose which one to load. 4. Loading only the selected dataset based on the user's input. ### Key Changes: 1. Available Splits/Subdatasets: The code checks for available splits/subdatasets using builder_instance.info.splits.keys(). 2. User Prompt: If splits are found, it prints them out and prompts the user to select one. 3. Loading Based on User Input: The dataset is loaded based on the user's choice. This way, the dataset loading function will interactively prompt the user to select which subdataset or split they want to load instead of automatically loading all of them.
2024-10-02T17:02:45
2024-11-10T08:48:21
2024-11-10T08:48:21
https://github.com/huggingface/datasets/pull/7191
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7191", "html_url": "https://github.com/huggingface/datasets/pull/7191", "diff_url": "https://github.com/huggingface/datasets/pull/7191.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7191.patch", "merged_at": null }
7,191
true
[ "I think the approach presented in https://github.com/huggingface/datasets/pull/6832 is the one we'll be taking.\r\n\r\nAsking user input is not a good idea since `load_dataset` is used a lot in server that don't have someone in front of them to select a split" ]
2,562,162,725
Datasets conflicts with fsspec 2024.9
open
### Describe the bug Installing both in latest versions are not possible `pip install "datasets==3.0.1" "fsspec==2024.9.0"` But using older version of datasets is ok `pip install "datasets==1.24.4" "fsspec==2024.9.0"` ### Steps to reproduce the bug `pip install "datasets==3.0.1" "fsspec==2024.9.0"` ### Expected behavior install both versions. ### Environment info debian 11. python 3.10.15
2024-10-02T16:43:46
2024-10-10T07:33:18
null
https://github.com/huggingface/datasets/issues/7190
null
7,190
false
[ "Yes, I need to use the latest version of fsspec and datasets for my usecase. \r\nhttps://github.com/fsspec/s3fs/pull/888#issuecomment-2404204606\r\nhttps://github.com/apache/arrow/issues/34363#issuecomment-2403553473\r\n\r\nlast version where things install without conflict is: 2.14.4\r\n\r\nSo this issue starts from:\r\nhttps://github.com/huggingface/datasets/releases/tag/2.14.5" ]
2,562,152,845
Audio preview in dataset viewer for audio array data without a path/filename
open
### Feature request Huggingface has quite a comprehensive set of guides for [audio datasets](https://huggingface.co/docs/datasets/en/audio_dataset). It seems, however, all these guides assume the audio array data to be decoded/inserted into a HF dataset always originates from individual files. The [Audio-dataclass](https://github.com/huggingface/datasets/blob/3.0.1/src/datasets/features/audio.py#L20) appears designed with this assumption in mind. Looking at its source code it returns a dictionary with the keys `path`, `array` and `sampling_rate`. However, sometimes users may have different pipelines where they themselves decode the audio array. This feature request has to do with wishing some clarification in guides on whether it is possible, and in such case how users can insert already decoded audio array data into datasets (pandas DataFrame, HF dataset or whatever) that are later saved as parquet, and still get a functioning audio preview in the dataset viewer. Do I perhaps need to write a tempfile of my audio array slice to wav and capture the bytes object with `io.BytesIO` and pass that to `Audio()`? ### Motivation I'm working with large audio datasets, and my pipeline reads (decodes) audio from larger files, and slices the relevant portions of audio from that larger file based on metadata I have available. The pipeline is designed this way to avoid having to store multiple copies of data, and to avoid having to store tens of millions of small files. I tried [test-uploading parquet files](https://huggingface.co/datasets/Lauler/riksdagen_test) where I store the audio array data of decoded slices of audio in an `audio` column with a dictionary with the keys `path`, `array` and `sampling_rate`. But I don't know the secret sauce of what the Huggingface Hub expects and requires to be able to display audio previews correctly. ### Your contribution I could contribute a tool agnostic guide of creating HF audio datasets directly as parquet to the HF documentation if there is an interest. Provided you help me figure out the secret sauce of what the dataset viewer expects to display the preview correctly.
2024-10-02T16:38:38
2024-10-02T17:01:40
null
https://github.com/huggingface/datasets/issues/7189
null
7,189
false
[]
2,560,712,689
Pin multiprocess<0.70.1 to align with dill<0.3.9
closed
Pin multiprocess<0.70.1 to align with dill<0.3.9. Note that multiprocess-0.70.1 requires dill-0.3.9: https://github.com/uqfoundation/multiprocess/releases/tag/0.70.17 Fix #7186.
2024-10-02T05:40:18
2024-10-02T06:08:25
2024-10-02T06:08:23
https://github.com/huggingface/datasets/pull/7188
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7188", "html_url": "https://github.com/huggingface/datasets/pull/7188", "diff_url": "https://github.com/huggingface/datasets/pull/7188.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7188.patch", "merged_at": "2024-10-02T06:08:23" }
7,188
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7188). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,560,501,308
shard_data_sources() got an unexpected keyword argument 'worker_id'
open
### Describe the bug ``` [rank0]: File "/home/qinghao/miniconda3/envs/doremi/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 238, in __iter__ [rank0]: for key_example in islice(self.generate_examples_fn(**gen_kwags), shard_example_idx_start, None): [rank0]: File "/home/qinghao/miniconda3/envs/doremi/lib/python3.10/site-packages/datasets/packaged_modules/generator/generator.py", line 32, in _generate_examples [rank0]: for idx, ex in enumerate(self.config.generator(**gen_kwargs)): [rank0]: File "/home/qinghao/workdir/doremi/doremi/dataloader.py", line 337, in take_data_generator [rank0]: for ex in ds: [rank0]: File "/home/qinghao/miniconda3/envs/doremi/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1791, in __iter__ [rank0]: yield from self._iter_pytorch() [rank0]: File "/home/qinghao/miniconda3/envs/doremi/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1704, in _iter_pytorch [rank0]: ex_iterable = ex_iterable.shard_data_sources(worker_id=worker_info.id, num_workers=worker_info.num_workers) [rank0]: TypeError: UpdatableRandomlyCyclingMultiSourcesExamplesIterable.shard_data_sources() got an unexpected keyword argument 'worker_id' ``` ### Steps to reproduce the bug IterableDataset cannot use ### Expected behavior can work on datasets==2.10, but will raise error for later versions. ### Environment info datasets==3.0.1
2024-10-02T01:26:35
2024-10-02T01:26:35
null
https://github.com/huggingface/datasets/issues/7187
null
7,187
false
[]
2,560,323,917
pinning `dill<0.3.9` without pinning `multiprocess`
closed
### Describe the bug The [latest `multiprocess` release](https://github.com/uqfoundation/multiprocess/releases/tag/0.70.17) requires `dill>=0.3.9` which causes issues when installing `datasets` without backtracking during package version resolution. Is it possible to add a pin for multiprocess so something like `multiprocess<=0.70.16` so that the `dill` version is compatible? ### Steps to reproduce the bug NA ### Expected behavior NA ### Environment info NA
2024-10-01T22:29:32
2024-10-02T06:08:24
2024-10-02T06:08:24
https://github.com/huggingface/datasets/issues/7186
null
7,186
false
[]
2,558,508,748
CI benchmarks are broken
closed
Since Aug 30, 2024, CI benchmarks are broken: https://github.com/huggingface/datasets/actions/runs/11108421214/job/30861323975 ``` {"level":"error","message":"Resource not accessible by integration","name":"HttpError","request":{"body":"{\"body\":\"<details>\\n<summary>Show benchmarks</summary>\\n\\nPyArrow==8.0.0\\n\\n<details>\\n<summary>Show updated benchmarks!</summary>\\n\\n### Benchmark: benchmark_array_xd.json\\n\\n| metric | read_batch_formatted_as_numpy after write_array2d | ... "headers":{"accept":"application/vnd.github.v3+json","authorization":"token [REDACTED]","content-type":"application/json; charset=utf-8","user-agent":"octokit-rest.js/18.0.0 octokit-core.js/3.6.0 Node.js/16.20.2 (linux; x64)"},"method":"POST","request":{"agent":{"_events":{},"_eventsCount":2,"cache": ... "response":{"data":{"documentation_url":"https://docs.github.com/rest/issues/comments#create-an-issue-comment","message":"Resource not accessible by integration","status":"403"}, ... "stack":"HttpError: Resource not accessible by integration\n at /usr/lib/node_modules/@dvcorg/cml/node_modules/@octokit/request/dist-node/index.js:86:21\n at processTicksAndRejections (node:internal/process/task_queues:96:5)\n at async Job.doExecute (/usr/lib/node_modules/@dvcorg/cml/node_modules/bottleneck/light.js:405:18)","status":403} ```
2024-10-01T08:16:08
2024-10-09T16:07:48
2024-10-09T16:07:48
https://github.com/huggingface/datasets/issues/7185
null
7,185
false
[ "Fixed by #7205" ]
2,556,855,150
Pin dill<0.3.9 to fix CI
closed
Pin dill<0.3.9 to fix CI for deps-latest. Note that dill-0.3.9 was released yesterday Sep 29, 2024: - https://pypi.org/project/dill/0.3.9/ - https://github.com/uqfoundation/dill/releases/tag/0.3.9 Fix #7183.
2024-09-30T14:26:25
2024-09-30T14:38:59
2024-09-30T14:38:57
https://github.com/huggingface/datasets/pull/7184
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7184", "html_url": "https://github.com/huggingface/datasets/pull/7184", "diff_url": "https://github.com/huggingface/datasets/pull/7184.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7184.patch", "merged_at": "2024-09-30T14:38:57" }
7,184
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7184). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,556,789,055
CI is broken for deps-latest
closed
See: https://github.com/huggingface/datasets/actions/runs/11106149906/job/30853879890 ``` =========================== short test summary info ============================ FAILED tests/test_arrow_dataset.py::BaseDatasetTest::test_filter_caching_on_disk - AssertionError: Lists differ: [{'fi[44 chars] {'filename': '/tmp/tmp6xcyyjs4/cache-9533fe2601cd3e48.arrow'}] != [{'fi[44 chars] {'filename': '/tmp/tmp6xcyyjs4/cache-e6e0a8b830976289.arrow'}] First differing element 1: {'filename': '/tmp/tmp6xcyyjs4/cache-9533fe2601cd3e48.arrow'} {'filename': '/tmp/tmp6xcyyjs4/cache-e6e0a8b830976289.arrow'} [{'filename': '/tmp/tmp6xcyyjs4/dataset0.arrow'}, - {'filename': '/tmp/tmp6xcyyjs4/cache-9533fe2601cd3e48.arrow'}] ? ^^^^^ -------- + {'filename': '/tmp/tmp6xcyyjs4/cache-e6e0a8b830976289.arrow'}] ? ++++++++++ ^^ + FAILED tests/test_arrow_dataset.py::BaseDatasetTest::test_map_caching_on_disk - AssertionError: Lists differ: [{'filename': '/tmp/tmp5gxrti_n/cache-e58d327daec8626f.arrow'}] != [{'filename': '/tmp/tmp5gxrti_n/cache-d87234c5763e54a3.arrow'}] First differing element 0: {'filename': '/tmp/tmp5gxrti_n/cache-e58d327daec8626f.arrow'} {'filename': '/tmp/tmp5gxrti_n/cache-d87234c5763e54a3.arrow'} - [{'filename': '/tmp/tmp5gxrti_n/cache-e58d327daec8626f.arrow'}] ? ^^ ----------- + [{'filename': '/tmp/tmp5gxrti_n/cache-d87234c5763e54a3.arrow'}] ? +++++++++++ ^^ FAILED tests/test_fingerprint.py::TokenizersHashTest::test_hash_regex - NameError: name 'log' is not defined FAILED tests/test_fingerprint.py::RecurseHashTest::test_hash_ignores_line_definition_of_function - AssertionError: '52e56ee04ad92499' != '0a4f75cec280f634' - 52e56ee04ad92499 + 0a4f75cec280f634 FAILED tests/test_fingerprint.py::RecurseHashTest::test_hash_ipython_function - AssertionError: 'a6bd2041ca63d6c0' != '517bf36b7eecdef5' - a6bd2041ca63d6c0 + 517bf36b7eecdef5 FAILED tests/test_fingerprint.py::HashingTest::test_hash_tiktoken_encoding - NameError: name 'log' is not defined FAILED tests/test_fingerprint.py::HashingTest::test_hash_torch_compiled_module - NameError: name 'log' is not defined FAILED tests/test_fingerprint.py::HashingTest::test_hash_torch_generator - NameError: name 'log' is not defined FAILED tests/test_fingerprint.py::HashingTest::test_hash_torch_tensor - NameError: name 'log' is not defined FAILED tests/test_fingerprint.py::HashingTest::test_set_doesnt_depend_on_order - NameError: name 'log' is not defined FAILED tests/test_fingerprint.py::HashingTest::test_set_stable - NameError: name 'log' is not defined ERROR tests/test_iterable_dataset.py::test_iterable_dataset_from_file - NameError: name 'log' is not defined = 11 failed, 2850 passed, 3 skipped, 23 warnings, 1 error in 191.06s (0:03:11) = ```
2024-09-30T14:02:07
2024-09-30T14:38:58
2024-09-30T14:38:58
https://github.com/huggingface/datasets/issues/7183
null
7,183
false
[]
2,556,333,671
Support features in metadata configs
closed
Support features in metadata configs, like: ``` configs: - config_name: default features: - name: id dtype: int64 - name: name dtype: string - name: score dtype: float64 ``` This will allow to avoid inference of data types. Currently, we allow passing this information in the `dataset_info` (instead of `configs`) field, but this is not intuitive and it is not properly documented. TODO: - [ ] Document usage
2024-09-30T11:14:53
2024-10-09T16:03:57
2024-10-09T16:03:54
https://github.com/huggingface/datasets/pull/7182
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7182", "html_url": "https://github.com/huggingface/datasets/pull/7182", "diff_url": "https://github.com/huggingface/datasets/pull/7182.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7182.patch", "merged_at": "2024-10-09T16:03:54" }
7,182
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7182). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "The CI issue is unrelated:\r\n- #7183" ]
2,554,917,019
Fix datasets export to JSON
closed
null
2024-09-29T12:45:20
2024-11-01T11:55:36
2024-11-01T11:55:36
https://github.com/huggingface/datasets/pull/7181
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7181", "html_url": "https://github.com/huggingface/datasets/pull/7181", "diff_url": "https://github.com/huggingface/datasets/pull/7181.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7181.patch", "merged_at": null }
7,181
true
[ "Linked Issue: #7037\r\nIdeas: #7039 ", "@albertvillanova / @lhoestq any early feedback?\r\n\r\nAFAIK there is no param `orient` in `load_dataset()`. So for orientations other than \"records\", the loading isn't very accurate. Any thoughts?", "`orient = \"split\"` can also be handled. I will add the changes soon", "Thanks for diving into this ! I don't think we want the JSON export to be that complex though, especially if people can do `ds.to_pandas().to_json(orient=...)`. Maybe we can just raise an error and suggest users to use pandas ? And also note that it loads the full dataset in memory so it's mainly for small scale datasets. The only acceptable option for large scale datasets is probably just JSON Lines anyway since it enables streaming.", "@lhoestq Simply doing `ds.to_pandas().to_json(orient=...)` is not going to give any batching or multiprocessing benefits right? Also, which function are you referring to - when you say that its meant for small scale datasets only?", "Yes indeed. Though I think it's fine since using something else than orient=\"lines\" is only suitable/useful for small datasets. Or you know a case where a big dataset need to be in a format that is not orient=\"lines\" ?", "@lhoestq Let me close this PR and open another one where I will add an error message, as suggested here.\r\n\r\n> Thanks for diving into this ! I don't think we want the JSON export to be that complex though, especially if people can do `ds.to_pandas().to_json(orient=...)`. Maybe we can just raise an error and suggest users to use pandas ? And also note that it loads the full dataset in memory so it's mainly for small scale datasets. The only acceptable option for large scale datasets is probably just JSON Lines anyway since it enables streaming.\r\n\r\n", "Addressed here: #7273 \r\n@lhoestq " ]
2,554,244,750
Memory leak when wrapping datasets into PyTorch Dataset without explicit deletion
closed
### Describe the bug I've encountered a memory leak when wrapping the HuggingFace dataset into a PyTorch Dataset. The RAM usage constantly increases during iteration if items are not explicitly deleted after use. ### Steps to reproduce the bug Steps to reproduce: Create a PyTorch Dataset wrapper for 'nebula/cc12m': ```` from torch.utils.data import Dataset from tqdm import tqdm from datasets import load_dataset from torchvision import transforms Image.MAX_IMAGE_PIXELS = None class CC12M(Dataset): def __init__(self, path_or_name='nebula/cc12m', split='train', transform=None, single_caption=True): self.raw_dataset = load_dataset(path_or_name)[split] if transform is None: self.transform = transforms.Compose([ transforms.Resize((224, 224)), transforms.CenterCrop(224), transforms.ToTensor(), transforms.Normalize( mean=[0.48145466, 0.4578275, 0.40821073], std=[0.26862954, 0.26130258, 0.27577711] ) ]) else: self.transform = transforms.Compose(transform) self.single_caption = single_caption self.length = len(self.raw_dataset) def __len__(self): return self.length def __getitem__(self, index): item = self.raw_dataset[index] caption = item['txt'] with io.BytesIO(item['webp']) as buffer: image = Image.open(buffer).convert('RGB') if self.transform: image = self.transform(image) # del item # Uncomment this line to prevent the memory leak return image, caption ```` Iterate through the dataset without the del item line in __getitem__. Observe RAM usage increasing constantly. Add del item at the end of __getitem__: ``` def __getitem__(self, index): item = self.raw_dataset[index] caption = item['txt'] with io.BytesIO(item['webp']) as buffer: image = Image.open(buffer).convert('RGB') if self.transform: image = self.transform(image) del item # This line prevents the memory leak return image, caption ``` Iterate through the dataset again and observe that RAM usage remains stable. ### Expected behavior Expected behavior: RAM usage should remain stable during iteration without needing to explicitly delete items. Actual behavior: RAM usage constantly increases unless items are explicitly deleted after use ### Environment info - `datasets` version: 2.21.0 - Platform: Linux-4.18.0-513.5.1.el8_9.x86_64-x86_64-with-glibc2.28 - Python version: 3.12.4 - `huggingface_hub` version: 0.24.6 - PyArrow version: 17.0.0 - Pandas version: 2.2.2 - `fsspec` version: 2024.6.1
2024-09-28T14:00:47
2024-09-30T12:07:56
2024-09-30T12:07:56
https://github.com/huggingface/datasets/issues/7180
null
7,180
false
[ "> I've encountered a memory leak when wrapping the HuggingFace dataset into a PyTorch Dataset. The RAM usage constantly increases during iteration if items are not explicitly deleted after use.\r\n\r\nDatasets are memory mapped so they work like SWAP memory. In particular as long as you have RAM available the data will stay in RAM, and get paged out once your system needs RAM for something else (no OOM).\r\n\r\nrelated: https://github.com/huggingface/datasets/issues/4883" ]
2,552,387,980
Support Python 3.11
closed
Support Python 3.11. Fix #7178.
2024-09-27T08:55:44
2024-10-08T16:21:06
2024-10-08T16:21:03
https://github.com/huggingface/datasets/pull/7179
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7179", "html_url": "https://github.com/huggingface/datasets/pull/7179", "diff_url": "https://github.com/huggingface/datasets/pull/7179.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7179.patch", "merged_at": "2024-10-08T16:21:03" }
7,179
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7179). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,552,378,330
Support Python 3.11
closed
Support Python 3.11: https://peps.python.org/pep-0664/
2024-09-27T08:50:47
2024-10-08T16:21:04
2024-10-08T16:21:04
https://github.com/huggingface/datasets/issues/7178
null
7,178
false
[]
2,552,371,082
Fix release instructions
closed
Fix release instructions. During last release, I had to make this additional update.
2024-09-27T08:47:01
2024-09-27T08:57:35
2024-09-27T08:57:32
https://github.com/huggingface/datasets/pull/7177
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7177", "html_url": "https://github.com/huggingface/datasets/pull/7177", "diff_url": "https://github.com/huggingface/datasets/pull/7177.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7177.patch", "merged_at": "2024-09-27T08:57:32" }
7,177
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7177). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,551,025,564
fix grammar in fingerprint.py
open
I see this error all the time and it was starting to get to me.
2024-09-26T16:13:42
2024-09-26T16:13:42
null
https://github.com/huggingface/datasets/pull/7176
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7176", "html_url": "https://github.com/huggingface/datasets/pull/7176", "diff_url": "https://github.com/huggingface/datasets/pull/7176.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7176.patch", "merged_at": null }
7,176
true
[]
2,550,957,337
[FSTimeoutError] load_dataset
closed
### Describe the bug When using `load_dataset`to load [HuggingFaceM4/VQAv2](https://huggingface.co/datasets/HuggingFaceM4/VQAv2), I am getting `FSTimeoutError`. ### Error ``` TimeoutError: The above exception was the direct cause of the following exception: FSTimeoutError Traceback (most recent call last) [/usr/local/lib/python3.10/dist-packages/fsspec/asyn.py](https://klh9mr78js-496ff2e9c6d22116-0-colab.googleusercontent.com/outputframe.html?vrz=colab_20240924-060116_RC00_678132060#) in sync(loop, func, timeout, *args, **kwargs) 99 if isinstance(return_result, asyncio.TimeoutError): 100 # suppress asyncio.TimeoutError, raise FSTimeoutError --> 101 raise FSTimeoutError from return_result 102 elif isinstance(return_result, BaseException): 103 raise return_result FSTimeoutError: ``` It usually fails around 5-6 GB. <img width="847" alt="Screenshot 2024-09-26 at 9 10 19 PM" src="https://github.com/user-attachments/assets/ff91995a-fb55-4de6-8214-94025d6c8470"> ### Steps to reproduce the bug To reproduce it, run this in colab notebook: ``` !pip install -q -U datasets from datasets import load_dataset ds = load_dataset('HuggingFaceM4/VQAv2', split="train[:10%]") ``` ### Expected behavior It should download properly. ### Environment info Using Colab Notebook.
2024-09-26T15:42:29
2025-02-01T09:09:35
2024-09-30T17:28:35
https://github.com/huggingface/datasets/issues/7175
null
7,175
false
[ "Is this `FSTimeoutError` due to download network issue from remote resource (from where it is being accessed)?", "It seems to happen for all datasets, not just a specific one, and especially for versions after 3.0. (3.0.0, 3.0.1 have this problem)\r\n\r\nI had the same error on a different dataset, but after downgrading to datasets==2.21.0, the problem was solved.", "Same as https://github.com/huggingface/datasets/issues/7164\r\n\r\nThis dataset is made of a python script that downloads data from elsewhere than HF, so availability depends on the original host. Ultimately it would be nice to host the files of this dataset on HF\r\n\r\nin `datasets` <3.0 there were lots of mechanisms that got removed after the decision to make datasets with python loading scripts legacy for security and maintenance reasons (we only do very basic support now)", "@lhoestq Thank you for the clarification! Closing the issue.", "I'm getting this too, and also at 5 minutes. But for `CSTR-Edinburgh/vctk`, so it's not just this dataset, it seems to be a timeout that was introduced and needs to be raised. The progress bar was moving along just fine before the timeout, and I get more or less of it depending on how fast the network is.", "You can change the `aiohttp` timeout from 5min to 1h like this:\r\n\r\n```python\r\nimport datasets, aiohttp\r\ndataset = datasets.load_dataset(\r\n dataset_name,\r\n storage_options={'client_kwargs': {'timeout': aiohttp.ClientTimeout(total=3600)}}\r\n)\r\n```", "@JonasLoos Solution solved a download timeout error I received when downloading `\"HuggingFaceM4/VQAv2\"` 🎉 " ]
2,549,892,315
Set dev version
closed
null
2024-09-26T08:30:11
2024-09-26T08:32:39
2024-09-26T08:30:21
https://github.com/huggingface/datasets/pull/7174
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7174", "html_url": "https://github.com/huggingface/datasets/pull/7174", "diff_url": "https://github.com/huggingface/datasets/pull/7174.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7174.patch", "merged_at": "2024-09-26T08:30:21" }
7,174
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7174). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,549,882,529
Release: 3.0.1
closed
null
2024-09-26T08:25:54
2024-09-26T08:28:29
2024-09-26T08:26:03
https://github.com/huggingface/datasets/pull/7173
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7173", "html_url": "https://github.com/huggingface/datasets/pull/7173", "diff_url": "https://github.com/huggingface/datasets/pull/7173.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7173.patch", "merged_at": "2024-09-26T08:26:03" }
7,173
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7173). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,549,781,691
Add torchdata as a regular test dependency
closed
Add `torchdata` as a regular test dependency. Note that previously, `torchdata` was installed from their repo and current main branch (0.10.0.dev) requires Python>=3.9. Also note they made a recent release: 0.8.0 on Jul 31, 2024. Fix #7171.
2024-09-26T07:45:55
2024-09-26T08:12:12
2024-09-26T08:05:40
https://github.com/huggingface/datasets/pull/7172
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7172", "html_url": "https://github.com/huggingface/datasets/pull/7172", "diff_url": "https://github.com/huggingface/datasets/pull/7172.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7172.patch", "merged_at": "2024-09-26T08:05:40" }
7,172
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7172). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2,549,738,919
CI is broken: No solution found when resolving dependencies
closed
See: https://github.com/huggingface/datasets/actions/runs/11046967444/job/30687294297 ``` Run uv pip install --system -r additional-tests-requirements.txt --no-deps × No solution found when resolving dependencies: ╰─▶ Because the current Python version (3.8.18) does not satisfy Python>=3.9 and torchdata==0.10.0a0+1a98f21 depends on Python>=3.9, we can conclude that torchdata==0.10.0a0+1a98f21 cannot be used. And because only torchdata==0.10.0a0+1a98f21 is available and you require torchdata, we can conclude that your requirements are unsatisfiable. Error: Process completed with exit code 1. ```
2024-09-26T07:24:58
2024-09-26T08:05:41
2024-09-26T08:05:41
https://github.com/huggingface/datasets/issues/7171
null
7,171
false
[]