id
int64
599M
3.26B
number
int64
1
7.7k
title
stringlengths
1
290
body
stringlengths
0
228k
state
stringclasses
2 values
html_url
stringlengths
46
51
created_at
timestamp[s]date
2020-04-14 10:18:02
2025-07-23 08:04:53
updated_at
timestamp[s]date
2020-04-27 16:04:17
2025-07-23 18:53:44
closed_at
timestamp[s]date
2020-04-14 12:01:40
2025-07-23 16:44:42
user
dict
labels
listlengths
0
4
is_pull_request
bool
2 classes
comments
listlengths
0
0
1,899,848,414
6,246
Add new column to dataset
### Describe the bug ``` --------------------------------------------------------------------------- KeyError Traceback (most recent call last) [<ipython-input-9-bd197b36b6a0>](https://localhost:8080/#) in <cell line: 1>() ----> 1 dataset['train']['/workspace/data'] 3 frames [/usr/local/lib/python3.10/dist-packages/datasets/formatting/formatting.py](https://localhost:8080/#) in _check_valid_column_key(key, columns) 518 def _check_valid_column_key(key: str, columns: List[str]) -> None: 519 if key not in columns: --> 520 raise KeyError(f"Column {key} not in the dataset. Current columns in the dataset: {columns}") 521 522 KeyError: "Column train not in the dataset. Current columns in the dataset: ['image', '/workspace/data']" ``` ### Steps to reproduce the bug please find the notebook for reference: https://colab.research.google.com/drive/10lZ_zLtU4itYVmIVTvIEVbjfOtCZaAZy?usp=sharing ### Expected behavior add column to the dataset ### Environment info colab pro
closed
https://github.com/huggingface/datasets/issues/6246
2023-09-17T16:59:48
2023-09-18T16:20:09
2023-09-18T16:20:09
{ "login": "andysingal", "id": 20493493, "type": "User" }
[]
false
[]
1,898,861,422
6,244
Add support for `fsspec>=2023.9.0`
Fix #6214
closed
https://github.com/huggingface/datasets/pull/6244
2023-09-15T17:58:25
2023-09-26T15:41:38
2023-09-26T15:32:51
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,898,532,784
6,243
Fix cast from fixed size list to variable size list
Fix #6242
closed
https://github.com/huggingface/datasets/pull/6243
2023-09-15T14:23:33
2023-09-19T18:02:21
2023-09-19T17:53:17
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,896,899,123
6,242
Data alteration when loading dataset with unspecified inner sequence length
### Describe the bug When a dataset saved with a specified inner sequence length is loaded without specifying that length, the original data is altered and becomes inconsistent. ### Steps to reproduce the bug ```python from datasets import Dataset, Features, Value, Sequence, load_dataset # Repository ID repo_id = "my_repo_id" # Define features with a specific length of 3 for each inner sequence specified_features = Features({"key": Sequence(Sequence(Value("float32"), length=3))}) # Create a dataset with the specified features data = [ [[1.0, 2.0, 3.0], [4.0, 5.0, 6.0]], [[7.0, 8.0, 9.0], [10.0, 11.0, 12.0]], ] dataset = Dataset.from_dict({"key": data}, features=specified_features) # Push the dataset to the hub dataset.push_to_hub(repo_id) # Define features without specifying the length unspecified_features = Features({"key": Sequence(Sequence(Value("float32")))}) # Load the dataset from the hub with this new feature definition dataset = load_dataset(f"qgallouedec/{repo_id}", split="train", features=unspecified_features) # The obtained data is altered print(dataset.to_dict()) # {'key': [[[1.0], [2.0]], [[3.0], [4.0]]]} ``` ### Expected behavior ```python print(dataset.to_dict()) # {'key': [[[1.0, 2.0, 3.0], [4.0, 5.0, 6.0]], [[7.0, 8.0, 9.0], [10.0, 11.0, 12.0]]]} ``` ### Environment info - `datasets` version: 2.14.4 - Platform: Linux-6.2.0-32-generic-x86_64-with-glibc2.35 - Python version: 3.9.12 - Huggingface_hub version: 0.15.1 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
closed
https://github.com/huggingface/datasets/issues/6242
2023-09-14T16:12:45
2023-09-19T17:53:18
2023-09-19T17:53:18
{ "login": "qgallouedec", "id": 45557362, "type": "User" }
[]
false
[]
1,896,429,694
6,241
Remove unused global variables in `audio.py`
null
closed
https://github.com/huggingface/datasets/pull/6241
2023-09-14T12:06:32
2023-09-15T15:57:10
2023-09-15T15:46:07
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,895,723,888
6,240
Dataloader stuck on multiple GPUs
### Describe the bug I am trying to get CLIP to fine-tuning with my code. When I tried to run it on multiple GPUs using accelerate, I encountered the following phenomenon. - Validation dataloader stuck in 2nd epoch only on multi-GPU Specifically, when the "for inputs in valid_loader:" process is finished, it does not proceed to the next step. train_loader process is completed. Also, both train and valid are working correctly in the first epoch. The accelerate command at that time is as follows. `accelerate launch --multi_gpu --num_processes=2 {script_name.py} {--arg1} {--arg2} ...` - This will not happen when single GPU is used. `CUDA_VISIBLE_DEVICES="0" accelerate launch {script_name.py} --arg1 --arg2 ...` - Setting num_workers=0 in dataloader did not change the result. ### Steps to reproduce the bug 1. The codes for fine-tuning the regular CLIP were updated for accelerate. 2. Run the code with the accelerate command as `accelerate launch --multi_gpu --num_processes=2 {script_name.py} {--arg1} {--arg2} ...` and the above problem will occur. 3. CUDA_VISIBLE_DEVICES="0" accelerate launch {script_name.py} --arg1 --arg2 ...` , it works fine. ### Expected behavior It Should end normally as if it was run on a single GPU. ### Environment info Since `datasets-cli env` did not work, the environment is described below. - OS: Ubuntu 22.04 with Docker - Docker: 24.0.5, build ced0996 - Python: 3.10.12 - torch==2.0.1 - accelerate==0.21.0 - transformers==4.33.1
closed
https://github.com/huggingface/datasets/issues/6240
2023-09-14T05:30:30
2023-09-14T23:54:42
2023-09-14T23:54:42
{ "login": "kuri54", "id": 40049003, "type": "User" }
[]
false
[]
1,895,349,382
6,239
Load local audio data doesn't work
### Describe the bug I get a RuntimeError from the following code: ```python audio_dataset = Dataset.from_dict({"audio": ["/kaggle/input/bengaliai-speech/train_mp3s/000005f3362c.mp3"]}).cast_column("audio", Audio()) audio_dataset[0] ``` ### Traceback <details> ```python RuntimeError Traceback (most recent call last) Cell In[33], line 1 ----> 1 train_dataset[0] File /opt/conda/lib/python3.10/site-packages/datasets/arrow_dataset.py:1764, in Dataset.__getitem__(self, key) 1762 def __getitem__(self, key): # noqa: F811 1763 """Can be used to index columns (by string names) or rows (by integer index or iterable of indices or bools).""" -> 1764 return self._getitem( 1765 key, 1766 ) File /opt/conda/lib/python3.10/site-packages/datasets/arrow_dataset.py:1749, in Dataset._getitem(self, key, decoded, **kwargs) 1747 formatter = get_formatter(format_type, features=self.features, decoded=decoded, **format_kwargs) 1748 pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None) -> 1749 formatted_output = format_table( 1750 pa_subtable, key, formatter=formatter, format_columns=format_columns, output_all_columns=output_all_columns 1751 ) 1752 return formatted_output File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:532, in format_table(table, key, formatter, format_columns, output_all_columns) 530 python_formatter = PythonFormatter(features=None) 531 if format_columns is None: --> 532 return formatter(pa_table, query_type=query_type) 533 elif query_type == "column": 534 if key in format_columns: File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:281, in Formatter.__call__(self, pa_table, query_type) 279 def __call__(self, pa_table: pa.Table, query_type: str) -> Union[RowFormat, ColumnFormat, BatchFormat]: 280 if query_type == "row": --> 281 return self.format_row(pa_table) 282 elif query_type == "column": 283 return self.format_column(pa_table) File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:312, in PythonFormatter.format_row(self, pa_table) 310 row = self.python_arrow_extractor().extract_row(pa_table) 311 if self.decoded: --> 312 row = self.python_features_decoder.decode_row(row) 313 return row File /opt/conda/lib/python3.10/site-packages/datasets/formatting/formatting.py:221, in PythonFeaturesDecoder.decode_row(self, row) 220 def decode_row(self, row: dict) -> dict: --> 221 return self.features.decode_example(row) if self.features else row File /opt/conda/lib/python3.10/site-packages/datasets/features/features.py:1386, in Features.decode_example(self, example) 1376 def decode_example(self, example: dict): 1377 """Decode example with custom feature decoding. 1378 1379 Args: (...) 1383 :obj:`dict[str, Any]` 1384 """ -> 1386 return { 1387 column_name: decode_nested_example(feature, value) 1388 if self._column_requires_decoding[column_name] 1389 else value 1390 for column_name, (feature, value) in zip_dict( 1391 {key: value for key, value in self.items() if key in example}, example 1392 ) 1393 } File /opt/conda/lib/python3.10/site-packages/datasets/features/features.py:1387, in <dictcomp>(.0) 1376 def decode_example(self, example: dict): 1377 """Decode example with custom feature decoding. 1378 1379 Args: (...) 1383 :obj:`dict[str, Any]` 1384 """ 1386 return { -> 1387 column_name: decode_nested_example(feature, value) 1388 if self._column_requires_decoding[column_name] 1389 else value 1390 for column_name, (feature, value) in zip_dict( 1391 {key: value for key, value in self.items() if key in example}, example 1392 ) 1393 } File /opt/conda/lib/python3.10/site-packages/datasets/features/features.py:1087, in decode_nested_example(schema, obj) 1085 # Object with special decoding: 1086 elif isinstance(schema, (Audio, Image)): -> 1087 return schema.decode_example(obj) if obj is not None else None 1088 return obj File /opt/conda/lib/python3.10/site-packages/datasets/features/audio.py:103, in Audio.decode_example(self, value) 101 raise ValueError(f"An audio sample should have one of 'path' or 'bytes' but both are None in {value}.") 102 elif path is not None and path.endswith("mp3"): --> 103 array, sampling_rate = self._decode_mp3(file if file else path) 104 elif path is not None and path.endswith("opus"): 105 if file: File /opt/conda/lib/python3.10/site-packages/datasets/features/audio.py:241, in Audio._decode_mp3(self, path_or_file) 238 except RuntimeError as err: 239 raise ImportError("To support decoding 'mp3' audio files, please install 'sox'.") from err --> 241 array, sampling_rate = torchaudio.load(path_or_file, format="mp3") 242 if self.sampling_rate and self.sampling_rate != sampling_rate: 243 if not hasattr(self, "_resampler") or self._resampler.orig_freq != sampling_rate: File /opt/conda/lib/python3.10/site-packages/torchaudio/backend/sox_io_backend.py:256, in load(filepath, frame_offset, num_frames, normalize, channels_first, format) 254 if ret is not None: 255 return ret --> 256 return _fallback_load(filepath, frame_offset, num_frames, normalize, channels_first, format) File /opt/conda/lib/python3.10/site-packages/torchaudio/backend/sox_io_backend.py:30, in _fail_load(filepath, frame_offset, num_frames, normalize, channels_first, format) 22 def _fail_load( 23 filepath: str, 24 frame_offset: int = 0, (...) 28 format: Optional[str] = None, 29 ) -> Tuple[torch.Tensor, int]: ---> 30 raise RuntimeError("Failed to load audio from {}".format(filepath)) RuntimeError: Failed to load audio from /kaggle/input/bengaliai-speech/train_mp3s/000005f3362c.mp3 ``` </details> ### Steps to reproduce the bug 1. - Create a custom dataset using Local files of type mp3. 3. - Try to read the first audio item. ### Expected behavior Expected output ```python audio_dataset[0]["audio"] {'array': array([ 0. , 0.00024414, -0.00024414, ..., -0.00024414, 0. , 0. ], dtype=float32), 'path': 'path/to/audio_1', 'sampling_rate': 16000} ``` ### Environment info N/A
closed
https://github.com/huggingface/datasets/issues/6239
2023-09-13T22:30:01
2023-09-15T14:32:10
2023-09-15T14:32:10
{ "login": "abodacs", "id": 554032, "type": "User" }
[]
false
[]
1,895,207,828
6,238
`dataset.filter` ALWAYS removes the first item from the dataset when using batched=True
### Describe the bug If you call batched=True when calling `filter`, the first item is _always_ filtered out, regardless of the filter condition. ### Steps to reproduce the bug Here's a minimal example: ```python def filter_batch_always_true(batch, indices): print("First index being passed into this filter function: ", indices[0]) return indices # Keep all indices data = {"value": list(range(10))} dataset = Dataset.from_dict(data) filtered_dataset = dataset.filter(filter_batch_always_true, with_indices=True, batched=True) print("Length of original dataset: ", len(dataset)) print("Length of filtered_dataset: ", len(filtered_dataset)) print("Is equal to original? ", len(filtered_dataset) == len(dataset)) print("First item of filtered dataset: ", filtered_dataset[0]) print("Last item of filtered dataset: ", filtered_dataset[-1]) ``` prints: ``` First index being passed into this filter function: 0 Length of original dataset: 10 Length of filtered_dataset: 9 Is equal to original? False First item of filtered dataset: {'value': 1} Last item of filtered dataset: {'value': 9} ``` ### Expected behavior Filter should respect the filter condition. ### Environment info - `datasets` version: 2.14.4 - Platform: macOS-13.5-arm64-arm-64bit - Python version: 3.9.18 - Huggingface_hub version: 0.17.1 - PyArrow version: 10.0.1 - Pandas version: 2.0.2
closed
https://github.com/huggingface/datasets/issues/6238
2023-09-13T20:20:37
2023-09-17T07:05:07
2023-09-17T07:05:07
{ "login": "Taytay", "id": 1330693, "type": "User" }
[]
false
[]
1,893,822,321
6,237
Tokenization with multiple workers is too slow
I am trying to tokenize a few million documents with multiple workers but the tokenization process is taking forever. Code snippet: ``` raw_datasets.map( encode_function, batched=False, num_proc=args.preprocessing_num_workers, load_from_cache_file=not args.overwrite_cache, remove_columns=[name for name in raw_datasets["train"].column_names if name not in ["input_ids", "labels", "attention_mask"]], desc="Tokenizing data", ) ``` Details: ``` transformers==4.28.0.dev0 datasets==4.28.0.dev0 preprocessing_num_workers==48 ``` tokenizer == decapoda-research/llama-7b-hf
closed
https://github.com/huggingface/datasets/issues/6237
2023-09-13T06:18:34
2023-09-19T21:54:58
2023-09-19T21:54:58
{ "login": "macabdul9", "id": 25720695, "type": "User" }
[]
false
[]
1,893,648,480
6,236
Support buffer shuffle for to_tf_dataset
### Feature request I'm using to_tf_dataset to convert a large dataset to tf.data.Dataset and use Keras fit to train model. Currently, to_tf_dataset only supports full size shuffle, which can be very slow on large dataset. tf.data.Dataset support buffer shuffle by default. shuffle( buffer_size, seed=None, reshuffle_each_iteration=None, name=None ) ### Motivation I'm very frustrated to find the loading with shuffling large dataset is very slow. It seems impossible to shuffle before training Keras with big dataset. ### Your contribution NA
open
https://github.com/huggingface/datasets/issues/6236
2023-09-13T03:19:44
2023-09-18T01:11:21
null
{ "login": "EthanRock", "id": 7635551, "type": "User" }
[ { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,893,337,083
6,235
Support multiprocessing for download/extract nestedly
### Feature request Current multiprocessing for download/extract is not done nestedly. For example, when processing SlimPajama, there is only 3 processes (for train/test/val), while there are many files inside these 3 folders ``` Downloading data files #0: 0%| | 0/1 [00:00<?, ?obj/s] Downloading data files #1: 0%| | 0/1 [00:00<?, ?obj/s] Downloading data files #2: 0%| | 0/1 [00:00<?, ?obj/s] Extracting data files #0: 0%| | 0/1 [00:00<?, ?obj/s] Extracting data files #1: 0%| | 0/1 [00:00<?, ?obj/s] Extracting data files #2: 0%| | 0/1 [00:00<?, ?obj/s] ``` ### Motivation speedup dataset loading ### Your contribution I can help test the feature
open
https://github.com/huggingface/datasets/issues/6235
2023-09-12T21:51:08
2023-09-12T21:51:08
null
{ "login": "hgt312", "id": 22725729, "type": "User" }
[ { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,891,804,286
6,233
Update README.md
fixed a typo
closed
https://github.com/huggingface/datasets/pull/6233
2023-09-12T06:53:06
2023-09-13T18:20:50
2023-09-13T18:10:04
{ "login": "NinoRisteski", "id": 95188570, "type": "User" }
[]
true
[]
1,891,109,762
6,232
Improve error message for missing function parameters
The error message in the fingerprint module was missing the f-string 'f' symbol, so the error message returned by fingerprint.py, line 469 was literally "function {func} is missing parameters {fingerprint_names} in signature." This has been fixed.
closed
https://github.com/huggingface/datasets/pull/6232
2023-09-11T19:11:58
2023-09-15T18:07:56
2023-09-15T17:59:02
{ "login": "suavemint", "id": 4016832, "type": "User" }
[]
true
[]
1,890,863,249
6,231
Overwrite legacy default config name in `dataset_infos.json` in packaged datasets
Currently if we push data as default config with `.push_to_hub` to a repo that has a legacy `dataset_infos.json` file containing a legacy default config name like `{username}--{dataset_name}`, new key `"default"` is added to `dataset_infos.json` along with the legacy one. I think the legacy one should be dropped in this case. Also, in `load.py` I suggest to check if a legacy config name is indeed a legacy config name because after this fix it might not be the case (this check was first introduced in https://github.com/huggingface/datasets/pull/6218)
open
https://github.com/huggingface/datasets/pull/6231
2023-09-11T16:27:09
2023-09-26T11:19:36
null
{ "login": "polinaeterna", "id": 16348744, "type": "User" }
[]
true
[]
1,890,521,006
6,230
Don't skip hidden files in `dl_manager.iter_files` when they are given as input
Required for `load_dataset(<format>, data_files=["path/to/.hidden_file"])` to work as expected
closed
https://github.com/huggingface/datasets/pull/6230
2023-09-11T13:29:19
2023-09-13T18:21:28
2023-09-13T18:12:09
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,889,050,954
6,229
Apply inference on all images in the dataset
### Describe the bug ``` --------------------------------------------------------------------------- NotImplementedError Traceback (most recent call last) Cell In[14], line 11 9 for idx, example in enumerate(dataset['train']): 10 image_path = example['image'] ---> 11 mask_image = process_image(image_path) 12 mask_image.save(f"mask_{idx}.png") Cell In[14], line 4, in process_image(image_path) 2 def process_image(image_path): 3 print("Processing image:", image_path) ----> 4 result = inferencer(image_path)['predictions'] 5 mask = np.where(result == 12, 255, 0).astype('uint8') 6 return Image.fromarray(mask) File /usr/local/lib/python3.10/dist-packages/mmseg/apis/mmseg_inferencer.py:183, in MMSegInferencer.__call__(self, inputs, return_datasamples, batch_size, show, wait_time, out_dir, img_out_dir, pred_out_dir, **kwargs) 180 pred_out_dir = '' 181 img_out_dir = '' --> 183 return super().__call__( 184 inputs=inputs, 185 return_datasamples=return_datasamples, 186 batch_size=batch_size, 187 show=show, 188 wait_time=wait_time, 189 img_out_dir=img_out_dir, 190 pred_out_dir=pred_out_dir, 191 **kwargs) File /usr/local/lib/python3.10/dist-packages/mmengine/infer/infer.py:221, in BaseInferencer.__call__(self, inputs, return_datasamples, batch_size, **kwargs) 218 inputs = self.preprocess( 219 ori_inputs, batch_size=batch_size, **preprocess_kwargs) 220 preds = [] --> 221 for data in (track(inputs, description='Inference') 222 if self.show_progress else inputs): 223 preds.extend(self.forward(data, **forward_kwargs)) 224 visualization = self.visualize( 225 ori_inputs, preds, 226 **visualize_kwargs) # type: ignore # noqa: E501 File /usr/local/lib/python3.10/dist-packages/rich/progress.py:168, in track(sequence, description, total, auto_refresh, console, transient, get_time, refresh_per_second, style, complete_style, finished_style, pulse_style, update_period, disable, show_speed) 157 progress = Progress( 158 *columns, 159 auto_refresh=auto_refresh, (...) 164 disable=disable, 165 ) 167 with progress: --> 168 yield from progress.track( 169 sequence, total=total, description=description, update_period=update_period 170 ) File /usr/local/lib/python3.10/dist-packages/rich/progress.py:1210, in Progress.track(self, sequence, total, task_id, description, update_period) 1208 if self.live.auto_refresh: 1209 with _TrackThread(self, task_id, update_period) as track_thread: -> 1210 for value in sequence: 1211 yield value 1212 track_thread.completed += 1 File /usr/local/lib/python3.10/dist-packages/mmengine/infer/infer.py:291, in BaseInferencer.preprocess(self, inputs, batch_size, **kwargs) 266 """Process the inputs into a model-feedable format. 267 268 Customize your preprocess by overriding this method. Preprocess should (...) 287 Any: Data processed by the ``pipeline`` and ``collate_fn``. 288 """ 289 chunked_data = self._get_chunk_data( 290 map(self.pipeline, inputs), batch_size) --> 291 yield from map(self.collate_fn, chunked_data) File /usr/local/lib/python3.10/dist-packages/mmengine/infer/infer.py:588, in BaseInferencer._get_chunk_data(self, inputs, chunk_size) 586 chunk_data = [] 587 for _ in range(chunk_size): --> 588 processed_data = next(inputs_iter) 589 chunk_data.append(processed_data) 590 yield chunk_data File /usr/local/lib/python3.10/dist-packages/mmcv/transforms/base.py:12, in BaseTransform.__call__(self, results) 9 def __call__(self, 10 results: Dict) -> Optional[Union[Dict, Tuple[List, List]]]: ---> 12 return self.transform(results) File /usr/local/lib/python3.10/dist-packages/mmcv/transforms/wrappers.py:88, in Compose.transform(self, results) 79 """Call function to apply transforms sequentially. 80 81 Args: (...) 85 dict or None: Transformed results. 86 """ 87 for t in self.transforms: ---> 88 results = t(results) # type: ignore 89 if results is None: 90 return None File /usr/local/lib/python3.10/dist-packages/mmcv/transforms/base.py:12, in BaseTransform.__call__(self, results) 9 def __call__(self, 10 results: Dict) -> Optional[Union[Dict, Tuple[List, List]]]: ---> 12 return self.transform(results) File /usr/local/lib/python3.10/dist-packages/mmseg/datasets/transforms/loading.py:496, in InferencerLoader.transform(self, single_input) 494 inputs = single_input 495 else: --> 496 raise NotImplementedError 498 if 'img' in inputs: 499 return self.from_ndarray(inputs) NotImplementedError: ```` ### Steps to reproduce the bug ``` from datasets import load_dataset dataset = load_dataset('Andyrasika/cat_kingdom') dataset from mmseg.apis import MMSegInferencer checkpoint_name = 'segformer_mit-b5_8xb2-160k_ade20k-640x640' inferencer = MMSegInferencer(model=checkpoint_name) # Define a function to apply the code to each image in the dataset def process_image(image_path): print("Processing image:", image_path) result = inferencer(image_path)['predictions'] mask = np.where(result == 12, 255, 0).astype('uint8') return Image.fromarray(mask) # Process and save masks for each image in the dataset for idx, example in enumerate(dataset['train']): image_path = example['image'] mask_image = process_image(image_path) mask_image.save(f"mask_{idx}.png") ``` ### Expected behavior create a separate column with masks in the dataset and further shows as a separate column in hub ### Environment info jupyter notebook RTX 3090
closed
https://github.com/huggingface/datasets/issues/6229
2023-09-10T08:36:12
2023-09-20T16:11:53
2023-09-20T16:11:52
{ "login": "andysingal", "id": 20493493, "type": "User" }
[]
false
[]
1,887,959,311
6,228
Remove RGB -> BGR image conversion in Object Detection tutorial
Fix #6225
closed
https://github.com/huggingface/datasets/pull/6228
2023-09-08T16:09:13
2023-09-08T18:02:49
2023-09-08T17:52:16
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,887,462,591
6,226
Add push_to_hub with multiple configs docs
null
closed
https://github.com/huggingface/datasets/pull/6226
2023-09-08T11:08:55
2023-09-08T12:29:21
2023-09-08T12:20:51
{ "login": "lhoestq", "id": 42851186, "type": "User" }
[]
true
[]
1,887,054,320
6,225
Conversion from RGB to BGR in Object Detection tutorial
The [tutorial](https://huggingface.co/docs/datasets/main/en/object_detection) mentions the necessity of conversion the input image from BGR to RGB > albumentations expects the image to be in BGR format, not RGB, so you’ll have to convert the image before applying the transform. [Link to tutorial](https://github.com/huggingface/datasets/blob/0a068dbf3b446417ffd89d32857608394ec699e6/docs/source/object_detection.mdx#L77) However, relevant albumentations' tutorials [on channels conversion](https://albumentations.ai/docs/examples/example/#read-the-image-from-the-disk-and-convert-it-from-the-bgr-color-space-to-the-rgb-color-space) and [on boxes](https://albumentations.ai/docs/examples/example_bboxes/) imply that it's not really true no more. I suggest removing this outdated conversion from the tutorial.
closed
https://github.com/huggingface/datasets/issues/6225
2023-09-08T06:49:19
2023-09-08T17:52:18
2023-09-08T17:52:17
{ "login": "samokhinv", "id": 33297401, "type": "User" }
[]
false
[]
1,886,043,692
6,224
Ignore `dataset_info.json` in data files resolution
`save_to_disk` creates this file, but also [`HugginFaceDatasetSever`](https://github.com/gradio-app/gradio/blob/26fef8c7f85a006c7e25cdbed1792df19c512d02/gradio/flagging.py#L214), so this is needed to avoid issues such as [this one](https://discord.com/channels/879548962464493619/1149295819938349107/1149295819938349107).
closed
https://github.com/huggingface/datasets/pull/6224
2023-09-07T14:43:51
2023-09-07T15:46:10
2023-09-07T15:37:20
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,885,710,696
6,223
Update README.md
fixed a few typos
closed
https://github.com/huggingface/datasets/pull/6223
2023-09-07T11:33:20
2023-09-13T22:32:31
2023-09-13T22:23:42
{ "login": "NinoRisteski", "id": 95188570, "type": "User" }
[]
true
[]
1,884,875,510
6,222
fix typo in Audio dataset documentation
There is a typo in the section of the documentation dedicated to creating an audio dataset. The Dataset is incorrectly suffixed with a `Config` https://huggingface.co/datasets/indonesian-nlp/librivox-indonesia/blob/main/librivox-indonesia.py#L59
closed
https://github.com/huggingface/datasets/pull/6222
2023-09-06T23:17:24
2023-10-03T14:18:41
2023-09-07T15:39:09
{ "login": "prassanna-ravishankar", "id": 3224332, "type": "User" }
[]
true
[]
1,884,324,631
6,221
Support saving datasets with custom formatting
Requested in https://discuss.huggingface.co/t/using-set-transform-on-a-dataset-leads-to-an-exception/53036. I am not sure if supporting this is the best idea for the following reasons: >For this to work, we would have to pickle a custom transform, which means the transform and the objects it references need to be serializable. Also, deserializing these bytes would make `load_from_disk` unsafe, so I'm not sure this is a good idea. @lhoestq WDYT?
open
https://github.com/huggingface/datasets/issues/6221
2023-09-06T16:03:32
2023-09-06T18:32:07
null
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
false
[]
1,884,285,980
6,220
Set dev version
null
closed
https://github.com/huggingface/datasets/pull/6220
2023-09-06T15:40:33
2023-09-06T15:52:33
2023-09-06T15:41:13
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[]
true
[]
1,884,244,334
6,219
Release: 2.14.5
null
closed
https://github.com/huggingface/datasets/pull/6219
2023-09-06T15:17:10
2023-09-06T15:46:20
2023-09-06T15:18:51
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[]
true
[]
1,883,734,000
6,218
Rename old push_to_hub configs to "default" in dataset_infos
Fix ```python from datasets import load_dataset_builder b = load_dataset_builder("lambdalabs/pokemon-blip-captions", "default") print(b.info) ``` which should return ``` DatasetInfo( features={'image': Image(decode=True, id=None), 'text': Value(dtype='string', id=None)}, dataset_name='pokemon-blip-captions', config_name='default', version=0.0.0, splits={'train': SplitInfo(name='train', num_bytes=119417410.0, num_examples=833, shard_lengths=None, dataset_name='pokemon-blip-captions')}, download_checksums=None, download_size=99672355, dataset_size=119417410.0, size_in_bytes=219089765.0, ... ) ``` instead of and empty dataset info. The dataset has a dataset_infos.json file with a deprecated config name "lambdalabs--pokemon-blip-captions". We switched those config names to "default" in 2.14, so the builder.info should take this into account.
closed
https://github.com/huggingface/datasets/pull/6218
2023-09-06T10:40:05
2023-09-07T08:31:29
2023-09-06T11:23:56
{ "login": "lhoestq", "id": 42851186, "type": "User" }
[]
true
[]
1,883,614,607
6,217
`Dataset.to_dict()` ignore `decode=True` with Image feature
### Describe the bug `Dataset.to_dict` seems to ignore the decoding instruction passed in features. ### Steps to reproduce the bug ```python import datasets import numpy as np from PIL import Image img = np.random.randint(0, 256, (5, 5, 3), dtype=np.uint8) img = Image.fromarray(img) features = datasets.Features({"image": datasets.Image(decode=True)}) dataset = datasets.Dataset.from_dict({"image": [img]}, features=features) print({key: dataset[key] for key in dataset.column_names}) # {'image': [<PIL.PngImagePlugin.PngImageFile image mode=RGB size=5x5 at 0x7EFBC80E15B0>]} print(dataset.to_dict()) # {'image': [{'bytes': b'\x89PNG\r\n\x1a\n\x00\x00\x00\rIHDR\x00\x00\x00\x05\x00\x00\x00\x05\x08\x02\x00\x00\x00\x02\r\xb1\xb2\x00\x00\x00[IDATx\x9c\x01P\x00\xaf\xff\x01\x13\x1b<7\xe7\xe0\xdc^6\xed\x04\xc7M\xd2\x9f\x00X\x1b\xb0?\x1ba\x15\xc5 o\xd0\x80\xbe\x19/\x01\xec\x95\x1f\x9f\xffj\xfa1\xa7\xc4X\xea\xbe\xa4g\x00\xc4\x15\xdeC\xc7 \xbbaqe\xc8\xb9\xa9q\xe7\x00,?M\xc0)\xdaD`}\xb1\xdci\x1e\xafC\xa9]%.@\xa6\xf0\xb3\x00\x00\x00\x00IEND\xaeB`\x82', 'path': None}]} ``` ### Expected behavior I would expect `{key: dataset[key] for key in dataset.column_names}` and `dataset.to_dict()` to be equivalent. If the previous behavior is expected, then it should be stated [in the doc](https://huggingface.co/docs/datasets/v2.14.4/en/package_reference/main_classes#datasets.Dataset.to_dict). ### Environment info - `datasets` version: 2.14.4 - Platform: Linux-6.2.0-31-generic-x86_64-with-glibc2.35 - Python version: 3.9.12 - Huggingface_hub version: 0.15.1 - PyArrow version: 12.0.1 - Pandas version: 2.0.3 - Pillow 9.5.0 - numpy 1.25.2
open
https://github.com/huggingface/datasets/issues/6217
2023-09-06T09:26:16
2023-09-08T17:08:52
null
{ "login": "qgallouedec", "id": 45557362, "type": "User" }
[]
false
[]
1,883,492,703
6,216
Release: 2.13.2
null
closed
https://github.com/huggingface/datasets/pull/6216
2023-09-06T08:15:32
2023-09-06T08:52:18
2023-09-06T08:22:43
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[]
true
[]
1,882,176,970
6,215
Fix checking patterns to infer packaged builder
Don't ignore results of pattern resolving if `self.data_files` is not None. Otherwise lines 854 and 1037 make no sense.
closed
https://github.com/huggingface/datasets/pull/6215
2023-09-05T15:10:47
2023-09-06T10:34:00
2023-09-06T10:25:00
{ "login": "polinaeterna", "id": 16348744, "type": "User" }
[]
true
[]
1,881,736,469
6,214
Unpin fsspec < 2023.9.0
Once root issue is fixed, remove temporary pin of fsspec < 2023.9.0 introduced by: - #6210 Related to issue: - #6209 After investigation, I think the root issue is related to the new glob behavior with double asterisk `**` they have introduced in: - https://github.com/fsspec/filesystem_spec/pull/1329
closed
https://github.com/huggingface/datasets/issues/6214
2023-09-05T11:02:58
2023-09-26T15:32:52
2023-09-26T15:32:52
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[ { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,880,592,987
6,213
Better list array values handling in cast/embed storage
Use [`array.flatten`](https://arrow.apache.org/docs/python/generated/pyarrow.ListArray.html#pyarrow.ListArray.flatten) that takes `.offset` into account instead of `array.values` in array cast/embed.
closed
https://github.com/huggingface/datasets/pull/6213
2023-09-04T16:21:23
2024-01-11T06:32:20
2023-10-05T15:24:34
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,880,399,516
6,212
Tilde (~) is not supported for data_files
### Describe the bug Attempting to `load_dataset` from a path starting with `~` (as a shorthand for the user's home directory) seems not to be fully working - at least as far as the `parquet` dataset builder is concerned. (the same file can be loaded correctly if providing its absolute path instead) I think that this is very similar to https://github.com/huggingface/datasets/issues/5757, but for `data_files` rather than `data_dir` ### Steps to reproduce the bug ```python import datasets # save a parquet file at ~/path/to/data.parquet data_files = "~/path/to/data.parquet" dataset = datasets.load_dataset("parquet", data_files=data_files) ``` ``` Downloading data files: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 12671.61it/s] Extracting data files: 100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 22671.91it/s] Generating train split: 0 examples [00:00, ? examples/s] Traceback (most recent call last): File ".venv/lib/python3.11/site-packages/datasets/builder.py", line 1949, in _prepare_split_single num_examples, num_bytes = writer.finalize() ^^^^^^^^^^^^^^^^^ File ".venv/lib/python3.11/site-packages/datasets/arrow_writer.py", line 598, in finalize raise SchemaInferenceError("Please pass `features` or at least one example when writing data") datasets.arrow_writer.SchemaInferenceError: Please pass `features` or at least one example when writing data The above exception was the direct cause of the following exception: Traceback (most recent call last): File ".venv/lib/python3.11/site-packages/datasets/load.py", line 2133, in load_dataset builder_instance.download_and_prepare( File ".venv/lib/python3.11/site-packages/datasets/builder.py", line 954, in download_and_prepare self._download_and_prepare( File ".venv/lib/python3.11/site-packages/datasets/builder.py", line 1049, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File ".venv/lib/python3.11/site-packages/datasets/builder.py", line 1813, in _prepare_split for job_id, done, content in self._prepare_split_single( File ".venv/lib/python3.11/site-packages/datasets/builder.py", line 1958, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.builder.DatasetGenerationError: An error occurred while generating the dataset ``` ### Expected behavior Can use `~` shorthand in paths when loading local (parquet) datasets. ### Environment info `datasets 2.14.3`
open
https://github.com/huggingface/datasets/issues/6212
2023-09-04T14:23:49
2023-09-05T08:28:39
null
{ "login": "exs-avianello", "id": 128361578, "type": "User" }
[]
false
[]
1,880,265,906
6,211
Fix empty splitinfo json
If a split is empty, then the JSON split info should mention num_bytes = 0 and num_examples = 0. Until now they were omited because the JSON dumps ignore the fields that are equal to the default values. This is needed in datasets-server since we parse this information to the viewer
closed
https://github.com/huggingface/datasets/pull/6211
2023-09-04T13:13:53
2023-09-04T14:58:34
2023-09-04T14:47:17
{ "login": "lhoestq", "id": 42851186, "type": "User" }
[]
true
[]
1,879,649,731
6,210
Temporarily pin fsspec < 2023.9.0
Temporarily pin fsspec < 2023.9.0 until permanent solution is found. Hot fix #6209.
closed
https://github.com/huggingface/datasets/pull/6210
2023-09-04T07:07:07
2023-09-04T07:40:23
2023-09-04T07:30:00
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[]
true
[]
1,879,622,000
6,209
CI is broken with AssertionError: 3 failed, 12 errors
Our CI is broken: 3 failed, 12 errors See: https://github.com/huggingface/datasets/actions/runs/6069947111/job/16465138041 ``` =========================== short test summary info ============================ FAILED tests/test_load.py::ModuleFactoryTest::test_LocalDatasetModuleFactoryWithoutScript_with_data_dir - AssertionError: assert ({NamedSplit('train'): ['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_LocalDatasetModuleFactory2/data_dir2/subdir1/train.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_LocalDatasetModuleFactory2/data_dir2/subdir1/train.txt'], NamedSplit('test'): ['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_LocalDatasetModuleFactory2/data_dir2/subdir1/test.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_LocalDatasetModuleFactory2/data_dir2/subdir1/test.txt']} is not None and 2 == 1) + where 2 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_LocalDatasetModuleFactory2/data_dir2/subdir1/train.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_LocalDatasetModuleFactory2/data_dir2/subdir1/train.txt']) FAILED tests/test_load.py::test_load_dataset_arrow[False] - AssertionError: assert 20 == 10 + where 20 = Dataset({\n features: ['col_1'],\n num_rows: 20\n}).num_rows FAILED tests/test_load.py::test_load_dataset_arrow[True] - assert 20 == 10 ERROR tests/packaged_modules/test_audiofolder.py::test_data_files_with_metadata_and_multiple_splits[jsonl-False] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_2/audiofolder_data_dir_with_metadata/train/audio_file.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_2/audiofolder_data_dir_with_metadata/train/audio_file2.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_2/audiofolder_data_dir_with_metadata/train/metadata.jsonl', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_2/audiofolder_data_dir_with_metadata/train/audio_file.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_2/audiofolder_data_dir_with_metadata/train/audio_file2.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_2/audiofolder_data_dir_with_metadata/train/metadata.jsonl']) ERROR tests/packaged_modules/test_audiofolder.py::test_data_files_with_metadata_and_multiple_splits[jsonl-True] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_3/audiofolder_data_dir_with_metadata/train/audio_file.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_3/audiofolder_data_dir_with_metadata/train/audio_file2.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_3/audiofolder_data_dir_with_metadata/train/metadata.jsonl', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_3/audiofolder_data_dir_with_metadata/train/audio_file.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_3/audiofolder_data_dir_with_metadata/train/audio_file2.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_3/audiofolder_data_dir_with_metadata/train/metadata.jsonl']) ERROR tests/packaged_modules/test_audiofolder.py::test_data_files_with_metadata_and_multiple_splits[csv-False] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_4/audiofolder_data_dir_with_metadata/train/audio_file.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_4/audiofolder_data_dir_with_metadata/train/audio_file2.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_4/audiofolder_data_dir_with_metadata/train/metadata.csv', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_4/audiofolder_data_dir_with_metadata/train/audio_file.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_4/audiofolder_data_dir_with_metadata/train/audio_file2.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_4/audiofolder_data_dir_with_metadata/train/metadata.csv']) ERROR tests/packaged_modules/test_audiofolder.py::test_data_files_with_metadata_and_multiple_splits[csv-True] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_5/audiofolder_data_dir_with_metadata/train/audio_file.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_5/audiofolder_data_dir_with_metadata/train/audio_file2.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_5/audiofolder_data_dir_with_metadata/train/metadata.csv', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_5/audiofolder_data_dir_with_metadata/train/audio_file.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_5/audiofolder_data_dir_with_metadata/train/audio_file2.wav', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_5/audiofolder_data_dir_with_metadata/train/metadata.csv']) ERROR tests/packaged_modules/test_folder_based_builder.py::test_data_files_with_metadata_and_splits[1-False] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_3/autofolder_data_dir_with_metadata_two_splits/train/file.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_3/autofolder_data_dir_with_metadata_two_splits/train/file2.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_3/autofolder_data_dir_with_metadata_two_splits/train/metadata.jsonl', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_3/autofolder_data_dir_with_metadata_two_splits/train/file.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_3/autofolder_data_dir_with_metadata_two_splits/train/file2.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_3/autofolder_data_dir_with_metadata_two_splits/train/metadata.jsonl']) ERROR tests/packaged_modules/test_folder_based_builder.py::test_data_files_with_metadata_and_splits[1-True] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_4/autofolder_data_dir_with_metadata_two_splits/train/file.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_4/autofolder_data_dir_with_metadata_two_splits/train/file2.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_4/autofolder_data_dir_with_metadata_two_splits/train/metadata.jsonl', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_4/autofolder_data_dir_with_metadata_two_splits/train/file.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_4/autofolder_data_dir_with_metadata_two_splits/train/file2.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_4/autofolder_data_dir_with_metadata_two_splits/train/metadata.jsonl']) ERROR tests/packaged_modules/test_folder_based_builder.py::test_data_files_with_metadata_and_splits[2-False] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_5/autofolder_data_dir_with_metadata_two_splits/train/file.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_5/autofolder_data_dir_with_metadata_two_splits/train/file2.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_5/autofolder_data_dir_with_metadata_two_splits/train/metadata.jsonl', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_5/autofolder_data_dir_with_metadata_two_splits/train/file.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_5/autofolder_data_dir_with_metadata_two_splits/train/file2.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_5/autofolder_data_dir_with_metadata_two_splits/train/metadata.jsonl']) ERROR tests/packaged_modules/test_imagefolder.py::test_data_files_with_metadata_and_multiple_splits[jsonl-False] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_12/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_12/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb2.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_12/imagefolder_data_dir_with_metadata_two_splits/train/metadata.jsonl', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_12/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_12/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb2.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_12/imagefolder_data_dir_with_metadata_two_splits/train/metadata.jsonl']) ERROR tests/packaged_modules/test_imagefolder.py::test_data_files_with_metadata_and_multiple_splits[jsonl-True] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_13/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_13/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb2.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_13/imagefolder_data_dir_with_metadata_two_splits/train/metadata.jsonl', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_13/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_13/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb2.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_13/imagefolder_data_dir_with_metadata_two_splits/train/metadata.jsonl']) ERROR tests/packaged_modules/test_folder_based_builder.py::test_data_files_with_metadata_and_splits[2-True] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_6/autofolder_data_dir_with_metadata_two_splits/train/file.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_6/autofolder_data_dir_with_metadata_two_splits/train/file2.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_6/autofolder_data_dir_with_metadata_two_splits/train/metadata.jsonl', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_6/autofolder_data_dir_with_metadata_two_splits/train/file.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_6/autofolder_data_dir_with_metadata_two_splits/train/file2.txt', '/tmp/pytest-of-runner/pytest-0/popen-gw0/test_data_files_with_metadata_6/autofolder_data_dir_with_metadata_two_splits/train/metadata.jsonl']) ERROR tests/packaged_modules/test_imagefolder.py::test_data_files_with_metadata_and_multiple_splits[csv-False] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_14/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_14/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb2.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_14/imagefolder_data_dir_with_metadata_two_splits/train/metadata.csv', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_14/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_14/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb2.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_14/imagefolder_data_dir_with_metadata_two_splits/train/metadata.csv']) ERROR tests/packaged_modules/test_imagefolder.py::test_data_files_with_metadata_and_multiple_splits[csv-True] - AssertionError: assert 6 == 3 + where 6 = len(['/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_15/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_15/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb2.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_15/imagefolder_data_dir_with_metadata_two_splits/train/metadata.csv', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_15/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_15/imagefolder_data_dir_with_metadata_two_splits/train/image_rgb2.jpg', '/tmp/pytest-of-runner/pytest-0/popen-gw1/test_data_files_with_metadata_15/imagefolder_data_dir_with_metadata_two_splits/train/metadata.csv']) = 3 failed, 2383 passed, 26 skipped, 9 warnings, 12 errors in 280.79s (0:04:40) = ```
closed
https://github.com/huggingface/datasets/issues/6209
2023-09-04T06:47:05
2023-09-04T07:30:01
2023-09-04T07:30:01
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[ { "name": "bug", "color": "d73a4a" } ]
false
[]
1,879,572,646
6,208
Do not filter out .zip extensions from no-script datasets
This PR is a hotfix of: - #6207 That PR introduced the filtering out of `.zip` extensions. This PR reverts that. Hot fix #6207. Maybe we should do patch releases: the bug was introduced in 2.13.1. CC: @lhoestq
closed
https://github.com/huggingface/datasets/pull/6208
2023-09-04T06:07:12
2023-09-04T09:22:19
2023-09-04T09:13:32
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[]
true
[]
1,879,555,234
6,207
No-script datasets with ZIP files do not load
While investigating an issue on a Hub dataset, I have discovered the no-script datasets containing ZIP files do not load. For example, that no-script dataset containing ZIP files, raises NonMatchingSplitsSizesError: ```python In [2]: ds = load_dataset("sidovic/LearningQ-qg") NonMatchingSplitsSizesError: [ { 'expected': SplitInfo(name='train', num_bytes=0, num_examples=188660, shard_lengths=None, dataset_name=None), 'recorded': SplitInfo(name='train', num_bytes=0, num_examples=0, shard_lengths=None, dataset_name='learning_q-qg') }, { 'expected': SplitInfo(name='validation', num_bytes=0, num_examples=20630, shard_lengths=None, dataset_name=None), 'recorded': SplitInfo(name='validation', num_bytes=0, num_examples=0, shard_lengths=None, dataset_name='learning_q-qg') }, { 'expected': SplitInfo(name='test', num_bytes=0, num_examples=18227, shard_lengths=None, dataset_name=None), 'recorded': SplitInfo(name='test', num_bytes=0, num_examples=0, shard_lengths=None, dataset_name='learning_q-qg') } ] ``` As another example, a no-script dataset containing just a (CSV)-ZIP file, raises a DatasetGenerationError: ``` > num_examples, num_bytes = writer.finalize() src/datasets/builder.py:1949: > raise SchemaInferenceError("Please pass `features` or at least one example when writing data") E datasets.arrow_writer.SchemaInferenceError: Please pass `features` or at least one example when writing data src/datasets/arrow_writer.py:598: SchemaInferenceError The above exception was the direct cause of the following exception: src/datasets/load.py:2143: in load_dataset builder_instance.download_and_prepare( src/datasets/builder.py:954: in download_and_prepare self._download_and_prepare( src/datasets/builder.py:1049: in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) src/datasets/builder.py:1813: in _prepare_split for job_id, done, content in self._prepare_split_single( > raise DatasetGenerationError("An error occurred while generating the dataset") from e E datasets.builder.DatasetGenerationError: An error occurred while generating the dataset src/datasets/builder.py:1958: DatasetGenerationError ``` After investigating, I think this bug was introduced in this PR: - #5972 Related to: - https://huggingface.co/datasets/sidovic/LearningQ-qg/discussions/1 CC: @lhoestq
closed
https://github.com/huggingface/datasets/issues/6207
2023-09-04T05:50:27
2023-09-04T09:13:33
2023-09-04T09:13:33
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[ { "name": "bug", "color": "d73a4a" } ]
false
[]
1,879,473,745
6,206
When calling load_dataset, raise error: pyarrow.lib.ArrowInvalid: offset overflow while concatenating arrays
### Describe the bug When calling load_dataset, raise error ``` Traceback (most recent call last): File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/builder.py", line 1694, in _pre pare_split_single writer.write(example, key) File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/arrow_writer.py", line 490, in write self.write_examples_on_file() File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/arrow_writer.py", line 448, in write_examples_on_file self.write_batch(batch_examples=batch_examples) File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/arrow_writer.py", line 559, in write_batch self.write_table(pa_table, writer_batch_size) File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/arrow_writer.py", line 571, in write_table pa_table = pa_table.combine_chunks() ^^^^^^^^^^^^^^^^^^^^^^^^^ File "pyarrow/table.pxi", line 3439, in pyarrow.lib.Table.combine_chunks File "pyarrow/error.pxi", line 144, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 100, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: offset overflow while concatenating arrays The above exception was the direct cause of the following exception: Traceback (most recent call last): dataset = load_dataset( ^^^^^^^^^^^^^ File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/load.py", line 2133, in load_da taset builder_instance.download_and_prepare( File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/builder.py", line 954, in downl oad_and_prepare self._download_and_prepare( File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/builder.py", line 1717, in _dow nload_and_prepare super()._download_and_prepare( File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/builder.py", line 1049, in _dow nload_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/builder.py", line 1555, in _pre pare_split for job_id, done, content in self._prepare_split_single( File "/home/aihao/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/builder.py", line 1712, in _pre pare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.builder.DatasetGenerationError: An error occurred while generating the dataset Setting num_proc from 8 back to 1 for the train split to disable multiprocessing as it only contains one shard. 09/04/2023 12:02:04 - WARNING - datasets.builder - Setting num_proc from 8 back to 1 for the train split to dis able multiprocessing as it only contains one shard. ``` ### Steps to reproduce the bug Call load_dataset with the large image as feature ### Expected behavior no error ### Environment info - `datasets` version: 2.14.3 - Platform: Linux-6.2.0-31-generic-x86_64-with-glibc2.35 - Python version: 3.11.4 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
closed
https://github.com/huggingface/datasets/issues/6206
2023-09-04T04:14:00
2024-04-17T15:53:29
2023-09-04T06:05:49
{ "login": "aihao2000", "id": 51043929, "type": "User" }
[]
false
[]
1,877,491,602
6,203
Support loading from a DVC remote repository
### Feature request Adding support for loading a file from a DVC repository, tracked remotely on a SCM. ### Motivation DVC is a popular version control system to version and manage datasets. The files are stored on a remote object storage platform, but they are tracked using Git. Integration with DVC is possible through the `DVCFileSystem`. I have a Gitlab repository where multiple files are tracked using DVC and stored in a GCP bucket. I would like to be able to load these files using `datasets` directly using an URL. My goal is to write a generic code that abstracts the storage layer, such that my users will only have to pass in an `fsspec`-compliant URL and the corresponding files will be loaded. ### Your contribution I managed to instantiate a `DVCFileSystem` pointing to a Gitlab repo from a `fsspec` chained URL in [this pull request](https://github.com/iterative/dvc/pull/9903) to DVC. ```python from fsspec.core import url_to_fs fs, _ = url_to_fs("dvc::https://gitlab.com/repository/group/my-repo") ``` From now I'm not sure how to continue, it seems that `datasets` expects the URL to be fully qualified like so: `dvc::https://gitlab.com/repository/group/my-repo/my-folder/my-file.json` but this fails because `DVCFileSystem` expects the URL to point to the root of an SCM repo. Is there a way to make this work with `datasets`?
closed
https://github.com/huggingface/datasets/issues/6203
2023-09-01T14:04:52
2023-09-15T15:11:27
2023-09-15T15:11:27
{ "login": "bilelomrani1", "id": 16692099, "type": "User" }
[ { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,876,630,351
6,202
avoid downgrading jax version
### Feature request Whenever I `pip install datasets[jax]` it downgrades jax to version 0.3.25. I seem to be able to install this library first then upgrade jax back to version 0.4.13. ### Motivation It would be nice to not overwrite currently installed version of jax if possible. ### Your contribution I would be willing to beta test. Or maybe write some code if I could get pointed in the right direction, I'm not super familiar with this codebase.
closed
https://github.com/huggingface/datasets/issues/6202
2023-09-01T02:57:57
2023-10-12T16:28:59
2023-10-12T16:28:59
{ "login": "chrisflesher", "id": 1332458, "type": "User" }
[ { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,875,256,775
6,201
Fix to_json ValueError and remove pandas pin
This PR fixes the root cause of the issue: - #6197 This PR also removes the temporary pin of `pandas` introduced by: - #6200 Note that for orient in ['records', 'values'], index value is ignored but - in `pandas` < 2.1.0, a ValueError is raised if not index and orient not in ['split', 'table'] - for orient = 'records', we need index = True - default index value is True - in `pandas` = 2.1.0, a ValueError is raised if index is True and orient in ['records', 'values'] - for orient = 'records', we need index = False or None - default index value is None This PR fixes the issue by not passing index and thus using default index value (valid for all pandas versions), unless orient is 'split' or 'table' (where we pass index = False, as it was done before this fix).
closed
https://github.com/huggingface/datasets/pull/6201
2023-08-31T10:38:08
2023-09-05T11:07:07
2023-09-05T10:58:21
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[]
true
[]
1,875,169,551
6,200
Temporarily pin pandas < 2.1.0
Temporarily pin `pandas` < 2.1.0 until permanent solution is found. Hot fix #6197.
closed
https://github.com/huggingface/datasets/pull/6200
2023-08-31T09:45:17
2023-08-31T10:33:24
2023-08-31T10:24:38
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[]
true
[]
1,875,165,185
6,199
Use load_dataset for local json files, but it not works
### Describe the bug when I use load_dataset to load my local datasets,it always goes to Hugging Face to download the data instead of loading the local dataset. ### Steps to reproduce the bug `raw_datasets = load_dataset( ‘json’, data_files=data_files)` ### Expected behavior ![image](https://github.com/huggingface/datasets/assets/50519434/add3747f-6481-4da7-b374-8f81c5a6472c) ### Environment info python version 3.8.5 datasets version 2.12 os version unbuntu 18.04
open
https://github.com/huggingface/datasets/issues/6199
2023-08-31T09:42:34
2023-08-31T19:05:07
null
{ "login": "Garen-in-bush", "id": 50519434, "type": "User" }
[]
false
[]
1,875,092,027
6,198
Preserve split order in DataFilesDict
After investigation, I have found that this copy forces the splits to be sorted alphabetically: https://github.com/huggingface/datasets/blob/029227a116c14720afca71b9b22e78eb2a1c09a6/src/datasets/builder.py#L556 This PR removes the alphabetically sort of `DataFilesDict` keys. - Note that for a `dict`, the order of keys is relevant when hashing: ```python hash1 = Hasher.hash({'train': 'train.csv', 'test': 'test.csv'}) hash2 = Hasher.hash({'test': 'test.csv', 'train': 'train.csv'}) assert hash1 != hash2 ``` - The `DataFilesDict` is a subclass of `dict`, thus the order should be relevant as well ```python hash1 = Hasher.hash(DataFilesDict({'train': 'train.csv', 'test': 'test.csv'})) hash2 = Hasher.hash(DataFilesDict({'test': 'test.csv', 'train': 'train.csv'})) assert hash1 != hash2 ``` Fix #6196.
closed
https://github.com/huggingface/datasets/pull/6198
2023-08-31T09:00:26
2023-08-31T13:57:31
2023-08-31T13:48:42
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[]
true
[]
1,875,078,155
6,197
ValueError: 'index=True' is only valid when 'orient' is 'split', 'table', 'index', or 'columns'
### Describe the bug Saving a dataset `.to_json()` fails with a `ValueError` since the latest `pandas` [release](https://pandas.pydata.org/docs/dev/whatsnew/v2.1.0.html) (`2.1.0`) In their latest release we have: > Improved error handling when using [DataFrame.to_json()](https://pandas.pydata.org/docs/dev/reference/api/pandas.DataFrame.to_json.html#pandas.DataFrame.to_json) with incompatible index and orient arguments ([GH 52143](https://github.com/pandas-dev/pandas/issues/52143)) i.e. an error is now raised for invalid combinations of `index` and `orient`. This means that unfortunately the custom logic at this line might sometimes lead to contradictions: https://github.com/huggingface/datasets/blob/029227a116c14720afca71b9b22e78eb2a1c09a6/src/datasets/io/json.py#L96 e.g. for the default case `orient=records` leads to `index=True`, which now raises a `ValueError` ### Steps to reproduce the bug ```python import datasets if __name__ == '__main__': dataset = datasets.Dataset.from_dict({"A": [1, 2, 3], "B": [4, 5, 6]}) dataset.to_json("dataset.json") ``` ```shell >>> ValueError: 'index=True' is only valid when 'orient' is 'split', 'table', 'index', or 'columns'. ``` ### Expected behavior The dataset is successfully saved as `.json` ### Environment info `python >= 3.9` `pandas >= 2.1.0`
closed
https://github.com/huggingface/datasets/issues/6197
2023-08-31T08:51:50
2023-09-01T10:35:10
2023-08-31T10:24:40
{ "login": "exs-avianello", "id": 128361578, "type": "User" }
[]
false
[]
1,875,070,972
6,196
Split order is not preserved
I have noticed that in some cases the split order is not preserved. For example, consider a no-script dataset with configs: ```yaml configs: - config_name: default data_files: - split: train path: train.csv - split: test path: test.csv ``` - Note the defined split order is [train, test] Once the dataset is loaded, the split order is not preserved: ```python In [16]: ds Out[16]: DatasetDict({ test: Dataset({ features: ['text', 'label'], num_rows: 1 }) train: Dataset({ features: ['text', 'label'], num_rows: 2 }) }) ``` - Note the obtained split order is [test, train]
closed
https://github.com/huggingface/datasets/issues/6196
2023-08-31T08:47:16
2023-08-31T13:48:43
2023-08-31T13:48:43
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[ { "name": "bug", "color": "d73a4a" } ]
false
[]
1,874,195,585
6,195
Force to reuse cache at given path
### Describe the bug I have run the official example of MLM like: ```bash python run_mlm.py \ --model_name_or_path roberta-base \ --dataset_name togethercomputer/RedPajama-Data-1T \ --dataset_config_name arxiv \ --per_device_train_batch_size 10 \ --preprocessing_num_workers 20 \ --validation_split_percentage 0 \ --cache_dir /project/huggingface_cache/datasets \ --line_by_line \ --do_train \ --pad_to_max_length \ --output_dir /project/huggingface_cache/test-mlm ``` it successfully runs and at my cache folder has `cache-1982fea76aa54a13_00001_of_00020.arrow`..... `cache-1982fea76aa54a13_00020_of_00020.arrow ` as tokenization cache of `map` method. And the cache works fine every time I run the command above. However, when I switched to jupyter notebook (since I do not want to load datasets every time when I changed other parameters not related to the dataloading). It is not recognizing the cache files and starts to re-run the entire tokenization process. I changed my code to ```python tokenized_datasets = raw_datasets["train"].map( tokenize_function, batched=True, num_proc=data_args.preprocessing_num_workers, remove_columns=[text_column_name], load_from_cache_file=True, desc="Running tokenizer on dataset line_by_line", # cache_file_names= {"train": "cache-1982fea76aa54a13.arrow"} cache_file_name="cache-1982fea76aa54a13.arrow", new_fingerprint="1982fea76aa54a13" ) ``` it still does not recognize the previously cached files and trying to re-run the tokenization process. ### Steps to reproduce the bug use jupyter notebook for dataset map function. ### Expected behavior the map function accepts the given cache_file_name and new_fingerprint then load the previously cached files. ### Environment info - `datasets` version: 2.14.4.dev0 - Platform: Linux-3.10.0-1160.59.1.el7.x86_64-x86_64-with-glibc2.10 - Python version: 3.8.8 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
closed
https://github.com/huggingface/datasets/issues/6195
2023-08-30T18:44:54
2023-11-03T10:14:21
2023-08-30T19:00:45
{ "login": "Luosuu", "id": 43507393, "type": "User" }
[]
false
[]
1,872,598,223
6,194
Support custom fingerprinting with `Dataset.from_generator`
### Feature request When using `Dataset.from_generator`, the generator is hashed when building the fingerprint. Similar to `.map`, it would be interesting to let the user bypass this hashing by accepting a `fingerprint` argument to `.from_generator`. ### Motivation Using the `.from_generator` constructor with a non-picklable generator fails. By accepting a `fingerprint` argument to `.from_generator`, the user would have the opportunity to manually fingerprint the dataset and thus bypass the crash. ### Your contribution If validated, I can try to submit a PR for this.
open
https://github.com/huggingface/datasets/issues/6194
2023-08-29T22:43:13
2024-12-22T01:14:39
null
{ "login": "bilelomrani1", "id": 16692099, "type": "User" }
[ { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,872,285,153
6,193
Dataset loading script method does not work with .pyc file
### Describe the bug The huggingface dataset library specifically looks for ‘.py’ file while loading the dataset using loading script approach and it does not work with ‘.pyc’ file. While deploying in production, it becomes an issue when we are restricted to use only .pyc files. Is there any work around for this ? ### Steps to reproduce the bug 1. Create a dataset loading script to read the custom data. 2. compile the code to make sure that .pyc file is created 3. Delete the loading script and re-run the code. Usually, python should make use of complied .pyc files. However, in this case, the dataset library errors out with the message that it's unable to find the data loader loading script. ### Expected behavior The code should make use of .pyc file and run without any error. ### Environment info NA
open
https://github.com/huggingface/datasets/issues/6193
2023-08-29T19:35:06
2023-08-31T19:47:29
null
{ "login": "riteshkumarumassedu", "id": 43389071, "type": "User" }
[]
false
[]
1,871,911,640
6,192
Set minimal fsspec version requirement to 2023.1.0
Fix https://github.com/huggingface/datasets/issues/6141 Colab installs 2023.6.0, so we should be good 🙂
closed
https://github.com/huggingface/datasets/pull/6192
2023-08-29T15:23:41
2023-08-30T14:01:56
2023-08-30T13:51:32
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,871,634,840
6,191
Add missing `revision` argument
I've noticed that when you're not working on the main branch, there are sometimes errors in the files returned. After some investigation, I realized that the revision was not properly passed everywhere. This PR proposes a fix.
closed
https://github.com/huggingface/datasets/pull/6191
2023-08-29T13:05:04
2023-09-04T06:38:17
2023-08-31T13:50:00
{ "login": "qgallouedec", "id": 45557362, "type": "User" }
[]
true
[]
1,871,582,175
6,190
`Invalid user token` even when correct user token is passed!
### Describe the bug I'm working on a dataset which comprises other datasets on the hub. URL: https://huggingface.co/datasets/open-asr-leaderboard/datasets-test-only Note: Some of the sub-datasets in this metadataset require explicit access. All the other datasets work fine, except, `common_voice`. ### Steps to reproduce the bug https://github.com/Vaibhavs10/scratchpad/blob/main/cv_datasets_bug_repro.ipynb ### Expected behavior It should work if the provided access token is valid (as it does for all the other datasets) ### Environment info datasets version -> 2.14.4
closed
https://github.com/huggingface/datasets/issues/6190
2023-08-29T12:37:03
2023-08-29T13:01:10
2023-08-29T13:01:09
{ "login": "Vaibhavs10", "id": 18682411, "type": "User" }
[]
false
[]
1,871,569,855
6,189
Don't alter input in Features.from_dict
null
closed
https://github.com/huggingface/datasets/pull/6189
2023-08-29T12:29:47
2023-08-29T13:04:59
2023-08-29T12:52:48
{ "login": "lhoestq", "id": 42851186, "type": "User" }
[]
true
[]
1,870,987,640
6,188
[Feature Request] Check the length of batch before writing so that empty batch is allowed
### Use Case I use `dataset.map(process_fn, batched=True)` to process the dataset, with data **augmentations or filtering**. However, when all examples within a batch is filtered out, i.e. **an empty batch is returned**, the following error will be thrown: ``` ValueError: Schema and number of arrays unequal ``` This is because the empty batch does not comply with the schema of other batches. I think an empty batch should be allowed to facilitate coding (one does not need to assign an empty list manually for all keys.) A simple fix is to check the length of `batch` before writing: ``` if len(batch): writer.write_batch(batch) ``` instead of https://github.com/huggingface/datasets/blob/74d60213dcbd7c99484c62ce1d3dfd90a1df0770/src/datasets/arrow_dataset.py#L3493
closed
https://github.com/huggingface/datasets/issues/6188
2023-08-29T06:37:34
2023-09-19T21:55:38
2023-09-19T21:55:37
{ "login": "namespace-Pt", "id": 61188463, "type": "User" }
[]
false
[]
1,870,936,143
6,187
Couldn't find a dataset script at /content/tsv/tsv.py or any data file in the same directory
### Describe the bug ``` --------------------------------------------------------------------------- FileNotFoundError Traceback (most recent call last) [<ipython-input-48-6a7b3e847019>](https://localhost:8080/#) in <cell line: 7>() 5 } 6 ----> 7 csv_datasets_reloaded = load_dataset("tsv", data_files=data_files) 8 csv_datasets_reloaded 2 frames [/usr/local/lib/python3.10/dist-packages/datasets/load.py](https://localhost:8080/#) in dataset_module_factory(path, revision, download_config, download_mode, dynamic_modules_path, data_dir, data_files, **download_kwargs) 1489 raise e1 from None 1490 if isinstance(e1, FileNotFoundError): -> 1491 raise FileNotFoundError( 1492 f"Couldn't find a dataset script at {relative_to_absolute_path(combined_path)} or any data file in the same directory. " 1493 f"Couldn't find '{path}' on the Hugging Face Hub either: {type(e1).__name__}: {e1}" FileNotFoundError: Couldn't find a dataset script at /content/tsv/tsv.py or any data file in the same directory. Couldn't find 'tsv' on the Hugging Face Hub either: FileNotFoundError: Dataset 'tsv' doesn't exist on the Hub ``` ### Steps to reproduce the bug ``` data_files = { "train": "/content/PUBHEALTH/train.tsv", "validation": "/content/PUBHEALTH/dev.tsv", "test": "/content/PUBHEALTH/test.tsv", } tsv_datasets_reloaded = load_dataset("tsv", data_files=data_files) tsv_datasets_reloaded ``` ``` --------------------------------------------------------------------------- FileNotFoundError Traceback (most recent call last) <ipython-input-48-6a7b3e847019> in <cell line: 7>() 5 } 6 ----> 7 csv_datasets_reloaded = load_dataset("tsv", data_files=data_files) 8 csv_datasets_reloaded 2 frames /usr/local/lib/python3.10/dist-packages/datasets/load.py in dataset_module_factory(path, revision, download_config, download_mode, dynamic_modules_path, data_dir, data_files, **download_kwargs) 1489 raise e1 from None 1490 if isinstance(e1, FileNotFoundError): -> 1491 raise FileNotFoundError( 1492 f"Couldn't find a dataset script at {relative_to_absolute_path(combined_path)} or any data file in the same directory. " 1493 f"Couldn't find '{path}' on the Hugging Face Hub either: {type(e1).__name__}: {e1}" FileNotFoundError: Couldn't find a dataset script at /content/tsv/tsv.py or any data file in the same directory. Couldn't find 'tsv' on the Hugging Face Hub either: FileNotFoundError: Dataset 'tsv' doesn't exist on the Hub ``` ### Expected behavior load the data, push to hub ### Environment info jupyter notebook RTX 3090
open
https://github.com/huggingface/datasets/issues/6187
2023-08-29T05:49:56
2023-08-29T16:21:45
null
{ "login": "andysingal", "id": 20493493, "type": "User" }
[]
false
[]
1,869,431,457
6,186
Feature request: add code example of multi-GPU processing
### Feature request Would be great to add a code example of how to do multi-GPU processing with 🤗 Datasets in the documentation. cc @stevhliu Currently the docs has a small [section](https://huggingface.co/docs/datasets/v2.3.2/en/process#map) on this saying "your big GPU call goes here", however it didn't work for me out-of-the-box. Let's say you have a PyTorch model that can do translation, and you have multiple GPUs. In that case, you'd like to duplicate the model on each GPU, each processing (translating) a chunk of the data in parallel. Here's how I tried to do that: ``` from datasets import load_dataset from transformers import AutoModelForSeq2SeqLM, AutoTokenizer from multiprocess import set_start_method import torch import os dataset = load_dataset("mlfoundations/datacomp_small") tokenizer = AutoTokenizer.from_pretrained("facebook/nllb-200-distilled-600M") model = AutoModelForSeq2SeqLM.from_pretrained("facebook/nllb-200-distilled-600M") # put model on each available GPU # also, should I do it like this or use nn.DataParallel? model.to("cuda:0") model.to("cuda:1") set_start_method("spawn") def translate_captions(batch, rank): os.environ["CUDA_VISIBLE_DEVICES"] = str(rank % torch.cuda.device_count()) texts = batch["text"] inputs = tokenizer(texts, padding=True, truncation=True, return_tensors="pt").to(model.device) translated_tokens = model.generate( **inputs, forced_bos_token_id=tokenizer.lang_code_to_id["eng_Latn"], max_length=30 ) translated_texts = tokenizer.batch_decode(translated_tokens, skip_special_tokens=True) batch["translated_text"] = translated_texts return batch updated_dataset = dataset.map(translate_captions, with_rank=True, num_proc=2, batched=True, batch_size=256) ``` I've personally tried running this script on a machine with 2 A100 GPUs. ## Error 1 Running the code snippet above from the terminal (python script.py) resulted in the following error: ``` Traceback (most recent call last): File "<string>", line 1, in <module> File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/spawn.py", line 116, in spawn_main exitcode = _main(fd, parent_sentinel) File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/spawn.py", line 125, in _main prepare(preparation_data) File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/spawn.py", line 236, in prepare _fixup_main_from_path(data['init_main_from_path']) File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/spawn.py", line 287, in _fixup_main_from_path main_content = runpy.run_path(main_path, File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/runpy.py", line 289, in run_path return _run_module_code(code, init_globals, run_name, File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/runpy.py", line 96, in _run_module_code _run_code(code, mod_globals, init_globals, File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/runpy.py", line 86, in _run_code exec(code, run_globals) File "/home/niels/python_projects/datacomp/datasets_multi_gpu.py", line 16, in <module> set_start_method("spawn") File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/context.py", line 247, in set_start_method raise RuntimeError('context has already been set') RuntimeError: context has already been set ``` ## Error 2 Then, based on [this Stackoverflow answer](https://stackoverflow.com/a/71616344/7762882), I put the `set_start_method("spawn")` section in a try: catch block. This resulted in the following error: ``` File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/datasets/dataset_dict.py", line 817, in <dictcomp> k: dataset.map( File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 2926, in map with Pool(nb_of_missing_shards, initargs=initargs, initializer=initializer) as pool: File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/context.py", line 119, in Pool return Pool(processes, initializer, initargs, maxtasksperchild, File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/pool.py", line 215, in __init__ self._repopulate_pool() File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/pool.py", line 306, in _repopulate_pool return self._repopulate_pool_static(self._ctx, self.Process, File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/pool.py", line 329, in _repopulate_pool_static w.start() File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/process.py", line 121, in start self._popen = self._Popen(self) File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/context.py", line 288, in _Popen return Popen(process_obj) File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/popen_spawn_posix.py", line 32, in __init__ super().__init__(process_obj) File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/popen_fork.py", line 19, in __init__ self._launch(process_obj) File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/popen_spawn_posix.py", line 42, in _launch prep_data = spawn.get_preparation_data(process_obj._name) File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/spawn.py", line 154, in get_preparation_data _check_not_importing_main() File "/home/niels/anaconda3/envs/datacomp/lib/python3.10/site-packages/multiprocess/spawn.py", line 134, in _check_not_importing_main raise RuntimeError(''' RuntimeError: An attempt has been made to start a new process before the current process has finished its bootstrapping phase. This probably means that you are not using fork to start your child processes and you have forgotten to use the proper idiom in the main module: if __name__ == '__main__': freeze_support() ... The "freeze_support()" line can be omitted if the program is not going to be frozen to produce an executable. ``` So then I put the last line under a `if __name__ == '__main__':` block. Then the code snippet seemed to work, but it seemed that it's only leveraging a single GPU (based on monitoring `nvidia-smi`): ``` Mon Aug 28 12:19:24 2023 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 515.65.01 Driver Version: 515.65.01 CUDA Version: 11.7 | |-------------------------------+----------------------+----------------------+ | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVIDIA A100-SXM... On | 00000000:01:00.0 Off | 0 | | N/A 55C P0 76W / 275W | 8747MiB / 81920MiB | 0% Default | | | | Disabled | +-------------------------------+----------------------+----------------------+ | 1 NVIDIA A100-SXM... On | 00000000:47:00.0 Off | 0 | | N/A 67C P0 274W / 275W | 59835MiB / 81920MiB | 100% Default | | | | Disabled | ``` Both GPUs should have equal GPU usage, but I've always noticed that the last GPU has way more usage than the other ones. This made me think that `os.environ["CUDA_VISIBLE_DEVICES"] = str(rank % torch.cuda.device_count())` might not work inside a Python script, especially if done after importing PyTorch? ### Motivation Would be great to clarify how to do multi-GPU data processing. ### Your contribution If my code snippet can be fixed, I can contribute it to the docs :)
closed
https://github.com/huggingface/datasets/issues/6186
2023-08-28T10:00:59
2024-10-07T09:39:51
2023-11-22T15:42:20
{ "login": "NielsRogge", "id": 48327001, "type": "User" }
[ { "name": "documentation", "color": "0075ca" }, { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,868,077,748
6,185
Error in saving the PIL image into *.arrow files using datasets.arrow_writer
### Describe the bug I am using the ArrowWriter from datasets.arrow_writer to save a json-style file as arrow files. Within the dictionary, it contains a feature called "image" which is a list of PIL.Image objects. I am saving the json using the following script: ``` def save_to_arrow(path,temp): with ArrowWriter(path=path,writer_batch_size=20) as writer: writer.write_batch(temp) writer.finalize() ``` However, when I attempt to restore the dataset and use the ```Dataset.from_file(path)``` function to load the arrow file, there seems to be an issue with the PIL.Image object in the dataset. The list of PIL.Images appears as follows rather than a normal PIL.Image object: ![1693051705440](https://github.com/huggingface/datasets/assets/14247682/03b204c2-d0fa-4d19-beff-6f4d7b83c848) ### Steps to reproduce the bug 1. Storing the data json into arrow files: ``` def save_to_arrow(path,temp): with ArrowWriter(path=path,writer_batch_size=20) as writer: writer.write_batch(temp) writer.finalize() save_to_arrow( path, json_file ) ``` 2. try to load the arrow file into the Dataset object using the ```Dataset.from_file(path)``` ### Expected behavior Except to saving the contained "image" feature as a list PIL.Image objects as the arrow file. And I can restore the dataset from the file. ### Environment info - `datasets` version: 2.12.0 - Platform: Linux-5.4.0-150-generic-x86_64-with-glibc2.17 - Python version: 3.8.17 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 1.4.4
open
https://github.com/huggingface/datasets/issues/6185
2023-08-26T12:15:57
2023-08-29T14:49:58
null
{ "login": "HaozheZhao", "id": 14247682, "type": "User" }
[]
false
[]
1,867,766,143
6,184
Map cache does not detect function changes in another module
```python # dataset.py import os import datasets if not os.path.exists('/tmp/test.json'): with open('/tmp/test.json', 'w') as file: file.write('[{"text": "hello"}]') def transform(example): text = example['text'] # text += ' world' return {'text': text} data = datasets.load_dataset('json', data_files=['/tmp/test.json'], split='train') data = data.map(transform) ``` ```python # test.py import dataset print(next(iter(dataset.data))) ``` Initialize cache ``` python3 test.py # {'text': 'hello'} ``` Edit dataset.py and uncomment the commented line, run again ``` python3 test.py # {'text': 'hello'} # expected: {'text': 'hello world'} ``` Clear cache and run again ``` rm -rf ~/.cache/huggingface/datasets/* python3 test.py # {'text': 'hello world'} ``` If instead the two files are combined, then changes to the function are detected correctly. But it's expected when working on any realistic codebase that things will be modularized into separate files.
closed
https://github.com/huggingface/datasets/issues/6184
2023-08-25T22:59:14
2023-08-29T20:57:07
2023-08-29T20:56:49
{ "login": "jonathanasdf", "id": 511073, "type": "User" }
[ { "name": "duplicate", "color": "cfd3d7" } ]
false
[]
1,867,743,276
6,183
Load dataset with non-existent file
### Describe the bug When load a dataset from datasets and pass a wrong path to json with the data, error message does not contain something abount "wrong path" or "file do not exist" - ```SchemaInferenceError: Please pass `features` or at least one example when writing data``` ### Steps to reproduce the bug ```python from datasets import load_dataset load_dataset('json', data_files='/home/alexey/unreal_file.json') ``` ### Expected behavior Raise os FileNotFound error or custom error with informative message ### Environment info ``` # packages in environment at /home/alexey/.conda/envs/alex_LoRA: # # Name Version Build Channel _libgcc_mutex 0.1 main _openmp_mutex 5.1 1_gnu accelerate 0.21.0 pypi_0 pypi aiohttp 3.8.5 pypi_0 pypi aiosignal 1.3.1 pypi_0 pypi antlr4-python3-runtime 4.9.3 pypi_0 pypi appdirs 1.4.4 pypi_0 pypi asttokens 2.0.5 pyhd3eb1b0_0 async-timeout 4.0.3 pypi_0 pypi attrs 23.1.0 pypi_0 pypi backcall 0.2.0 pyhd3eb1b0_0 bitsandbytes 0.41.1 pypi_0 pypi bzip2 1.0.8 h7b6447c_0 ca-certificates 2023.05.30 h06a4308_0 certifi 2023.7.22 pypi_0 pypi charset-normalizer 3.2.0 pypi_0 pypi click 8.1.6 pypi_0 pypi cmake 3.27.2 pypi_0 pypi comm 0.1.2 py310h06a4308_0 contourpy 1.1.0 pypi_0 pypi cycler 0.11.0 pypi_0 pypi datasets 2.14.4 pypi_0 pypi debugpy 1.6.7 py310h6a678d5_0 decorator 5.1.1 pyhd3eb1b0_0 dill 0.3.7 pypi_0 pypi docker-pycreds 0.4.0 pypi_0 pypi executing 0.8.3 pyhd3eb1b0_0 filelock 3.12.2 pypi_0 pypi fire 0.5.0 pypi_0 pypi fonttools 4.42.0 pypi_0 pypi frozenlist 1.4.0 pypi_0 pypi fsspec 2023.6.0 pypi_0 pypi gitdb 4.0.10 pypi_0 pypi gitpython 3.1.32 pypi_0 pypi huggingface-hub 0.16.4 pypi_0 pypi idna 3.4 pypi_0 pypi ipykernel 6.25.0 py310h2f386ee_0 ipython 8.12.2 py310h06a4308_0 ipython-genutils 0.2.0 pypi_0 pypi ipywidgets 8.0.4 py310h06a4308_0 jedi 0.18.1 py310h06a4308_1 jinja2 3.1.2 pypi_0 pypi jsonschema 4.19.0 pypi_0 pypi jsonschema-specifications 2023.7.1 pypi_0 pypi jupyter_client 8.1.0 py310h06a4308_0 jupyter_core 5.3.0 py310h06a4308_0 jupyterlab_widgets 3.0.5 py310h06a4308_0 kiwisolver 1.4.4 pypi_0 pypi ld_impl_linux-64 2.38 h1181459_1 libffi 3.3 he6710b0_2 libgcc-ng 11.2.0 h1234567_1 libgomp 11.2.0 h1234567_1 libsodium 1.0.18 h7b6447c_0 libstdcxx-ng 11.2.0 h1234567_1 libuuid 1.41.5 h5eee18b_0 lightning-utilities 0.9.0 pypi_0 pypi lit 16.0.6 pypi_0 pypi markupsafe 2.1.3 pypi_0 pypi matplotlib 3.7.2 pypi_0 pypi matplotlib-inline 0.1.6 py310h06a4308_0 mpmath 1.3.0 pypi_0 pypi multidict 6.0.4 pypi_0 pypi multiprocess 0.70.15 pypi_0 pypi nbformat 4.2.0 pypi_0 pypi ncurses 6.4 h6a678d5_0 nest-asyncio 1.5.6 py310h06a4308_0 networkx 3.1 pypi_0 pypi numpy 1.25.2 pypi_0 pypi nvidia-cublas-cu11 11.10.3.66 pypi_0 pypi nvidia-cuda-cupti-cu11 11.7.101 pypi_0 pypi nvidia-cuda-nvrtc-cu11 11.7.99 pypi_0 pypi nvidia-cuda-runtime-cu11 11.7.99 pypi_0 pypi nvidia-cudnn-cu11 8.5.0.96 pypi_0 pypi nvidia-cufft-cu11 10.9.0.58 pypi_0 pypi nvidia-curand-cu11 10.2.10.91 pypi_0 pypi nvidia-cusolver-cu11 11.4.0.1 pypi_0 pypi nvidia-cusparse-cu11 11.7.4.91 pypi_0 pypi nvidia-nccl-cu11 2.14.3 pypi_0 pypi nvidia-nvtx-cu11 11.7.91 pypi_0 pypi omegaconf 2.3.0 pypi_0 pypi openssl 1.1.1v h7f8727e_0 packaging 23.0 py310h06a4308_0 pandas 2.0.3 pypi_0 pypi parso 0.8.3 pyhd3eb1b0_0 pathtools 0.1.2 pypi_0 pypi peft 0.4.0 pypi_0 pypi pexpect 4.8.0 pyhd3eb1b0_3 pickleshare 0.7.5 pyhd3eb1b0_1003 pillow 10.0.0 pypi_0 pypi pip 23.2.1 py310h06a4308_0 platformdirs 2.5.2 py310h06a4308_0 plotly 5.16.1 pypi_0 pypi prompt-toolkit 3.0.36 py310h06a4308_0 protobuf 4.24.0 pypi_0 pypi psutil 5.9.0 py310h5eee18b_0 ptyprocess 0.7.0 pyhd3eb1b0_2 pure_eval 0.2.2 pyhd3eb1b0_0 pyarrow 12.0.1 pypi_0 pypi pygments 2.15.1 py310h06a4308_1 pyparsing 3.0.9 pypi_0 pypi python 3.10.0 h12debd9_5 python-dateutil 2.8.2 pyhd3eb1b0_0 pytorch-lightning 2.0.6 pypi_0 pypi pytz 2023.3 pypi_0 pypi pyyaml 6.0.1 pypi_0 pypi pyzmq 25.1.0 py310h6a678d5_0 readline 8.2 h5eee18b_0 referencing 0.30.2 pypi_0 pypi regex 2023.8.8 pypi_0 pypi requests 2.31.0 pypi_0 pypi rpds-py 0.9.2 pypi_0 pypi safetensors 0.3.2 pypi_0 pypi scipy 1.11.1 pypi_0 pypi sentencepiece 0.1.99 pypi_0 pypi sentry-sdk 1.29.2 pypi_0 pypi setproctitle 1.3.2 pypi_0 pypi setuptools 68.0.0 py310h06a4308_0 six 1.16.0 pyhd3eb1b0_1 smmap 5.0.0 pypi_0 pypi sqlite 3.41.2 h5eee18b_0 stack_data 0.2.0 pyhd3eb1b0_0 sympy 1.12 pypi_0 pypi tenacity 8.2.3 pypi_0 pypi termcolor 2.3.0 pypi_0 pypi tk 8.6.12 h1ccaba5_0 tokenizers 0.13.3 pypi_0 pypi torch 2.0.1 pypi_0 pypi torchmetrics 1.0.3 pypi_0 pypi tornado 6.3.2 py310h5eee18b_0 tqdm 4.66.1 pypi_0 pypi traitlets 5.7.1 py310h06a4308_0 transformers 4.31.0 pypi_0 pypi triton 2.0.0 pypi_0 pypi typing-extensions 4.7.1 pypi_0 pypi tzdata 2023.3 pypi_0 pypi urllib3 2.0.4 pypi_0 pypi wandb 0.15.8 pypi_0 pypi wcwidth 0.2.5 pyhd3eb1b0_0 wheel 0.38.4 py310h06a4308_0 widgetsnbextension 4.0.5 py310h06a4308_0 xxhash 3.3.0 pypi_0 pypi xz 5.4.2 h5eee18b_0 yarl 1.9.2 pypi_0 pypi zeromq 4.3.4 h2531618_0 zlib 1.2.13 h5eee18b_0 active environment : None user config file : /home/alexey/.condarc populated config files : conda version : 23.1.0 conda-build version : 3.22.0 python version : 3.9.13.final.0 virtual packages : __archspec=1=x86_64 __cuda=12.0=0 __glibc=2.35=0 __linux=5.19.0=0 __unix=0=0 base environment : /opt/anaconda/anaconda3 (read only) conda av data dir : /opt/anaconda/anaconda3/etc/conda conda av metadata url : None channel URLs : https://repo.anaconda.com/pkgs/main/linux-64 https://repo.anaconda.com/pkgs/main/noarch https://repo.anaconda.com/pkgs/r/linux-64 https://repo.anaconda.com/pkgs/r/noarch package cache : /opt/anaconda/anaconda3/pkgs /home/alexey/.conda/pkgs envs directories : /home/alexey/.conda/envs /opt/anaconda/anaconda3/envs platform : linux-64 user-agent : conda/23.1.0 requests/2.31.0 CPython/3.9.13 Linux/5.19.0-46-generic ubuntu/22.04.2 glibc/2.35 UID:GID : 1009:1009 netrc file : /home/alexey/.netrc offline mode : False ```
closed
https://github.com/huggingface/datasets/issues/6183
2023-08-25T22:21:22
2023-08-29T13:26:22
2023-08-29T13:26:22
{ "login": "freQuensy23-coder", "id": 64750224, "type": "User" }
[]
false
[]
1,867,203,131
6,182
Loading Meteor metric in HF evaluate module crashes due to datasets import issue
### Describe the bug When using python3.9 and ```evaluate``` module loading Meteor metric crashes at a non-existent import from ```datasets.config``` in ```datasets v2.14``` ### Steps to reproduce the bug ``` from evaluate import load meteor = load("meteor") ``` produces the following error: ``` from datasets.config import importlib_metadata, version ImportError: cannot import name 'importlib_metadata' from 'datasets.config' (<path_to_project>/venv/lib/python3.9/site-packages/datasets/config.py) ``` ### Expected behavior ```datasets``` of v2.10 has the following workaround in ```config.py```: ``` if PY_VERSION < version.parse("3.8"): import importlib_metadata else: import importlib.metadata as importlib_metadata ``` However, it's absent in v2.14 which might be the cause of the issue. ### Environment info - `datasets` version: 2.14.4 - Platform: macOS-13.5-arm64-arm-64bit - Python version: 3.9.6 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3 - Evaluate version: 0.4.0
closed
https://github.com/huggingface/datasets/issues/6182
2023-08-25T14:54:06
2023-09-04T16:41:11
2023-08-31T14:38:23
{ "login": "dsashulya", "id": 42322648, "type": "User" }
[]
false
[]
1,867,035,522
6,181
Fix import in `image_load` doc
Reported on [Discord](https://discord.com/channels/879548962464493619/1144295822209581168/1144295822209581168)
closed
https://github.com/huggingface/datasets/pull/6181
2023-08-25T13:12:19
2023-08-25T16:12:46
2023-08-25T16:02:24
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,867,032,578
6,180
Use `hf-internal-testing` repos for hosting test dataset repos
Use `hf-internal-testing` for hosting instead of the maintainers' dataset repos.
closed
https://github.com/huggingface/datasets/pull/6180
2023-08-25T13:10:26
2023-08-25T16:58:02
2023-08-25T16:46:22
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,867,009,016
6,179
Map cache with tokenizer
Similar issue to https://github.com/huggingface/datasets/issues/5985, but across different sessions rather than two calls in the same session. Unlike that issue, explicitly calling tokenizer(my_args) before the map() doesn't help, because the tokenizer was created with a different hash to begin with... setup ``` from transformers import AutoTokenizer AutoTokenizer.from_pretrained('bert-base-uncased').save_pretrained("tok") ``` this prints different value each time ``` from transformers import AutoTokenizer from datasets.utils.py_utils import dumps # Huggingface datasets print(hash(dumps(AutoTokenizer.from_pretrained("tok")))) ```
open
https://github.com/huggingface/datasets/issues/6179
2023-08-25T12:55:18
2023-08-31T15:17:24
null
{ "login": "jonathanasdf", "id": 511073, "type": "User" }
[]
false
[]
1,866,610,102
6,178
'import datasets' throws "invalid syntax error"
### Describe the bug Hi, I have been trying to import the datasets library but I keep gtting this error. `Traceback (most recent call last): File /opt/local/jupyterhub/lib64/python3.9/site-packages/IPython/core/interactiveshell.py:3508 in run_code exec(code_obj, self.user_global_ns, self.user_ns) Cell In[2], line 1 import datasets File /opt/local/jupyterhub/lib64/python3.9/site-packages/datasets/__init__.py:22 from .arrow_dataset import Dataset File /opt/local/jupyterhub/lib64/python3.9/site-packages/datasets/arrow_dataset.py:67 from .arrow_writer import ArrowWriter, OptimizedTypedSequence File /opt/local/jupyterhub/lib64/python3.9/site-packages/datasets/arrow_writer.py:27 from .features import Features, Image, Value File /opt/local/jupyterhub/lib64/python3.9/site-packages/datasets/features/__init__.py:17 from .audio import Audio File /opt/local/jupyterhub/lib64/python3.9/site-packages/datasets/features/audio.py:11 from ..download.streaming_download_manager import xopen, xsplitext File /opt/local/jupyterhub/lib64/python3.9/site-packages/datasets/download/__init__.py:10 from .streaming_download_manager import StreamingDownloadManager File /opt/local/jupyterhub/lib64/python3.9/site-packages/datasets/download/streaming_download_manager.py:18 from aiohttp.client_exceptions import ClientError File /opt/local/jupyterhub/lib64/python3.9/site-packages/aiohttp/__init__.py:7 from .connector import * # noqa File /opt/local/jupyterhub/lib64/python3.9/site-packages/aiohttp/connector.py:12 from .client import ClientRequest File /opt/local/jupyterhub/lib64/python3.9/site-packages/aiohttp/client.py:144 yield from asyncio.async(resp.release(), loop=loop) ^ SyntaxError: invalid syntax` I have simply used these commands: `import datasets` and `from datasets import load_dataset` ### Environment info The library has been installed a virtual machine on JupyterHub. Although I have used this library multiple times (on the same VM) before, to train/test an ASR or other ML models, I had never encountered this error.
closed
https://github.com/huggingface/datasets/issues/6178
2023-08-25T08:35:14
2023-09-27T17:33:39
2023-09-27T17:33:39
{ "login": "elia-ashraf", "id": 128580829, "type": "User" }
[]
false
[]
1,865,490,962
6,177
Use object detection images from `huggingface/documentation-images`
null
closed
https://github.com/huggingface/datasets/pull/6177
2023-08-24T16:16:09
2023-08-25T16:30:00
2023-08-25T16:21:17
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,864,436,408
6,176
how to limit the size of memory mapped file?
### Describe the bug Huggingface datasets use memory-mapped file to map large datasets in memory for fast access. However, it seems like huggingface will occupy all the memory for memory-mapped files, which makes a troublesome situation since we cluster will distribute a small portion of memory to me (once it's over the limit, memory cannot be allocated), however, when the dataset checks the total memory, all of the memory will be taken into account which makes huggingface dataset try to allocate more memory than allowed. So is there a way to explicitly limit the size of memory mapped file? ### Steps to reproduce the bug python >>> from datasets import load_dataset >>> dataset = load_dataset("c4", "en", streaming=True) ### Expected behavior In a normal environment, this will not have any problem. However, when the system allocates a portion of the memory to the program and when the dataset checks the total memory, all of the memory will be taken into account which makes huggingface dataset try to allocate more memory than allowed. ### Environment info linux cluster with SGE(Sun Grid Engine)
open
https://github.com/huggingface/datasets/issues/6176
2023-08-24T05:33:45
2023-10-11T06:00:10
null
{ "login": "williamium3000", "id": 47763855, "type": "User" }
[]
false
[]
1,863,592,678
6,175
PyArrow 13 CI fixes
Fixes: * bumps the PyArrow version check in the `cast_array_to_feature` to avoid the offset bug (still not fixed) * aligns the Pandas formatting tests with the Numpy ones (the current test fails due to https://github.com/apache/arrow/pull/35656, which requires `.to_pandas(coerce_temporal_nanoseconds=True)` to always return `datetime [ns]` objects) Fix #6173
closed
https://github.com/huggingface/datasets/pull/6175
2023-08-23T15:45:53
2023-08-25T13:15:59
2023-08-25T13:06:52
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,863,422,065
6,173
Fix CI for pyarrow 13.0.0
pyarrow 13.0.0 just came out ``` FAILED tests/test_formatting.py::ArrowExtractorTest::test_pandas_extractor - AssertionError: Attributes of Series are different Attribute "dtype" are different [left]: datetime64[us, UTC] [right]: datetime64[ns, UTC] ``` ``` FAILED tests/test_table.py::test_cast_sliced_fixed_size_array_to_features - TypeError: Couldn't cast array of type fixed_size_list<item: int32>[3] to Sequence(feature=Value(dtype='int64', id=None), length=3, id=None) ``` e.g. in https://github.com/huggingface/datasets/actions/runs/5952253963/job/16143847230 first error may be related to https://github.com/apache/arrow/issues/33321 second one maybe because `feature.length * len(array) == len(array_values)` is not satisfied anymore somehow ?
closed
https://github.com/huggingface/datasets/issues/6173
2023-08-23T14:11:20
2023-08-25T13:06:53
2023-08-25T13:06:53
{ "login": "lhoestq", "id": 42851186, "type": "User" }
[]
false
[]
1,863,318,027
6,172
Make Dataset streaming queries retryable
### Feature request Streaming datasets, as intended, do not load the entire dataset in memory or disk. However, while querying the next data chunk from the remote, sometimes it is possible that the service is down or there might be other issues that may cause the query to fail. In such a scenario, it would be nice to make these queries retryable (perhaps with a backoff strategy). ### Motivation I was working on a model and the model checkpoints after every 1000 steps. At step 1800 I got a 504 HTTP status code error from Huggingface hub for my pytorch `dataloader`. Given the size of my model and data, it took around 2 hours to reach 1800 steps and now it will take about an hour to recover the lost 800. It would be better to get a retryable querying strategy. ### Your contribution It would be better if someone having experience in this area takes this up as this would require some testing.
open
https://github.com/huggingface/datasets/issues/6172
2023-08-23T13:15:38
2023-11-06T13:54:16
null
{ "login": "rojagtap", "id": 42299342, "type": "User" }
[ { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,862,922,767
6,171
Fix typo in about_mapstyle_vs_iterable.mdx
null
closed
https://github.com/huggingface/datasets/pull/6171
2023-08-23T09:21:11
2023-08-23T09:32:59
2023-08-23T09:21:19
{ "login": "lhoestq", "id": 42851186, "type": "User" }
[]
true
[]
1,862,705,731
6,170
feat: Return the name of the currently loaded file
Added an optional parameter return_file_name in the load_dataset function. When it is set to True, the function will include the name of the file corresponding to the current line as a feature in the returned output. I added this here https://github.com/huggingface/datasets/blob/main/src/datasets/packaged_modules/json/json.py#L92. fixes #5806
open
https://github.com/huggingface/datasets/pull/6170
2023-08-23T07:08:17
2023-08-29T12:41:05
null
{ "login": "Amitesh-Patel", "id": 124021133, "type": "User" }
[]
true
[]
1,862,360,199
6,169
Configurations in yaml not working
### Dataset configurations cannot be created in YAML/README Hello! I'm trying to follow the docs here in order to create structure in my dataset as added from here (#5331): https://github.com/huggingface/datasets/blob/8b8e6ee067eb74e7965ca2a6768f15f9398cb7c8/docs/source/repository_structure.mdx#L110-L118 I have the exact example in my config file for [my data repo](https://huggingface.co/datasets/tsor13/test): ``` configs: - config_name: main_data data_files: "main_data.csv" - config_name: additional_data data_files: "additional_data.csv" ``` Yet, I'm unable to load different configurations: ``` from datasets import get_dataset_config_names get_dataset_config_names('tsor13/test', use_auth_token=True) ``` returns a single split, `['tsor13--test']` Does anyone have any insights? @polinaeterna thank you for adding this feature, it is super useful. Do you happen to have any ideas? ### Steps to reproduce the bug from datasets import get_dataset_config_names get_dataset_config_names('tsor13/test') ### Expected behavior I would expect there to be two splits, `main_data` and `additional_data`. However, only `['tsor13--test']` test is returned. ### Environment info - `datasets` version: 2.14.4 - Platform: macOS-13.4-arm64-arm-64bit - Python version: 3.11.4 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 1.5.1
open
https://github.com/huggingface/datasets/issues/6169
2023-08-23T00:13:22
2023-08-23T15:35:31
null
{ "login": "tsor13", "id": 45085098, "type": "User" }
[]
false
[]
1,861,867,274
6,168
Fix ArrayXD YAML conversion
Replace the `shape` tuple with a list in the `ArrayXD` YAML conversion. Fix #6112
closed
https://github.com/huggingface/datasets/pull/6168
2023-08-22T17:02:54
2023-12-12T15:06:59
2023-12-12T15:00:43
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,861,474,327
6,167
Allow hyphen in split name
To fix https://discuss.huggingface.co/t/error-when-setting-up-the-dataset-viewer-streamingrowserror/51276.
closed
https://github.com/huggingface/datasets/pull/6167
2023-08-22T13:30:59
2024-01-11T06:31:31
2023-08-22T15:38:53
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,861,259,055
6,166
Document BUILDER_CONFIG_CLASS
Related to https://github.com/huggingface/datasets/issues/6130
closed
https://github.com/huggingface/datasets/pull/6166
2023-08-22T11:27:41
2023-08-23T14:01:25
2023-08-23T13:52:36
{ "login": "lhoestq", "id": 42851186, "type": "User" }
[]
true
[]
1,861,124,284
6,165
Fix multiprocessing with spawn in iterable datasets
The "Spawn" method is preferred when multiprocessing on macOS or Windows systems, instead of the "Fork" method on linux systems. This causes some methods of Iterable Datasets to break when using a dataloader with more than 0 workers. I fixed the issue by replacing lambda and local methods which are not pickle-able. See the example below: ```python from datasets import load_dataset from torch.utils.data import DataLoader if __name__ == "__main__": dataset = load_dataset("lhoestq/demo1", split="train") dataset = dataset.to_iterable_dataset(num_shards=3) dataset = dataset.remove_columns(["package_name"]) dataset = dataset.rename_columns({ "review": "review1" }) dataset = dataset.rename_column("date", "date1") for sample in DataLoader(dataset, batch_size=None, num_workers=3): print(sample) ``` To notice the fix on a linux system, adding these lines should do the trick: ```python import multiprocessing multiprocessing.set_start_method('spawn') ``` I also removed what looks like code duplication between rename_colums and rename_column
closed
https://github.com/huggingface/datasets/pull/6165
2023-08-22T10:07:23
2023-08-29T13:27:14
2023-08-29T13:18:11
{ "login": "bruno-hays", "id": 48770768, "type": "User" }
[]
true
[]
1,859,560,007
6,164
Fix: Missing a MetadataConfigs init when the repo has a `datasets_info.json` but no README
When I try to push to an arrow repo (can provide the link on Slack), it uploads the files but fails to update the metadata, with ``` File "app.py", line 123, in add_new_eval eval_results[level].push_to_hub(my_repo, token=TOKEN, split=SPLIT) File "blabla_my_env_path/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 5501, in push_to_hub if not metadata_configs: UnboundLocalError: local variable 'metadata_configs' referenced before assignment ``` This fixes it.
closed
https://github.com/huggingface/datasets/pull/6164
2023-08-21T14:57:54
2023-08-21T16:27:05
2023-08-21T16:18:26
{ "login": "clefourrier", "id": 22726840, "type": "User" }
[]
true
[]
1,857,682,241
6,163
Error type: ArrowInvalid Details: Failed to parse string: '[254,254]' as a scalar of type int32
### Describe the bug I am getting the following error while I am trying to upload the CSV sheet to train a model. My CSV sheet content is exactly same as shown in the example CSV file in the Auto Train page. Attaching screenshot of error for reference. I have also tried converting the index of the answer that are integer into string by placing inverted commas and also without inverted commas. Can anyone please help me out? FYI : I am using Chrome browser. Error type: ArrowInvalid Details: Failed to parse string: '[254,254]' as a scalar of type int32 ![Screenshot 2023-08-19 165827](https://github.com/huggingface/datasets/assets/90616801/95fad96e-7dce-4bb5-9f83-9f1659a32891) ### Steps to reproduce the bug Kindly let me know how to fix this? ### Expected behavior Kindly let me know how to fix this? ### Environment info Kindly let me know how to fix this?
open
https://github.com/huggingface/datasets/issues/6163
2023-08-19T11:34:40
2025-07-22T12:04:46
null
{ "login": "shishirCTC", "id": 90616801, "type": "User" }
[]
false
[]
1,856,198,342
6,162
load_dataset('json',...) from togethercomputer/RedPajama-Data-1T errors when jsonl rows contains different data fields
### Describe the bug When loading some jsonl from redpajama-data-1T github source [togethercomputer/RedPajama-Data-1T](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T) fails due to one row of the file containing an extra field called **symlink_target: string>**. When deleting that line the loading is successful. We also tried loading this file with the discrepancy using this function and it is successful ```python os.environ["RED_PAJAMA_DATA_DIR"] ="/path_to_local_copy_of_RedPajama-Data-1T" ds = load_dataset('togethercomputer/RedPajama-Data-1T', 'github',cache_dir="/path_to_folder_with_jsonl",streaming=True)['train'] ``` ### Steps to reproduce the bug Steps to reproduce the behavior: 1. Load one jsonl from the redpajama-data-1T ```bash wget https://data.together.xyz/redpajama-data-1T/v1.0.0/github/filtered_27f05c041a1c401783f90b9415e40e4b.sampled.jsonl ``` 2.Load dataset will give error: ```python from datasets import load_dataset ds = load_dataset('json', data_files='/path_to/filtered_27f05c041a1c401783f90b9415e40e4b.sampled.jsonl') ``` _TypeError: Couldn't cast array of type Struct <content_hash: string, timestamp: string, source: string, line_count: int64, max_line_length: int64, avg_line_length: double, alnum_prop: double, repo_name: string, id: string, size: string, binary: bool, copies: string, ref: string, path: string, mode: string, license: string, language: list<item: struct<name: string, bytes: string>>, **symlink_target: string>** to {'content_hash': Value(dtype='string', id=None), 'timestamp': Value(dtype='string', id=None), 'source': Value(dtype='string', id=None), 'line_count': Value(dtype='int64', id=None), 'max_line_length': Value(dtype='int64', id=None), 'avg_line_length': Value(dtype='float64', id=None), 'alnum_prop': Value(dtype='float64', id=None), 'repo_name': Value(dtype='string', id=None), 'id': Value(dtype='string', id=None), 'size': Value(dtype='string', id=None), 'binary': Value(dtype='bool', id=None), 'copies': Value(dtype='string', id=None), 'ref': Value(dtype='string', id=None), 'path': Value(dtype='string', id=None), 'mode': Value(dtype='string', id=None), 'license': Value(dtype='string', id=None), 'language': [{'name': Value(dtype='string', id=None), 'bytes': Value(dtype='string', id=None)}]}_ 3. To remove the line causing the problem that includes the **symlink_target: string>** do: ```bash sed -i '112252d' filtered_27f05c041a1c401783f90b9415e40e4b.sampled.jsonl ``` 4. Rerun the loading function now is succesful: ```python from datasets import load_dataset ds = load_dataset('json', data_files='/path_to/filtered_27f05c041a1c401783f90b9415e40e4b.sampled.jsonl') ``` ### Expected behavior Have a clean dataset without discrepancies on the jsonl fields or have the load_dataset('json',...) method not error out. ### Environment info - `datasets` version: 2.14.1 - Platform: Linux-4.18.0-425.13.1.el8_7.x86_64-x86_64-with-glibc2.28 - Python version: 3.9.17 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
open
https://github.com/huggingface/datasets/issues/6162
2023-08-18T07:19:39
2023-08-18T17:00:35
null
{ "login": "rbrugaro", "id": 82971690, "type": "User" }
[]
false
[]
1,855,794,354
6,161
Fix protocol prefix for Beam
Fix #6147
closed
https://github.com/huggingface/datasets/pull/6161
2023-08-17T22:40:37
2024-03-18T17:01:21
2024-03-18T17:01:21
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,855,760,543
6,160
Fix Parquet loading with `columns`
Fix #6149
closed
https://github.com/huggingface/datasets/pull/6160
2023-08-17T21:58:24
2023-08-17T22:44:59
2023-08-17T22:36:04
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[]
true
[]
1,855,691,512
6,159
Add `BoundingBox` feature
... to make working with object detection datasets easier. Currently, `Sequence(int_or_float, length=4)` can be used to represent this feature optimally (in the storage backend), so I only see this feature being useful if we make it work with the viewer. Also, bounding boxes usually come in 4 different formats (explained [here](https://albumentations.ai/docs/getting_started/bounding_boxes_augmentation/)), so we need to decide which one to support (or maybe all of them). cc @NielsRogge @severo
open
https://github.com/huggingface/datasets/issues/6159
2023-08-17T20:49:51
2024-11-18T17:58:43
null
{ "login": "mariosasko", "id": 47462742, "type": "User" }
[ { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,855,374,220
6,158
[docs] Complete `to_iterable_dataset`
Finishes the `to_iterable_dataset` documentation by adding it to the relevant sections in the tutorial and guide.
closed
https://github.com/huggingface/datasets/pull/6158
2023-08-17T17:02:11
2023-08-17T19:24:20
2023-08-17T19:13:15
{ "login": "stevhliu", "id": 59462357, "type": "User" }
[]
true
[]
1,855,265,663
6,157
DatasetInfo.__init__() got an unexpected keyword argument '_column_requires_decoding'
### Describe the bug When I was in load_dataset, it said "DatasetInfo.__init__() got an unexpected keyword argument '_column_requires_decoding'". The second time I ran it, there was no error and the dataset object worked ```python --------------------------------------------------------------------------- TypeError Traceback (most recent call last) Cell In[3], line 1 ----> 1 dataset = load_dataset( 2 "/home/aihao/workspace/DeepLearningContent/datasets/manga", 3 data_dir="/home/aihao/workspace/DeepLearningContent/datasets/manga", 4 split="train", 5 ) File [~/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/load.py:2146](https://vscode-remote+ssh-002dremote-002bhome.vscode-resource.vscode-cdn.net/home/aihao/workspace/DeepLearningContent/datasets/~/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/load.py:2146), in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, ignore_verifications, keep_in_memory, save_infos, revision, token, use_auth_token, task, streaming, num_proc, storage_options, **config_kwargs) 2142 # Build dataset for splits 2143 keep_in_memory = ( 2144 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size) 2145 ) -> 2146 ds = builder_instance.as_dataset(split=split, verification_mode=verification_mode, in_memory=keep_in_memory) 2147 # Rename and cast features to match task schema 2148 if task is not None: 2149 # To avoid issuing the same warning twice File [~/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/builder.py:1190](https://vscode-remote+ssh-002dremote-002bhome.vscode-resource.vscode-cdn.net/home/aihao/workspace/DeepLearningContent/datasets/~/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/builder.py:1190), in DatasetBuilder.as_dataset(self, split, run_post_process, verification_mode, ignore_verifications, in_memory) 1187 verification_mode = VerificationMode(verification_mode or VerificationMode.BASIC_CHECKS) 1189 # Create a dataset for each of the given splits -> 1190 datasets = map_nested( 1191 partial( 1192 self._build_single_dataset, ... File [~/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/info.py:379](https://vscode-remote+ssh-002dremote-002bhome.vscode-resource.vscode-cdn.net/home/aihao/workspace/DeepLearningContent/datasets/~/miniconda3/envs/torch/lib/python3.11/site-packages/datasets/info.py:379), in DatasetInfo.copy(self) 378 def copy(self) -> "DatasetInfo": --> 379 return self.__class__(**{k: copy.deepcopy(v) for k, v in self.__dict__.items()}) TypeError: DatasetInfo.__init__() got an unexpected keyword argument '_column_requires_decoding' ``` ### Steps to reproduce the bug /home/aihao/workspace/DeepLearningContent/datasets/images/images.py ```python from logging import config import datasets import os from PIL import Image import csv import json class ImagesConfig(datasets.BuilderConfig): def __init__(self, **kwargs): super(ImagesConfig, self).__init__(**kwargs) class Images(datasets.GeneratorBasedBuilder): def _split_generators(self, dl_manager: datasets.DownloadManager): return [ datasets.SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={"split": datasets.Split.TRAIN}, ) ] BUILDER_CONFIGS = [ ImagesConfig( name="similar_pairs", description="simliar pair dataset,item is a pair of similar images", ), ImagesConfig( name="image_prompt_pairs", description="image prompt pairs", ), ] def _info(self): if self.config.name == "similar_pairs": return datasets.Features( { "image1": datasets.features.Image(), "image2": datasets.features.Image(), "similarity": datasets.Value("float32"), } ) elif self.config.name == "image_prompt_pairs": return datasets.Features( {"image": datasets.features.Image(), "prompt": datasets.Value("string")} ) def _generate_examples(self, split): data_path = os.path.join(self.config.data_dir, "data") if self.config.name == "similar_pairs": prompts = {} with open(os.path.join(data_path ,"prompts.json"), "r") as f: prompts = json.load(f) with open(os.path.join(data_path, "similar_pairs.csv"), "r") as f: reader = csv.reader(f) for row in reader: image1_path, image2_path, similarity = row yield image1_path + ":" + image2_path + ":", { "image1": Image.open(image1_path), "prompt1": prompts[image1_path], "image2": Image.open(image2_path), "prompt2": prompts[image2_path], "similarity": float(similarity), } ``` Code that indicates an error: ```python from datasets import load_dataset import json import csv import ast import torch data_dir = "/home/aihao/workspace/DeepLearningContent/datasets/images" dataset = load_dataset(data_dir, data_dir=data_dir, name="similar_pairs") ``` ### Expected behavior The first execution gives an error, but it works fine ### Environment info - `datasets` version: 2.14.3 - Platform: Linux-6.2.0-26-generic-x86_64-with-glibc2.35 - Python version: 3.11.4 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
closed
https://github.com/huggingface/datasets/issues/6157
2023-08-17T15:48:11
2023-09-27T17:36:14
2023-09-27T17:36:14
{ "login": "aihao2000", "id": 51043929, "type": "User" }
[]
false
[]
1,854,768,618
6,156
Why not use self._epoch as seed to shuffle in distributed training with IterableDataset
### Describe the bug Currently, distributed training with `IterableDataset` needs to pass fixed seed to shuffle to keep each node use the same seed to avoid overlapping. https://github.com/huggingface/datasets/blob/a7f8d9019e7cb104eac4106bdc6ec0292f0dc61a/src/datasets/iterable_dataset.py#L1174-L1177 My question is why not directly use `self._epoch` which is set by `set_epoch` as seed? It's almost the same across nodes. https://github.com/huggingface/datasets/blob/a7f8d9019e7cb104eac4106bdc6ec0292f0dc61a/src/datasets/iterable_dataset.py#L1790-L1801 If not using `self._epoch` as shuffling seed, what does this method do to prepare an epoch seeded generator? https://github.com/huggingface/datasets/blob/a7f8d9019e7cb104eac4106bdc6ec0292f0dc61a/src/datasets/iterable_dataset.py#L1206 ### Steps to reproduce the bug As mentioned above. ### Expected behavior As mentioned above. ### Environment info Not related
closed
https://github.com/huggingface/datasets/issues/6156
2023-08-17T10:58:20
2023-08-17T14:33:15
2023-08-17T14:33:14
{ "login": "npuichigo", "id": 11533479, "type": "User" }
[]
false
[]
1,854,661,682
6,155
Raise FileNotFoundError when passing data_files that don't exist
e.g. when running `load_dataset("parquet", data_files="doesnt_exist.parquet")`
closed
https://github.com/huggingface/datasets/pull/6155
2023-08-17T09:49:48
2023-08-18T13:45:58
2023-08-18T13:35:13
{ "login": "lhoestq", "id": 42851186, "type": "User" }
[]
true
[]
1,854,595,943
6,154
Use yaml instead of get data patterns when possible
This would make the data files resolution faster: no need to list all the data files to infer the dataset builder to use. fix https://github.com/huggingface/datasets/issues/6140
closed
https://github.com/huggingface/datasets/pull/6154
2023-08-17T09:17:05
2023-08-17T20:46:25
2023-08-17T20:37:19
{ "login": "lhoestq", "id": 42851186, "type": "User" }
[]
true
[]
1,852,494,646
6,152
FolderBase Dataset automatically resolves under current directory when data_dir is not specified
### Describe the bug FolderBase Dataset automatically resolves under current directory when data_dir is not specified. For example: ``` load_dataset("audiofolder") ``` takes long time to resolve and collect data_files from current directory. But I think it should reach out to this line for error handling https://github.com/huggingface/datasets/blob/cb8c5de5145c7e7eee65391cb7f4d92f0d565d62/src/datasets/packaged_modules/folder_based_builder/folder_based_builder.py#L58-L59 ### Steps to reproduce the bug ``` load_dataset("audiofolder") ``` ### Expected behavior Error report ### Environment info - `datasets` version: 2.14.4 - Platform: Linux-5.15.0-78-generic-x86_64-with-glibc2.17 - Python version: 3.8.15 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 1.5.3
closed
https://github.com/huggingface/datasets/issues/6152
2023-08-16T04:38:09
2025-06-18T14:18:42
2025-06-18T14:18:42
{ "login": "npuichigo", "id": 11533479, "type": "User" }
[ { "name": "good first issue", "color": "7057ff" } ]
false
[]
1,851,497,818
6,151
Faster sorting for single key items
### Feature request A faster way to sort a dataset which contains a large number of rows. ### Motivation The current sorting implementations took significantly longer than expected when I was running on a dataset trying to sort by timestamps. **Code snippet:** ```python ds = datasets.load_dataset( "json", **{"data_files": {"train": "path-to-jsonlines"}, "split": "train"}, num_proc=os.cpu_count(), keep_in_memory=True) sorted_ds = ds.sort("pubDate", keep_in_memory=True) ``` However, once I switched to a different method which 1. unpacked to a list of tuples 2. sorted tuples by key 3. run `.select` with the sorted list of indices It was significantly faster (orders of magnitude, especially with M's of rows) ### Your contribution I'd be happy to implement a crude single key sorting algorithm so that other users can benefit from this trick. Broadly, this would take a `Dataset` and perform; ```python # ds is a Dataset object # key_name is the sorting key class Dataset: ... def _sort(key_name: str) -> Dataset: index_keys = [(i,x) for i,x in enumerate(self[key_name])] sorted_rows = sorted(row_pubdate, key=lambda x: x[1]) sorted_indicies = [x[0] for x in sorted_rows] return self.select(sorted_indicies) ```
closed
https://github.com/huggingface/datasets/issues/6151
2023-08-15T14:02:31
2023-08-21T14:38:26
2023-08-21T14:38:25
{ "login": "jackapbutler", "id": 47942453, "type": "User" }
[ { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,850,740,456
6,150
Allow dataset implement .take
### Feature request I want to do: ``` dataset.take(512) ``` but it only works with streaming = True ### Motivation uniform interface to data sets. Really surprising the above only works with streaming = True. ### Your contribution Should be trivial to copy paste the IterableDataset .take to use the local path in the data (when streaming = False)
open
https://github.com/huggingface/datasets/issues/6150
2023-08-15T00:17:51
2023-08-17T13:49:37
null
{ "login": "brando90", "id": 1855278, "type": "User" }
[ { "name": "enhancement", "color": "a2eeef" } ]
false
[]
1,850,700,624
6,149
Dataset.from_parquet cannot load subset of columns
### Describe the bug When using `Dataset.from_parquet(path_or_paths, columns=[...])` and a subset of columns, loading fails with a variant of the following ``` ValueError: Couldn't cast a: int64 -- schema metadata -- pandas: '{"index_columns": [], "column_indexes": [], "columns": [{"name":' + 273 to {'a': Value(dtype='int64', id=None), 'b': Value(dtype='int64', id=None)} because column names don't match The above exception was the direct cause of the following exception: ``` Looks to be triggered by https://github.com/huggingface/datasets/blob/c02a44715c036b5261686669727394b1308a3a4b/src/datasets/table.py#L2285-L2286 ### Steps to reproduce the bug ``` import pandas as pd from datasets import Dataset pd.DataFrame([{"a": 1, "b": 2}]).to_parquet("test.pq") Dataset.from_parquet("test.pq", columns=["a"]) ``` ### Expected behavior A subset of columns should be loaded without error ### Environment info - `datasets` version: 2.14.4 - Platform: Linux-5.10.0-23-cloud-amd64-x86_64-with-glibc2.2.5 - Python version: 3.8.16 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3
closed
https://github.com/huggingface/datasets/issues/6149
2023-08-14T23:28:22
2023-08-17T22:36:05
2023-08-17T22:36:05
{ "login": "dwyatte", "id": 2512762, "type": "User" }
[]
false
[]
1,849,524,683
6,148
Ignore parallel warning in map_nested
This warning message was shown every time you pass num_proc to `load_dataset` because of `map_nested` ``` parallel_map is experimental and might be subject to breaking changes in the future ``` This PR removes it for `map_nested`. If someone uses another parallel backend they're already warned when `parallel_backend` is called anyway
closed
https://github.com/huggingface/datasets/pull/6148
2023-08-14T10:43:41
2023-08-17T08:54:06
2023-08-17T08:43:58
{ "login": "lhoestq", "id": 42851186, "type": "User" }
[]
true
[]
1,848,914,830
6,147
ValueError when running BeamBasedBuilder with GCS path in cache_dir
### Describe the bug When running the BeamBasedBuilder with a GCS path specified in the cache_dir, the following ValueError occurs: ``` ValueError: Unable to get filesystem from specified path, please use the correct path or ensure the required dependency is installed, e.g., pip install apache-beam[gcp]. Path specified: gcs://my-bucket/huggingface_datasets/my_beam_dataset/default/0.0.0/my_beam_dataset-train [while running 'train/Save to parquet/Write/WriteImpl/InitializeWrite'] ``` Same error occurs after running `pip install apache-beam[gcp]` as instructed. ### Steps to reproduce the bug Put `my_beam_dataset.py`: ```python import datasets class MyBeamDataset(datasets.BeamBasedBuilder): def _info(self): features = datasets.Features({"value": datasets.Value("int64")}) return datasets.DatasetInfo(features=features) def _split_generators(self, dl_manager, pipeline): return [datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={})] def _build_pcollection(self, pipeline): import apache_beam as beam return pipeline | beam.Create([{"value": i} for i in range(10)]) ``` Run: ```bash datasets-cli run_beam my_beam_dataset.py --cache_dir=gs://my-bucket/huggingface_datasets/ --beam_pipeline_options="runner=DirectRunner" ``` ### Expected behavior Running the BeamBasedBuilder with a GCS cache path without any errors. ### Environment info - `datasets` version: 2.14.4 - Platform: macOS-13.4-arm64-arm-64bit - Python version: 3.9.17 - Huggingface_hub version: 0.16.4 - PyArrow version: 9.0.0 - Pandas version: 2.0.3
closed
https://github.com/huggingface/datasets/issues/6147
2023-08-14T03:11:34
2024-03-18T16:59:15
2024-03-18T16:59:14
{ "login": "ktrk115", "id": 13844767, "type": "User" }
[]
false
[]
1,848,417,366
6,146
DatasetGenerationError when load glue benchmark datasets from `load_dataset`
### Describe the bug Package version: datasets-2.14.4 When I run the codes: ``` from datasets import load_dataset dataset = load_dataset("glue", "ax") ``` I got the following errors: --------------------------------------------------------------------------- SchemaInferenceError Traceback (most recent call last) File ~/anaconda3/envs/python3/lib/python3.10/site-packages/datasets/builder.py:1949, in ArrowBasedBuilder._prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id) 1948 num_shards = shard_id + 1 -> 1949 num_examples, num_bytes = writer.finalize() 1950 writer.close() File ~/anaconda3/envs/python3/lib/python3.10/site-packages/datasets/arrow_writer.py:598, in ArrowWriter.finalize(self, close_stream) 597 self.stream.close() --> 598 raise SchemaInferenceError("Please pass `features` or at least one example when writing data") 599 logger.debug( 600 f"Done writing {self._num_examples} {self.unit} in {self._num_bytes} bytes {self._path if self._path else ''}." 601 ) SchemaInferenceError: Please pass `features` or at least one example when writing data The above exception was the direct cause of the following exception: DatasetGenerationError Traceback (most recent call last) Cell In[5], line 3 1 from datasets import load_dataset ----> 3 dataset = load_dataset("glue", "ax") File ~/anaconda3/envs/python3/lib/python3.10/site-packages/datasets/load.py:2136, in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, ignore_verifications, keep_in_memory, save_infos, revision, token, use_auth_token, task, streaming, num_proc, storage_options, **config_kwargs) 2133 try_from_hf_gcs = path not in _PACKAGED_DATASETS_MODULES 2135 # Download and prepare data -> 2136 builder_instance.download_and_prepare( 2137 download_config=download_config, 2138 download_mode=download_mode, 2139 verification_mode=verification_mode, 2140 try_from_hf_gcs=try_from_hf_gcs, 2141 num_proc=num_proc, 2142 storage_options=storage_options, 2143 ) 2145 # Build dataset for splits 2146 keep_in_memory = ( 2147 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size) 2148 ) File ~/anaconda3/envs/python3/lib/python3.10/site-packages/datasets/builder.py:954, in DatasetBuilder.download_and_prepare(self, output_dir, download_config, download_mode, verification_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, file_format, max_shard_size, num_proc, storage_options, **download_and_prepare_kwargs) 952 if num_proc is not None: 953 prepare_split_kwargs["num_proc"] = num_proc --> 954 self._download_and_prepare( 955 dl_manager=dl_manager, 956 verification_mode=verification_mode, 957 **prepare_split_kwargs, 958 **download_and_prepare_kwargs, 959 ) 960 # Sync info 961 self.info.dataset_size = sum(split.num_bytes for split in self.info.splits.values()) File ~/anaconda3/envs/python3/lib/python3.10/site-packages/datasets/builder.py:1049, in DatasetBuilder._download_and_prepare(self, dl_manager, verification_mode, **prepare_split_kwargs) 1045 split_dict.add(split_generator.split_info) 1047 try: 1048 # Prepare split will record examples associated to the split -> 1049 self._prepare_split(split_generator, **prepare_split_kwargs) 1050 except OSError as e: 1051 raise OSError( 1052 "Cannot find data file. " 1053 + (self.manual_download_instructions or "") 1054 + "\nOriginal error:\n" 1055 + str(e) 1056 ) from None File ~/anaconda3/envs/python3/lib/python3.10/site-packages/datasets/builder.py:1813, in ArrowBasedBuilder._prepare_split(self, split_generator, file_format, num_proc, max_shard_size) 1811 job_id = 0 1812 with pbar: -> 1813 for job_id, done, content in self._prepare_split_single( 1814 gen_kwargs=gen_kwargs, job_id=job_id, **_prepare_split_args 1815 ): 1816 if done: 1817 result = content File ~/anaconda3/envs/python3/lib/python3.10/site-packages/datasets/builder.py:1958, in ArrowBasedBuilder._prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id) 1956 if isinstance(e, SchemaInferenceError) and e.__context__ is not None: 1957 e = e.__context__ -> 1958 raise DatasetGenerationError("An error occurred while generating the dataset") from e 1960 yield job_id, True, (total_num_examples, total_num_bytes, writer._features, num_shards, shard_lengths) DatasetGenerationError: An error occurred while generating the dataset ### Steps to reproduce the bug from datasets import load_dataset dataset = load_dataset("glue", "ax") ### Expected behavior When generating the train split: Generating train split: 0/0 [00:00<?, ? examples/s] It raise the error: DatasetGenerationError: An error occurred while generating the dataset ### Environment info datasets-2.14.4. Python 3.10
closed
https://github.com/huggingface/datasets/issues/6146
2023-08-13T05:17:56
2023-08-26T22:09:09
2023-08-26T22:09:09
{ "login": "yusx-swapp", "id": 78742415, "type": "User" }
[]
false
[]
1,852,630,074
6,153
custom load dataset to hub
### System Info kaggle notebook i transformed dataset: ``` dataset = load_dataset("Dahoas/first-instruct-human-assistant-prompt") ``` to formatted_dataset: ``` Dataset({ features: ['message_tree_id', 'message_tree_text'], num_rows: 33143 }) ``` but would like to know how to upload to hub ### Who can help? @ArthurZucker @younesbelkada ### Information - [ ] The official example scripts - [ ] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...) - [ ] My own task or dataset (give details below) ### Reproduction shared above ### Expected behavior load dataset to hub
closed
https://github.com/huggingface/datasets/issues/6153
2023-08-13T04:42:22
2023-11-21T11:50:28
2023-10-08T17:04:16
{ "login": "andysingal", "id": 20493493, "type": "User" }
[]
false
[]
1,847,811,310
6,145
Export to_iterable_dataset to document
Fix the export of a missing method of `Dataset`
closed
https://github.com/huggingface/datasets/pull/6145
2023-08-12T07:00:14
2023-08-15T17:04:01
2023-08-15T16:55:24
{ "login": "npuichigo", "id": 11533479, "type": "User" }
[]
true
[]
1,847,296,711
6,144
NIH exporter file not found
### Describe the bug can't use or download the nih exporter pile data. ``` 15 experiment_compute_diveristy_coeff_single_dataset_then_combined_datasets_with_domain_weights() 16 File "/lfs/ampere1/0/brando9/beyond-scale-language-data-diversity/src/diversity/div_coeff.py", line 474, in experiment_compute_diveristy_coeff_single_dataset_then_combined_datasets_with_domain_weights 17 column_names = next(iter(dataset)).keys() 18 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 1353, in __iter__ 19 for key, example in ex_iterable: 20 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/datasets/iterable_dataset.py", line 207, in __iter__ 21 yield from self.generate_examples_fn(**self.kwargs) 22 File "/lfs/ampere1/0/brando9/.cache/huggingface/modules/datasets_modules/datasets/EleutherAI--pile/ebea56d358e91cf4d37b0fde361d563bed1472fbd8221a21b38fc8bb4ba554fb/pile.py", line 236, in _generate_examples 23 with zstd.open(open(files[subset], "rb"), "rt", encoding="utf-8") as f: 24 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/datasets/streaming.py", line 74, in wrapper 25 return function(*args, download_config=download_config, **kwargs) 26 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/datasets/download/streaming_download_manager.py", line 496, in xopen 27 file_obj = fsspec.open(file, mode=mode, *args, **kwargs).open() 28 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/fsspec/core.py", line 134, in open 29 return self.__enter__() 30 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/fsspec/core.py", line 102, in __enter__ 31 f = self.fs.open(self.path, mode=mode) 32 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/fsspec/spec.py", line 1241, in open 33 f = self._open( 34 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/fsspec/implementations/http.py", line 356, in _open 35 size = size or self.info(path, **kwargs)["size"] 36 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/fsspec/asyn.py", line 121, in wrapper 37 return sync(self.loop, func, *args, **kwargs) 38 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/fsspec/asyn.py", line 106, in sync 39 raise return_result 40 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/fsspec/asyn.py", line 61, in _runner 41 result[0] = await coro 42 File "/lfs/ampere1/0/brando9/miniconda/envs/beyond_scale/lib/python3.10/site-packages/fsspec/implementations/http.py", line 430, in _info 43 raise FileNotFoundError(url) from exc 44 FileNotFoundError: https://the-eye.eu/public/AI/pile_preliminary_components/NIH_ExPORTER_awarded_grant_text.jsonl.zst ``` ### Steps to reproduce the bug run this: ``` from datasets import load_dataset path, name = 'EleutherAI/pile', 'nih_exporter' # -- Get data set dataset = load_dataset(path, name, streaming=True, split="train").with_format("torch") batch = dataset.take(512) print(f'{batch=}') ``` ### Expected behavior print the batch ### Environment info ``` (beyond_scale) brando9@ampere1:~/beyond-scale-language-data-diversity$ datasets-cli env Copy-and-paste the text below in your GitHub issue. - `datasets` version: 2.14.4 - Platform: Linux-5.4.0-122-generic-x86_64-with-glibc2.31 - Python version: 3.10.11 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 2.0.3 ```
open
https://github.com/huggingface/datasets/issues/6144
2023-08-11T19:05:25
2023-08-14T23:28:38
null
{ "login": "brando90", "id": 1855278, "type": "User" }
[]
false
[]
1,846,205,216
6,142
the-stack-dedup fails to generate
### Describe the bug I'm getting an error generating the-stack-dedup with datasets 2.13.1, and with 2.14.4 nothing happens. ### Steps to reproduce the bug My code: ``` import os import datasets as ds MY_CACHE_DIR = "/home/ubuntu/the-stack-dedup-local" MY_TOKEN="my-token" the_stack_ds = ds.load_dataset("bigcode/the-stack-dedup", split="train", download_mode="reuse_cache_if_exists", cache_dir=MY_CACHE_DIR, use_auth_token=MY_TOKEN, num_proc=64) ``` The exception: ``` Generating train split: 233248251 examples [54:31, 57280.00 examples/s] multiprocess.pool.RemoteTraceback: """ Traceback (most recent call last): File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/build er.py", line 1879, in _prepare_split_single for _, table in generator: File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/packa ged_modules/parquet/parquet.py", line 82, in _generate_tables yield f"{file_idx}_{batch_idx}", self._cast_table(pa_table) File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/packa ged_modules/parquet/parquet.py", line 61, in _cast_table pa_table = table_cast(pa_table, self.info.features.arrow_schema) File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/table .py", line 2324, in table_cast return cast_table_to_schema(table, schema) File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/table .py", line 2282, in cast_table_to_schema raise ValueError(f"Couldn't cast\n{table.schema}\nto\n{features}\nb ecause column names don't match") ValueError: Couldn't cast hexsha: string size: int64 ext: string lang: string max_stars_repo_path: string max_stars_repo_name: string max_stars_repo_head_hexsha: string max_stars_repo_licenses: list<item: string> child 0, item: string max_stars_count: int64 max_stars_repo_stars_event_min_datetime: string max_stars_repo_stars_event_max_datetime: string max_issues_repo_path: string max_issues_repo_name: string max_issues_repo_head_hexsha: string max_issues_repo_licenses: list<item: string> child 0, item: string max_issues_count: int64 max_issues_repo_issues_event_min_datetime: string max_issues_repo_issues_event_max_datetime: string max_forks_repo_path: string max_forks_repo_name: string max_forks_repo_head_hexsha: string max_forks_repo_licenses: list<item: string> child 0, item: string max_forks_count: int64 max_forks_repo_forks_event_min_datetime: string max_forks_repo_forks_event_max_datetime: string content: string avg_line_length: double max_line_length: int64 alphanum_fraction: double __id__: int64 -- schema metadata -- huggingface: '{"info": {"features": {"hexsha": {"dtype": "string", "_type' + 1979 to {'hexsha': Value(dtype='string', id=None), 'size': Value(dtype='int64', id=None), 'ext': Value(dtype='string', id=None), 'lang': Value(dtype='string', id=None), 'max_stars_repo_path': Value(dtype='string', id=None), 'max_stars_repo_name': Value(dtype='string', id=None), 'max_stars_repo_head_hexsha': Value(dtype='string', id=None), 'max_stars_repo_licenses': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'max_stars_count': Value(dtype='int64', id=None), 'max_stars_repo_stars_event_min_datetime': Value(dtype='string', id=None), 'max_stars_repo_stars_event_max_datetime': Value(dtype='string', id=None), 'max_issues_repo_path': Value(dtype='string', id=None), 'max_issues_repo_name': Value(dtype='string', id=None), 'max_issues_repo_head_hexsha': Value(dtype='string', id=None), 'max_issues_repo_licenses': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'max_issues_count': Value(dtype='int64', id=None), 'max_issues_repo_issues_event_min_datetime': Value(dtype='string', id=None), 'max_issues_repo_issues_event_max_datetime': Value(dtype='string', id=None), 'max_forks_repo_path': Value(dtype='string', id=None), 'max_forks_repo_name': Value(dtype='string', id=None), 'max_forks_repo_head_hexsha': Value(dtype='string', id=None), 'max_forks_repo_licenses': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'max_forks_count': Value(dtype='int64', id=None), 'max_forks_repo_forks_event_min_datetime': Value(dtype='string', id=None), 'max_forks_repo_forks_event_max_datetime': Value(dtype='string', id=None), 'content': Value(dtype='string', id=None), 'avg_line_length': Value(dtype='float64', id=None), 'max_line_length': Value(dtype='int64', id=None), 'alphanum_fraction': Value(dtype='float64', id=None)} because column names don't match The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/ubuntu/.local/lib/python3.10/site-packages/multiprocess/p ool.py", line 125, in worker result = (True, func(*args, **kwds)) File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/utils /py_utils.py", line 1328, in _write_generator_to_queue for i, result in enumerate(func(**kwargs)): File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/build er.py", line 1912, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating th e dataset") from e datasets.builder.DatasetGenerationError: An error occurred while genera ting the dataset """ The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/ubuntu/download_the_stack.py", line 7, in <module> the_stack_ds = ds.load_dataset("bigcode/the-stack-dedup", split="tr ain", download_mode="reuse_cache_if_exists", cache_dir=MY_CACHE_DIR, us e_auth_token=MY_TOKEN, num_proc=64) File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/load. py", line 1809, in load_dataset builder_instance.download_and_prepare( File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/build er.py", line 909, in download_and_prepare self._download_and_prepare( File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/build er.py", line 1004, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/build er.py", line 1796, in _prepare_split for job_id, done, content in iflatmap_unordered( File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/utils /py_utils.py", line 1354, in iflatmap_unordered [async_result.get(timeout=0.05) for async_result in async_results] File "/home/ubuntu/.local/lib/python3.10/site-packages/datasets/utils /py_utils.py", line 1354, in <listcomp> [async_result.get(timeout=0.05) for async_result in async_results] File "/home/ubuntu/.local/lib/python3.10/site-packages/multiprocess/p ool.py", line 774, in get raise self._value datasets.builder.DatasetGenerationError: An error occurred while generating the dataset ``` ### Expected behavior The dataset downloads properly. @lhoestq @loub ### Environment info Datasets 2.13.1, large VM with 2TB RAM, Ubuntu 20.04
closed
https://github.com/huggingface/datasets/issues/6142
2023-08-11T05:10:49
2023-08-17T09:26:13
2023-08-17T09:26:13
{ "login": "michaelroyzen", "id": 45830328, "type": "User" }
[]
false
[]
1,846,117,729
6,141
TypeError: ClientSession._request() got an unexpected keyword argument 'https'
### Describe the bug Hello, when I ran the [code snippet](https://huggingface.co/docs/datasets/v2.14.4/en/loading#json) on the document, I encountered the following problem: ``` Python 3.10.9 (main, Mar 1 2023, 18:23:06) [GCC 11.2.0] on linux Type "help", "copyright", "credits" or "license" for more information. >>> from datasets import load_dataset >>> base_url = "https://rajpurkar.github.io/SQuAD-explorer/dataset/" >>> dataset = load_dataset("json", data_files={"train": base_url + "train-v1.1.json", "validation": base_url + "dev-v1.1.json"}, field="data") Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/liushuai/anaconda3/lib/python3.10/site-packages/datasets/load.py", line 2112, in load_dataset builder_instance = load_dataset_builder( File "/home/liushuai/anaconda3/lib/python3.10/site-packages/datasets/load.py", line 1798, in load_dataset_builder dataset_module = dataset_module_factory( File "/home/liushuai/anaconda3/lib/python3.10/site-packages/datasets/load.py", line 1413, in dataset_module_factory ).get_module() File "/home/liushuai/anaconda3/lib/python3.10/site-packages/datasets/load.py", line 949, in get_module data_files = DataFilesDict.from_patterns( File "/home/liushuai/anaconda3/lib/python3.10/site-packages/datasets/data_files.py", line 672, in from_patterns DataFilesList.from_patterns( File "/home/liushuai/anaconda3/lib/python3.10/site-packages/datasets/data_files.py", line 578, in from_patterns resolve_pattern( File "/home/liushuai/anaconda3/lib/python3.10/site-packages/datasets/data_files.py", line 340, in resolve_pattern for filepath, info in fs.glob(pattern, detail=True).items() File "/home/liushuai/anaconda3/lib/python3.10/site-packages/fsspec/asyn.py", line 113, in wrapper return sync(self.loop, func, *args, **kwargs) File "/home/liushuai/anaconda3/lib/python3.10/site-packages/fsspec/asyn.py", line 98, in sync raise return_result File "/home/liushuai/anaconda3/lib/python3.10/site-packages/fsspec/asyn.py", line 53, in _runner result[0] = await coro File "/home/liushuai/anaconda3/lib/python3.10/site-packages/fsspec/implementations/http.py", line 449, in _glob elif await self._exists(path): File "/home/liushuai/anaconda3/lib/python3.10/site-packages/fsspec/implementations/http.py", line 306, in _exists r = await session.get(self.encode_url(path), **kw) File "/home/liushuai/anaconda3/lib/python3.10/site-packages/aiohttp/client.py", line 922, in get self._request(hdrs.METH_GET, url, allow_redirects=allow_redirects, **kwargs) TypeError: ClientSession._request() got an unexpected keyword argument 'https' ``` ### Steps to reproduce the bug ``` from datasets import load_dataset base_url = "https://rajpurkar.github.io/SQuAD-explorer/dataset/" dataset = load_dataset("json", data_files={"train": base_url + "train-v1.1.json", "validation": base_url + "dev-v1.1.json"}, field="data") ``` ### Expected behavior able to load normally ### Environment info - `datasets` version: 2.14.4 - Platform: Linux-5.4.54-2-x86_64-with-glibc2.27 - Python version: 3.10.9 - Huggingface_hub version: 0.16.4 - PyArrow version: 12.0.1 - Pandas version: 1.5.3
closed
https://github.com/huggingface/datasets/issues/6141
2023-08-11T02:40:32
2023-08-30T13:51:33
2023-08-30T13:51:33
{ "login": "q935970314", "id": 35994018, "type": "User" }
[]
false
[]
1,845,384,712
6,140
Misalignment between file format specified in configs metadata YAML and the inferred builder
There is a misalignment between the format of the `data_files` specified in the configs metadata YAML (CSV): ```yaml configs: - config_name: default data_files: - split: train path: data.csv ``` and the inferred builder (JSON). Note there are multiple JSON files in the repo, but they do not appear in the configs metadata YAML. See: https://huggingface.co/datasets/freddyaboulton/chatinterface_with_image_csv/discussions/1 CC: @freddyaboulton @polinaeterna
closed
https://github.com/huggingface/datasets/issues/6140
2023-08-10T15:07:34
2023-08-17T20:37:20
2023-08-17T20:37:20
{ "login": "albertvillanova", "id": 8515462, "type": "User" }
[ { "name": "bug", "color": "d73a4a" } ]
false
[]