id
int64
599M
3.29B
url
stringlengths
58
61
html_url
stringlengths
46
51
number
int64
1
7.72k
title
stringlengths
1
290
state
stringclasses
2 values
comments
int64
0
70
created_at
timestamp[s]date
2020-04-14 10:18:02
2025-08-05 09:28:51
updated_at
timestamp[s]date
2020-04-27 16:04:17
2025-08-05 11:39:56
closed_at
timestamp[s]date
2020-04-14 12:01:40
2025-08-01 05:15:45
user_login
stringlengths
3
26
labels
listlengths
0
4
body
stringlengths
0
228k
is_pull_request
bool
2 classes
1,042,887,291
https://api.github.com/repos/huggingface/datasets/issues/3200
https://github.com/huggingface/datasets/pull/3200
3,200
Catch token invalid error in CI
closed
0
2021-11-02T21:56:26
2021-11-03T09:41:08
2021-11-03T09:41:08
lhoestq
[]
The staging back end sometimes returns invalid token errors when trying to delete a repo. I modified the fixture in the test that uses staging to ignore this error
true
1,042,860,935
https://api.github.com/repos/huggingface/datasets/issues/3199
https://github.com/huggingface/datasets/pull/3199
3,199
Bump huggingface_hub
closed
0
2021-11-02T21:29:10
2021-11-14T01:48:11
2021-11-02T21:41:40
lhoestq
[]
huggingface_hub just released its first minor version, so we need to update the dependency It was supposed to be part of 1.15.0 but I'm adding it for 1.15.1
true
1,042,679,548
https://api.github.com/repos/huggingface/datasets/issues/3198
https://github.com/huggingface/datasets/pull/3198
3,198
Add Multi-Lingual LibriSpeech
closed
0
2021-11-02T18:23:59
2021-11-04T17:09:22
2021-11-04T17:09:22
patrickvonplaten
[]
Add https://www.openslr.org/94/
true
1,042,541,127
https://api.github.com/repos/huggingface/datasets/issues/3197
https://github.com/huggingface/datasets/pull/3197
3,197
Fix optimized encoding for arrays
closed
0
2021-11-02T15:55:53
2021-11-02T19:12:24
2021-11-02T19:12:23
lhoestq
[]
Hi ! #3124 introduced a regression that made the benchmarks CI fail because of a bad array comparison when checking the first encoded element. This PR fixes this by making sure that encoding is applied on all sequence types except lists. cc @eladsegal fyi (no big deal)
true
1,042,223,913
https://api.github.com/repos/huggingface/datasets/issues/3196
https://github.com/huggingface/datasets/pull/3196
3,196
QOL improvements: auto-flatten_indices and desc in map calls
closed
0
2021-11-02T11:28:50
2021-11-02T15:41:09
2021-11-02T15:41:08
mariosasko
[]
This PR: * automatically calls `flatten_indices` where needed: in `unique` and `save_to_disk` to avoid saving the indices file * adds descriptions to the map calls Fix #3040
true
1,042,204,044
https://api.github.com/repos/huggingface/datasets/issues/3195
https://github.com/huggingface/datasets/pull/3195
3,195
More robust `None` handling
closed
5
2021-11-02T11:15:10
2021-12-09T14:27:00
2021-12-09T14:26:58
mariosasko
[]
PyArrow has explicit support for `null` values, so it makes sense to support Nones on our side as well. [Colab Notebook with examples](https://colab.research.google.com/drive/1zcK8BnZYnRe3Ao2271u1T19ag9zLEiy3?usp=sharing) Changes: * allow None for the features types with special encoding (`ClassLabel, TranslationVariableLanguages, Value, _ArrayXD`) * handle None in `class_encode_column` (also there is an option to stringify Nones and treat them as a class) * support None sorting in `sort` (use pandas for that) * handle None in align_labels_with_mapping * support for None in ArrayXD (converts `None` to `np.nan` to align the behavior with PyArrow) * support for None in the Audio/Image feature * allow promotion when concatenating tables (`pa.concat_tables(table_list, promote=True)`) and `null` row/~~column~~ broadcasting similar to pandas Additional notes: * use `null` instead of `none` for function arguments for consistency with existing `disable_nullable` * fixes a bug with the `update_metadata_with_features` call in `Dataset.rename_columns` * had to update some tests, let me know if that's ok TODO: - [x] check how the Audio features behaves with Nones - [x] Better None handling in `concatenate_datasets`/`add_item` - [x] Fix formatting with Nones - [x] Add Colab with examples - [x] Tests TODOs for subsequent PRs: - Mention None handling in the docs - Add `drop_null`/`fill_null` to `Dataset`/`DatasetDict` Fix #3181 #3253
true
1,041,999,535
https://api.github.com/repos/huggingface/datasets/issues/3194
https://github.com/huggingface/datasets/pull/3194
3,194
Update link to Datasets Tagging app in Spaces
closed
0
2021-11-02T08:13:50
2021-11-08T10:36:23
2021-11-08T10:36:22
albertvillanova
[]
Fix #3193.
true
1,041,971,117
https://api.github.com/repos/huggingface/datasets/issues/3193
https://github.com/huggingface/datasets/issues/3193
3,193
Update link to datasets-tagging app
closed
0
2021-11-02T07:39:59
2021-11-08T10:36:22
2021-11-08T10:36:22
albertvillanova
[]
Once datasets-tagging has been transferred to Spaces: - huggingface/datasets-tagging#22 We should update the link in Datasets.
false
1,041,308,086
https://api.github.com/repos/huggingface/datasets/issues/3192
https://github.com/huggingface/datasets/issues/3192
3,192
Multiprocessing filter/map (tests) not working on Windows
open
0
2021-11-01T15:36:08
2021-11-01T15:57:03
null
BramVanroy
[ "bug" ]
While running the tests, I found that the multiprocessing examples fail on Windows, or rather they do not complete: they cause a deadlock. I haven't dug deep into it, but they do not seem to work as-is. I currently have no time to tests this in detail but at least the tests seem not to run correctly (deadlocking). ## Steps to reproduce the bug ```shell pytest tests/test_arrow_dataset.py -k "test_filter_multiprocessing" pytest tests/test_arrow_dataset.py -k "test_map_multiprocessing" ``` ## Expected results The functionality to work on all platforms. ## Actual results Deadlock. ## Environment info - `datasets` version: 1.14.1.dev0 - Platform: Windows-10-10.0.19041-SP0 - Python version: 3.9.2, also tested with 3.7.9 - PyArrow version: 4.0.1
false
1,041,225,111
https://api.github.com/repos/huggingface/datasets/issues/3191
https://github.com/huggingface/datasets/issues/3191
3,191
Dataset viewer issue for '*compguesswhat*'
closed
4
2021-11-01T14:16:49
2022-09-12T08:02:29
2022-09-12T08:02:29
benotti
[ "streaming" ]
## Dataset viewer issue for '*compguesswhat*' **Link:** https://huggingface.co/datasets/compguesswhat File not found Am I the one who added this dataset ? No
false
1,041,153,631
https://api.github.com/repos/huggingface/datasets/issues/3190
https://github.com/huggingface/datasets/issues/3190
3,190
combination of shuffle and filter results in a bug
closed
3
2021-11-01T13:07:29
2021-11-02T10:50:49
2021-11-02T10:50:49
rabeehk
[ "bug" ]
## Describe the bug Hi, I would like to shuffle a dataset, then filter it based on each existing label. however, the combination of `filter`, `shuffle` seems to results in a bug. In the minimal example below, as you see in the filtered results, the filtered labels are not unique, meaning filter has not worked. Any suggestions as a temporary fix is appreciated @lhoestq. Thanks. Best regards Rabeeh ## Steps to reproduce the bug ```python import numpy as np import datasets datasets = datasets.load_dataset('super_glue', 'rte', script_version="master") shuffled_data = datasets["train"].shuffle(seed=42) for label in range(2): print("label ", label) data = shuffled_data.filter(lambda example: int(example['label']) == label) print("length ", len(data), np.unique(data['label'])) ``` ## Expected results Filtering per label, should only return the data with that specific label. ## Actual results As you can see, filtered data per label, has still two labels of [0, 1] ``` label 0 length 1249 [0 1] label 1 length 1241 [0 1] ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.1 - Platform: linux - Python version: 3.7.11 - PyArrow version: 5.0.0
false
1,041,044,986
https://api.github.com/repos/huggingface/datasets/issues/3189
https://github.com/huggingface/datasets/issues/3189
3,189
conll2003 incorrect label explanation
closed
1
2021-11-01T11:03:30
2021-11-09T10:40:58
2021-11-09T10:40:58
BramVanroy
[ "bug" ]
In the [conll2003](https://huggingface.co/datasets/conll2003#data-fields) README, the labels are described as follows > - `id`: a `string` feature. > - `tokens`: a `list` of `string` features. > - `pos_tags`: a `list` of classification labels, with possible values including `"` (0), `''` (1), `#` (2), `$` (3), `(` (4). > - `chunk_tags`: a `list` of classification labels, with possible values including `O` (0), `B-ADJP` (1), `I-ADJP` (2), `B-ADVP` (3), `I-ADVP` (4). > - `ner_tags`: a `list` of classification labels, with possible values including `O` (0), `B-PER` (1), `I-PER` (2), `B-ORG` (3), `I-ORG` (4) `B-LOC` (5), `I-LOC` (6) `B-MISC` (7), `I-MISC` (8). First of all, it would be great if we can get a list of ALL possible pos_tags. Second, the chunk tags labels cannot be correct. The description says the values go from 0 to 4 whereas the data shows values from at least 11 to 21 and 0. EDIT: not really a bug, sorry for mistagging.
false
1,040,980,712
https://api.github.com/repos/huggingface/datasets/issues/3188
https://github.com/huggingface/datasets/issues/3188
3,188
conll2002 issues
closed
4
2021-11-01T09:49:24
2021-11-15T13:50:59
2021-11-12T17:18:11
BramVanroy
[ "dataset-viewer" ]
**Link:** https://huggingface.co/datasets/conll2002 The dataset viewer throws a server error when trying to preview the dataset. ``` Message: Extraction protocol 'train' for file at 'https://raw.githubusercontent.com/teropa/nlp/master/resources/corpora/conll2002/esp.train' is not implemented yet ``` In addition, the "point of contact" has encoding issues and does not work when clicked. Am I the one who added this dataset ? No, @lhoestq did
false
1,040,412,869
https://api.github.com/repos/huggingface/datasets/issues/3187
https://github.com/huggingface/datasets/pull/3187
3,187
Add ChrF(++) (as implemented in sacrebleu)
closed
0
2021-10-31T08:53:58
2021-11-02T14:50:50
2021-11-02T14:31:26
BramVanroy
[]
Similar to my [PR for TER](https://github.com/huggingface/datasets/pull/3153), it feels only right to also include ChrF and friends. These are present in Sacrebleu and are therefore very similar to implement as TER and sacrebleu. I tested the implementation with sacrebleu's tests to verify. You can try this below for yourself ```python import datasets EPSILON = 1e-4 chrf = datasets.load_metric(r"path\to\datasets\metrics\chrf") test_cases = [ (["abcdefg"], ["hijklmnop"], 0.0), (["a"], ["b"], 0.0), ([""], ["b"], 0.0), ([""], ["ref"], 0.0), ([""], ["reference"], 0.0), (["aa"], ["ab"], 8.3333), (["a", "b"], ["a", "c"], 8.3333), (["a"], ["a"], 16.6667), (["a b c"], ["a b c"], 50.0), (["a b c"], ["abc"], 50.0), ([" risk assessment must be made of those who are qualified and expertise in the sector - these are the scientists ."], ["risk assessment has to be undertaken by those who are qualified and expert in that area - that is the scientists ."], 63.361730), ([" Die Beziehung zwischen Obama und Netanjahu ist nicht gerade freundlich. "], ["Das Verhältnis zwischen Obama und Netanyahu ist nicht gerade freundschaftlich."], 64.1302698), (["Niemand hat die Absicht, eine Mauer zu errichten"], ["Niemand hat die Absicht, eine Mauer zu errichten"], 100.0), ] for hyp, ref, score in test_cases: # Note the reference transformation which is different from scarebleu's input format results = chrf.compute(predictions=hyp, references=[[r] for r in ref], char_order=6, word_order=0, beta=3, eps_smoothing=True) if abs(score - results["score"]) > EPSILON: print(f"expected {score}, got {results['score']} for {hyp} - {ref}") test_cases_effective_order = [ (["a"], ["a"], 100.0), ([""], ["reference"], 0.0), (["a b c"], ["a b c"], 100.0), (["a b c"], ["abc"], 100.0), ([""], ["c"], 0.0), (["a", "b"], ["a", "c"], 50.0), (["aa"], ["ab"], 25.0), ] for hyp, ref, score in test_cases_effective_order: # Note the reference transformation which is different from scarebleu's input format results = chrf.compute(predictions=hyp, references=[[r] for r in ref], char_order=6, word_order=0, beta=3, eps_smoothing=False) if abs(score - results["score"]) > EPSILON: print(f"expected {score}, got {results['score']} for {hyp} - {ref}") test_cases_keep_whitespace = [ ( ["Die Beziehung zwischen Obama und Netanjahu ist nicht gerade freundlich."], ["Das Verhältnis zwischen Obama und Netanyahu ist nicht gerade freundschaftlich."], 67.3481606, ), ( ["risk assessment must be made of those who are qualified and expertise in the sector - these are the scientists ."], ["risk assessment has to be undertaken by those who are qualified and expert in that area - that is the scientists ."], 65.2414427, ), ] for hyp, ref, score in test_cases_keep_whitespace: # Note the reference transformation which is different from scarebleu's input format results = chrf.compute(predictions=hyp, references=[[r] for r in ref], char_order=6, word_order=0, beta=3, whitespace=True) if abs(score - results["score"]) > EPSILON: print(f"expected {score}, got {results['score']} for {hyp} - {ref}") predictions = ["The relationship between Obama and Netanyahu is not exactly friendly."] references = [["The ties between Obama and Netanyahu are not particularly friendly."]] print(chrf.compute(predictions=predictions, references=references)) ```
true
1,040,369,397
https://api.github.com/repos/huggingface/datasets/issues/3186
https://github.com/huggingface/datasets/issues/3186
3,186
Dataset viewer for nli_tr
closed
6
2021-10-31T03:56:33
2022-09-12T09:15:34
2022-09-12T08:43:09
e-budur
[ "streaming" ]
## Dataset viewer issue for '*nli_tr*' **Link:** https://huggingface.co/datasets/nli_tr Hello, Thank you for the new dataset preview feature that will help the users to view the datasets online. We just noticed that the dataset viewer widget in the `nli_tr` dataset shows the error below. The error must be due to a temporary problem that may have blocked access to the dataset through the dataset viewer. But the dataset is currently accessible through the link in the error message. May we kindly ask if it would be possible to rerun the job so that it can access the dataset for the dataset viewer function? Thank you. Emrah ------------------------------------------ Server Error Status code: 404 Exception: FileNotFoundError Message: [Errno 2] No such file or directory: 'zip://snli_tr_1.0_train.jsonl::https://tabilab.cmpe.boun.edu.tr/datasets/nli_datasets/snli_tr_1.0.zip ------------------------------------------ Am I the one who added this dataset ? Yes
false
1,040,291,961
https://api.github.com/repos/huggingface/datasets/issues/3185
https://github.com/huggingface/datasets/issues/3185
3,185
7z dataset preview not implemented?
closed
2
2021-10-30T20:18:27
2022-04-12T11:48:16
2022-04-12T11:48:07
Kirili4ik
[ "dataset-viewer" ]
## Dataset viewer issue for dataset 'samsum' **Link:** https://huggingface.co/datasets/samsum Server Error Status code: 400 Exception: NotImplementedError Message: Extraction protocol '7z' for file at 'https://arxiv.org/src/1911.12237v2/anc/corpus.7z' is not implemented yet
false
1,040,114,102
https://api.github.com/repos/huggingface/datasets/issues/3184
https://github.com/huggingface/datasets/pull/3184
3,184
RONEC v2
closed
2
2021-10-30T10:50:03
2021-11-02T16:02:23
2021-11-02T16:02:22
dumitrescustefan
[]
Hi, as we've recently finished with the new RONEC (Romanian Named Entity Corpus), we'd like to update the dataset here as well. It's actually essential as links to V1 are no longer valid. In reality we'd like to replace completely v1, as v2 is a full re-annotation of v1 with additional data (up to 2x size vs v1). I've run the make style and all the dummy and real data test, and they passed. I hope it's okay to merge the new RONEC v2 in the datasets. Thanks!
true
1,039,761,120
https://api.github.com/repos/huggingface/datasets/issues/3183
https://github.com/huggingface/datasets/pull/3183
3,183
Add missing docstring to DownloadConfig
closed
0
2021-10-29T16:56:35
2021-11-02T10:25:38
2021-11-02T10:25:37
mariosasko
[]
Document the `use_etag` and `num_proc` attributes in `DownloadConig`.
true
1,039,739,606
https://api.github.com/repos/huggingface/datasets/issues/3182
https://github.com/huggingface/datasets/pull/3182
3,182
Don't memoize strings when hashing since two identical strings may have different python ids
closed
1
2021-10-29T16:26:17
2021-11-02T09:35:38
2021-11-02T09:35:37
lhoestq
[]
When hashing an object that has several times the same string, the hashing could return a different hash if the identical strings share the same python `id()` or not. Here is an example code that shows how the issue can affect the caching: ```python import json import pyarrow as pa from datasets.features import Features from datasets.fingerprint import Hasher schema = pa.schema([pa.field("some_string", pa.string()), pa.field("another_string", pa.string())]) features_from_schema = Features.from_arrow_schema(schema) Hasher.hash(features_from_schema) # dffa9dca9a73fd8c features_dict = json.loads('{"some_string": {"dtype": "string", "id": null, "_type": "Value"}, "another_string": {"dtype": "string", "id": null, "_type": "Value"}}') features_from_json = Features.from_dict(features_dict) Hasher.hash(features_from_json) # 3812e76b15e6420e features_from_schema == features_from_json # True ``` This is because in `features_dict`, some strings like "dtype" are repeated but don't share the same id, contrary to the ones in `features_from_schema`. I fixed that by disabling memoization for strings. This could be optimized in the future by implementing a smarter memoization with a special handling for strings.
true
1,039,682,097
https://api.github.com/repos/huggingface/datasets/issues/3181
https://github.com/huggingface/datasets/issues/3181
3,181
`None` converted to `"None"` when loading a dataset
closed
9
2021-10-29T15:23:53
2021-12-11T01:16:40
2021-12-09T14:26:57
eladsegal
[ "bug" ]
## Describe the bug When loading a dataset `None` values of the type `NoneType` are converted to `'None'` of the type `str`. ## Steps to reproduce the bug ```python from datasets import load_dataset qasper = load_dataset("qasper", split="train", download_mode="reuse_cache_if_exists") print(qasper[60]["full_text"]["section_name"]) ``` When installing version 1.1.40, the output is `[None, 'Introduction', 'Benchmark Datasets', ...]` When installing from the master branch, the output is `['None', 'Introduction', 'Benchmark Datasets', ...]` Notice how the first element was changed from `NoneType` to `str`. ## Expected results `None` should stay as is. ## Actual results `None` is converted to a string. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: master - Platform: Linux-4.4.0-19041-Microsoft-x86_64-with-glibc2.17 - Python version: 3.8.10 - PyArrow version: 4.0.1
false
1,039,641,316
https://api.github.com/repos/huggingface/datasets/issues/3180
https://github.com/huggingface/datasets/pull/3180
3,180
fix label mapping
closed
3
2021-10-29T14:42:24
2021-11-02T13:41:07
2021-11-02T10:37:12
VictorSanh
[]
Fixing label mapping for hlgd. 0 correponds to same event and 1 corresponds to different event <img width="642" alt="Capture d’écran 2021-10-29 à 10 39 58 AM" src="https://user-images.githubusercontent.com/16107619/139454810-1f225e3d-ad48-44a8-b8b1-9205c9533839.png"> <img width="638" alt="Capture d’écran 2021-10-29 à 10 40 09 AM" src="https://user-images.githubusercontent.com/16107619/139454813-93066a3c-7d33-4f56-b133-2f1a7661e438.png"> nt
true
1,039,571,928
https://api.github.com/repos/huggingface/datasets/issues/3179
https://github.com/huggingface/datasets/issues/3179
3,179
Cannot load dataset when the config name is "special"
closed
1
2021-10-29T13:30:47
2021-10-29T13:35:21
2021-10-29T13:35:21
severo
[ "bug", "dataset-viewer" ]
## Describe the bug After https://github.com/huggingface/datasets/pull/3159, we can get the config name of "Check/region_1", which is "Check___region_1". But now we cannot load the dataset (not sure it's related to the above PR though). It's the case for all the similar datasets, listed in https://github.com/huggingface/datasets-preview-backend/issues/78 ## Steps to reproduce the bug ```python >>> from datasets import get_dataset_config_names >>> get_dataset_config_names("Check/region_1") ['Check___region_1'] >>> load_dataset("Check/region_1") Using custom data configuration Check___region_1-d2b3bc48f11c9be2 Downloading and preparing dataset json/Check___region_1 to /home/slesage/.cache/huggingface/datasets/json/Check___region_1-d2b3bc48f11c9be2/0.0.0/c2d554c3377ea79c7664b93dc65d0803b45e3279000f993c7bfd18937fd7f426... 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 4443.12it/s] 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 1277.19it/s] Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.9/site-packages/datasets/load.py", line 1632, in load_dataset builder_instance.download_and_prepare( File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.9/site-packages/datasets/builder.py", line 607, in download_and_prepare self._download_and_prepare( File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.9/site-packages/datasets/builder.py", line 697, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1159, in _prepare_split writer.write_table(table) File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 442, in write_table pa_table = pa.Table.from_arrays([pa_table[name] for name in self._schema.names], schema=self._schema) File "/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 442, in <listcomp> pa_table = pa.Table.from_arrays([pa_table[name] for name in self._schema.names], schema=self._schema) File "pyarrow/table.pxi", line 1249, in pyarrow.lib.Table.__getitem__ File "pyarrow/table.pxi", line 1825, in pyarrow.lib.Table.column File "pyarrow/table.pxi", line 1800, in pyarrow.lib.Table._ensure_integer_index KeyError: 'Field "builder_name" does not exist in table schema' ``` Loading in streaming mode also returns something strange: ```python >>> list(load_dataset("Check/region_1", streaming=True, split="train")) Using custom data configuration Check___region_1-d2b3bc48f11c9be2 [{'builder_name': None, 'citation': '', 'config_name': None, 'dataset_size': None, 'description': '', 'download_checksums': None, 'download_size': None, 'features': {'speech': {'feature': {'dtype': 'float64', 'id': None, '_type': 'Value'}, 'length': -1, 'id': None, '_type': 'Sequence'}, 'sampling_rate': {'dtype': 'int64', 'id': None, '_type': 'Value'}, 'label': {'dtype': 'string', 'id': None, '_type': 'Value'}}, 'homepage': '', 'license': '', 'post_processed': None, 'post_processing_size': None, 'size_in_bytes': None, 'splits': None, 'supervised_keys': None, 'task_templates': None, 'version': None}, {'_data_files': [{'filename': 'dataset.arrow'}], '_fingerprint': 'f1702bb5533c549c', '_format_columns': ['speech', 'sampling_rate', 'label'], '_format_kwargs': {}, '_format_type': None, '_indexes': {}, '_indices_data_files': None, '_output_all_columns': False, '_split': None}] ``` ## Expected results The dataset should be loaded ## Actual results An error occurs ## Environment info - `datasets` version: 1.14.1.dev0 - Platform: Linux-5.11.0-1020-aws-x86_64-with-glibc2.31 - Python version: 3.9.6 - PyArrow version: 4.0.1
false
1,039,539,076
https://api.github.com/repos/huggingface/datasets/issues/3178
https://github.com/huggingface/datasets/issues/3178
3,178
"Property couldn't be hashed properly" even though fully picklable
closed
26
2021-10-29T12:56:09
2024-08-19T13:03:49
2022-11-02T17:18:43
BramVanroy
[ "bug" ]
## Describe the bug I am trying to tokenize a dataset with spaCy. I found that no matter what I do, the spaCy language object (`nlp`) prevents `datasets` from pickling correctly - or so the warning says - even though manually pickling is no issue. It should not be an issue either, since spaCy objects are picklable. ## Steps to reproduce the bug Here is a [colab](https://colab.research.google.com/drive/1gt75LCBIzsmBMvvipEOvWulvyZseBiA7?usp=sharing) but for some reason I cannot reproduce it there. That may have to do with logging/tqdm on Colab, or with running things in notebooks. I tried below code on Windows and Ubuntu as a Python script and getting the same issue (warning below). ```python import pickle from datasets import load_dataset import spacy class Processor: def __init__(self): self.nlp = spacy.load("en_core_web_sm", disable=["tagger", "parser", "ner", "lemmatizer"]) @staticmethod def collate(batch): return [d["en"] for d in batch] def parse(self, batch): batch = batch["translation"] return {"translation_tok": [{"en_tok": " ".join([t.text for t in doc])} for doc in self.nlp.pipe(self.collate(batch))]} def process(self): ds = load_dataset("wmt16", "de-en", split="train[:10%]") ds = ds.map(self.parse, batched=True, num_proc=6) if __name__ == '__main__': pr = Processor() # succeeds with open("temp.pkl", "wb") as f: pickle.dump(pr, f) print("Successfully pickled!") pr.process() ``` --- Here is a small change that includes `Hasher.hash` that shows that the hasher cannot seem to successfully pickle parts form the NLP object. ```python from datasets.fingerprint import Hasher import pickle from datasets import load_dataset import spacy class Processor: def __init__(self): self.nlp = spacy.load("en_core_web_sm", disable=["tagger", "parser", "ner", "lemmatizer"]) @staticmethod def collate(batch): return [d["en"] for d in batch] def parse(self, batch): batch = batch["translation"] return {"translation_tok": [{"en_tok": " ".join([t.text for t in doc])} for doc in self.nlp.pipe(self.collate(batch))]} def process(self): ds = load_dataset("wmt16", "de-en", split="train[:10]") return ds.map(self.parse, batched=True) if __name__ == '__main__': pr = Processor() # succeeds with open("temp.pkl", "wb") as f: pickle.dump(pr, f) print("Successfully pickled class instance!") # succeeds with open("temp.pkl", "wb") as f: pickle.dump(pr.nlp, f) print("Successfully pickled nlp!") # fails print(Hasher.hash(pr.nlp)) pr.process() ``` ## Expected results This to be picklable, working (fingerprinted), and no warning. ## Actual results In the first snippet, I get this warning Parameter 'function'=<function Processor.parse at 0x7f44982247a0> of the transform datasets.arrow_dataset.Dataset._map_single couldn't be hashed properly, a random hash was used instead. Make sure your transforms and parameters are serializable with pickle or dill for the dataset fingerprinting and caching to work. If you reuse this transform, the caching mechanism will consider it to be different from the previous calls and recompute everything. This warning is only showed once. Subsequent hashing failures won't be showed. In the second, I get this traceback which directs to the `Hasher.hash` line. ``` Traceback (most recent call last): File " \Python\Python36\lib\pickle.py", line 918, in save_global obj2, parent = _getattribute(module, name) File " \Python\Python36\lib\pickle.py", line 266, in _getattribute .format(name, obj)) AttributeError: Can't get local attribute 'add_codes.<locals>.ErrorsWithCodes' on <function add_codes at 0x00000296FF606EA0> During handling of the above exception, another exception occurred: Traceback (most recent call last): File " scratch_4.py", line 40, in <module> print(Hasher.hash(pr.nlp)) File " \lib\site-packages\datasets\fingerprint.py", line 191, in hash return cls.hash_default(value) File " \lib\site-packages\datasets\fingerprint.py", line 184, in hash_default return cls.hash_bytes(dumps(value)) File " \lib\site-packages\datasets\utils\py_utils.py", line 345, in dumps dump(obj, file) File " \lib\site-packages\datasets\utils\py_utils.py", line 320, in dump Pickler(file, recurse=True).dump(obj) File " \lib\site-packages\dill\_dill.py", line 498, in dump StockPickler.dump(self, obj) File " \Python\Python36\lib\pickle.py", line 409, in dump self.save(obj) File " \Python\Python36\lib\pickle.py", line 521, in save self.save_reduce(obj=obj, *rv) File " \Python\Python36\lib\pickle.py", line 634, in save_reduce save(state) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \lib\site-packages\dill\_dill.py", line 990, in save_module_dict StockPickler.save_dict(pickler, obj) File " \Python\Python36\lib\pickle.py", line 821, in save_dict self._batch_setitems(obj.items()) File " \Python\Python36\lib\pickle.py", line 847, in _batch_setitems save(v) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \Python\Python36\lib\pickle.py", line 781, in save_list self._batch_appends(obj) File " \Python\Python36\lib\pickle.py", line 805, in _batch_appends save(x) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \Python\Python36\lib\pickle.py", line 736, in save_tuple save(element) File " \Python\Python36\lib\pickle.py", line 521, in save self.save_reduce(obj=obj, *rv) File " \Python\Python36\lib\pickle.py", line 634, in save_reduce save(state) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \Python\Python36\lib\pickle.py", line 736, in save_tuple save(element) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \lib\site-packages\dill\_dill.py", line 990, in save_module_dict StockPickler.save_dict(pickler, obj) File " \Python\Python36\lib\pickle.py", line 821, in save_dict self._batch_setitems(obj.items()) File " \Python\Python36\lib\pickle.py", line 847, in _batch_setitems save(v) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \lib\site-packages\dill\_dill.py", line 1176, in save_instancemethod0 pickler.save_reduce(MethodType, (obj.__func__, obj.__self__), obj=obj) File " \Python\Python36\lib\pickle.py", line 610, in save_reduce save(args) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \Python\Python36\lib\pickle.py", line 736, in save_tuple save(element) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \lib\site-packages\datasets\utils\py_utils.py", line 523, in save_function obj=obj, File " \Python\Python36\lib\pickle.py", line 610, in save_reduce save(args) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \Python\Python36\lib\pickle.py", line 751, in save_tuple save(element) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \lib\site-packages\dill\_dill.py", line 990, in save_module_dict StockPickler.save_dict(pickler, obj) File " \Python\Python36\lib\pickle.py", line 821, in save_dict self._batch_setitems(obj.items()) File " \Python\Python36\lib\pickle.py", line 847, in _batch_setitems save(v) File " \Python\Python36\lib\pickle.py", line 521, in save self.save_reduce(obj=obj, *rv) File " \Python\Python36\lib\pickle.py", line 605, in save_reduce save(cls) File " \Python\Python36\lib\pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File " \lib\site-packages\dill\_dill.py", line 1439, in save_type StockPickler.save_global(pickler, obj, name=name) File " \Python\Python36\lib\pickle.py", line 922, in save_global (obj, module_name, name)) _pickle.PicklingError: Can't pickle <class 'spacy.errors.add_codes.<locals>.ErrorsWithCodes'>: it's not found as spacy.errors.add_codes.<locals>.ErrorsWithCodes ``` ## Environment info Tried on both Linux and Windows - `datasets` version: 1.14.0 - Platform: Windows-10-10.0.19041-SP0 + Python 3.7.9; Linux-5.11.0-38-generic-x86_64-with-Ubuntu-20.04-focal + Python 3.7.12 - PyArrow version: 6.0.0
false
1,039,487,780
https://api.github.com/repos/huggingface/datasets/issues/3177
https://github.com/huggingface/datasets/issues/3177
3,177
More control over TQDM when using map/filter with multiple processes
closed
2
2021-10-29T11:56:16
2023-02-13T20:16:40
2023-02-13T20:16:40
BramVanroy
[ "enhancement" ]
It would help with the clutter in my terminal if tqdm is only shown for rank 0 when using `num_proces>0` in the map and filter methods of datasets. ```python dataset.map(lambda examples: tokenize(examples["text"]), batched=True, num_proc=6) ``` The above snippet leads to a lot of TQDM bars and depending on your terminal, these will not overwrite but keep pushing each other down. ``` #0: 0%| | 0/13 [00:00<?, ?ba/s] #1: 0%| | 0/13 [00:00<?, ?ba/s] #2: 0%| | 0/13 [00:00<?, ?ba/s] #3: 0%| | 0/13 [00:00<?, ?ba/s] #4: 0%| | 0/13 [00:00<?, ?ba/s] #5: 0%| | 0/13 [00:00<?, ?ba/s] #0: 8%| | 1/13 [00:00<?, ?ba/s] #1: 8%| | 1/13 [00:00<?, ?ba/s] ... ``` Instead, it would be welcome if we had the option to only show the progress of rank 0.
false
1,039,068,312
https://api.github.com/repos/huggingface/datasets/issues/3176
https://github.com/huggingface/datasets/pull/3176
3,176
OpenSLR dataset: update generate_examples to properly extract data for SLR83
closed
1
2021-10-29T00:59:27
2021-11-04T16:20:45
2021-10-29T10:04:09
tyrius02
[]
Fixed #3168. The SLR38 indices are CSV files and there wasn't any code in openslr.py to process these files properly. The end result was an empty table. I've added code to properly process these CSV files.
true
1,038,945,271
https://api.github.com/repos/huggingface/datasets/issues/3175
https://github.com/huggingface/datasets/pull/3175
3,175
Add docs for `to_tf_dataset`
closed
2
2021-10-28T20:55:22
2021-11-03T15:39:36
2021-11-03T10:07:23
stevhliu
[ "documentation" ]
This PR adds some documentation for new features released in v1.13.0, with the main addition being `to_tf_dataset`: - Show how to use `to_tf_dataset` in the tutorial, and move `set_format(type='tensorflow'...)` to the Process section (let me know if I'm missing anything @Rocketknight1 😅). - Add an example for loading dataset from multiple zipped CSV files to the Load section. - Add an example for removing columns for an `IterableDataset`. - Add graphic for visualizing streaming.
true
1,038,427,245
https://api.github.com/repos/huggingface/datasets/issues/3174
https://github.com/huggingface/datasets/pull/3174
3,174
Asserts replaced by exceptions (huggingface#3171)
closed
1
2021-10-28T11:55:45
2021-11-06T06:35:32
2021-10-29T13:08:43
joseporiolayats
[]
I've replaced two asserts with their proper exceptions following the guidelines described in issue #3171 by following the contributing guidelines. PS: This is one of my first PRs, hoping I don't break anything!
true
1,038,404,300
https://api.github.com/repos/huggingface/datasets/issues/3173
https://github.com/huggingface/datasets/pull/3173
3,173
Fix issue with filelock filename being too long on encrypted filesystems
closed
0
2021-10-28T11:28:57
2021-10-29T09:42:24
2021-10-29T09:42:24
mariosasko
[]
Infer max filename length in filelock on Unix-like systems. Should fix problems on encrypted filesystems such as eCryptfs. Fix #2924 cc: @lmmx
true
1,038,351,587
https://api.github.com/repos/huggingface/datasets/issues/3172
https://github.com/huggingface/datasets/issues/3172
3,172
`SystemError 15` thrown in `Dataset.__del__` when using `Dataset.map()` with `num_proc>1`
closed
12
2021-10-28T10:29:00
2024-04-02T18:13:21
2021-11-03T11:26:10
vlievin
[ "bug" ]
## Describe the bug I use `datasets.map` to preprocess some data in my application. The error `SystemError 15` is thrown at the end of the execution of `Dataset.map()` (only with `num_proc>1`. Traceback included bellow. The exception is raised only when the code runs within a specific context. Despite ~10h spent investigating this issue, I have failed to isolate the bug, so let me describe my setup. In my project, `Dataset` is wrapped into a `LightningDataModule` and the data is preprocessed when calling `LightningDataModule.setup()`. Calling `.setup()` in an isolated script works fine (even when wrapped with `hydra.main()`). However, when calling `.setup()` within the experiment script (depends on `pytorch_lightning`), the script crashes and `SystemError 15`. I could avoid throwing this error by modifying ` Dataset.__del__()` (see bellow), but I believe this only moves the problem somewhere else. I am completely stuck with this issue, any hint would be welcome. ```python class Dataset() ... def __del__(self): if hasattr(self, "_data"): _ = self._data # <- ugly trick that allows avoiding the issue. del self._data if hasattr(self, "_indices"): del self._indices ``` ## Steps to reproduce the bug ```python # Unfortunately I couldn't isolate the bug. ``` ## Expected results Calling `Dataset.map()` without throwing an exception. Or at least raising a more detailed exception/traceback. ## Actual results ``` Exception ignored in: <function Dataset.__del__ at 0x7f7cec179160>███████████████████████████████████████████████████| 5/5 [00:05<00:00, 1.17ba/s] Traceback (most recent call last): File ".../python3.8/site-packages/datasets/arrow_dataset.py", line 906, in __del__ del self._data File ".../python3.8/site-packages/ray/worker.py", line 1033, in sigterm_handler sys.exit(signum) SystemExit: 15 ``` ## Environment info Tested on 2 environments: **Environment 1.** - `datasets` version: 1.14.0 - Platform: macOS-10.16-x86_64-i386-64bit - Python version: 3.8.8 - PyArrow version: 6.0.0 **Environment 2.** - `datasets` version: 1.14.0 - Platform: Linux-4.18.0-305.19.1.el8_4.x86_64-x86_64-with-glibc2.28 - Python version: 3.9.7 - PyArrow version: 6.0.0
false
1,037,728,059
https://api.github.com/repos/huggingface/datasets/issues/3171
https://github.com/huggingface/datasets/issues/3171
3,171
Raise exceptions instead of using assertions for control flow
closed
4
2021-10-27T18:26:52
2021-12-23T16:40:37
2021-12-23T16:40:37
mariosasko
[ "good first issue" ]
Motivated by https://github.com/huggingface/transformers/issues/12789 in Transformers, one welcoming change would be replacing assertions with proper exceptions. The only type of assertions we should keep are those used as sanity checks. Currently, there is a total of 87 files with the `assert` statements (located under `datasets` and `src/datasets`), so when working on this, to manage the PR size, only modify 4-5 files at most before submitting a PR.
false
1,037,601,926
https://api.github.com/repos/huggingface/datasets/issues/3170
https://github.com/huggingface/datasets/pull/3170
3,170
Preserve ordering in `zip_dict`
closed
0
2021-10-27T16:07:30
2021-10-29T13:09:37
2021-10-29T13:09:37
mariosasko
[]
Replace `set` with the `unique_values` generator in `zip_dict`. This PR fixes the problem with the different ordering of the example keys across different Python sessions caused by the `zip_dict` call in `Features.decode_example`.
true
1,036,773,357
https://api.github.com/repos/huggingface/datasets/issues/3169
https://github.com/huggingface/datasets/pull/3169
3,169
Configurable max filename length in file locks
closed
2
2021-10-26T21:52:55
2021-10-28T16:14:14
2021-10-28T16:14:13
lmmx
[]
Resolve #2924 (https://github.com/huggingface/datasets/issues/2924#issuecomment-952330956) wherein the assumption of file lock maximum filename length to be 255 raises an OSError on encrypted drives (ecryptFS on Linux uses part of the lower filename, reducing the maximum filename size to 143). Allowing this limit to be set in the config module allows this to be modified by users. Will not affect Windows users, as their class passes 255 on init explicitly. Reproduced with the following example ([the first few lines of a script from Lightning Flash](https://lightning-flash.readthedocs.io/en/latest/reference/speech_recognition.html), fine-tuning a HF model): ```py import torch import flash from flash.audio import SpeechRecognition, SpeechRecognitionData from flash.core.data.utils import download_data # 1. Create the DataModule download_data("https://pl-flash-data.s3.amazonaws.com/timit_data.zip", "./data") datamodule = SpeechRecognitionData.from_json( input_fields="file", target_fields="text", train_file="data/timit/train.json", test_file="data/timit/test.json", ) ``` Which gave this traceback: ```py Traceback (most recent call last): File "lf_ft.py", line 10, in <module> datamodule = SpeechRecognitionData.from_json( File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/flash/core/data/data_module.py", line 1005, in from_json return cls.from_data_source( File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/flash/core/data/data_module.py", line 571, in from_data_source train_dataset, val_dataset, test_dataset, predict_dataset = data_source.to_datasets( File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/flash/core/data/data_source.py", line 307, in to_datasets train_dataset = self.generate_dataset(train_data, RunningStage.TRAINING) File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/flash/core/data/data_source.py", line 344, in generate_dataset data = load_data(data, mock_dataset) File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/flash/audio/speech_recognition/data.py", line 103, in load_data dataset_dict = load_dataset(self.filetype, data_files={stage: str(file)}) File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/datasets/load.py", line 1599, in load_dataset builder_instance = load_dataset_builder( File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/datasets/load.py", line 1457, in load_dataset_builder builder_instance: DatasetBuilder = builder_cls( File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/datasets/builder.py", line 285, in __init__ with FileLock(lock_path): File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/datasets/utils/filelock.py", line 323, in __enter__ self.acquire() File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/datasets/utils/filelock.py", line 272, in acquire self._acquire() File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/datasets/utils/filelock.py", line 403, in _acquire fd = os.open(self._lock_file, open_mode) OSError: [Errno 36] File name too long: '/home/louis/.cache/huggingface/datasets/_home_louis_.cache_huggingface_datasets_json_default-98e6813a547f72fa_0.0.0_c2d554c3377ea79c7664b93dc65d0803b45e3279000f993c7bfd18937fd7f426.lock' ``` Note the filename is 145 chars long: ``` >>> len("_home_louis_.cache_huggingface_datasets_json_default-98e6813a547f72fa_0.0.0_c2d554c3377ea79c7664b93dc65d0803b45e3279000f993c7bfd18937fd7f426.lock") 145 ``` After installing datasets as an editable local package and modifying the script I was running to first include: ```py import datasets datasets.config.MAX_DATASET_CONFIG_ID_READABLE_LENGTH = 143 ``` The error goes away. If I instead deliberately set the value incorrectly as 144, the OSError returns: ``` Traceback (most recent call last): File "lf_ft.py", line 14, in <module> datamodule = SpeechRecognitionData.from_json( File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/flash/core/data/data_module.py", line 1005, in from_json return cls.from_data_source( File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/flash/core/data/data_module.py", line 571, in from_data_source train_dataset, val_dataset, test_dataset, predict_dataset = data_source.to_datasets( File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/flash/core/data/data_source.py", line 307, in to_datasets train_dataset = self.generate_dataset(train_data, RunningStage.TRAINING) File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/flash/core/data/data_source.py", line 344, in generate_dataset data = load_data(data, mock_dataset) File "/home/louis/miniconda3/envs/w2vlf/lib/python3.8/site-packages/flash/audio/speech_recognition/data.py", line 103, in load_data dataset_dict = load_dataset(self.filetype, data_files={stage: str(file)}) File "/home/louis/dev/hf_datasets/src/datasets/load.py", line 1605, in load_dataset builder_instance = load_dataset_builder( File "/home/louis/dev/hf_datasets/src/datasets/load.py", line 1463, in load_dataset_builder builder_instance: DatasetBuilder = builder_cls( File "/home/louis/dev/hf_datasets/src/datasets/builder.py", line 285, in __init__ with FileLock(lock_path): File "/home/louis/dev/hf_datasets/src/datasets/utils/filelock.py", line 326, in __enter__ self.acquire() File "/home/louis/dev/hf_datasets/src/datasets/utils/filelock.py", line 275, in acquire self._acquire() File "/home/louis/dev/hf_datasets/src/datasets/utils/filelock.py", line 406, in _acquire fd = os.open(self._lock_file, open_mode) OSError: [Errno 36] File name too long: '/home/louis/.cache/huggingface/datasets/_home_louis_.cache_huggingface_datasets_json_default-32c812b5c1272d64_0.0.0_c2d554c3377ea79c7664b93dc65d0803b45e3279...-5794079643713042223.lock' ```
true
1,036,673,263
https://api.github.com/repos/huggingface/datasets/issues/3168
https://github.com/huggingface/datasets/issues/3168
3,168
OpenSLR/83 is empty
closed
3
2021-10-26T19:42:21
2021-10-29T10:04:09
2021-10-29T10:04:09
tyrius02
[ "bug" ]
## Describe the bug As the summary says, openslr / SLR83 / train is empty. The dataset returned after loading indicates there are **zero** rows. The correct number should be **17877**. ## Steps to reproduce the bug ```python import datasets datasets.load_dataset('openslr', 'SLR83') ``` ## Expected results ``` DatasetDict({ train: Dataset({ features: ['path', 'audio', 'sentence'], num_rows: 17877 }) }) ``` ## Actual results ``` DatasetDict({ train: Dataset({ features: ['path', 'audio', 'sentence'], num_rows: 0 }) }) ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.14.1.dev0 (master HEAD) - Platform: Ubuntu 20.04 - Python version: 3.7.10 - PyArrow version: 3.0.0
false
1,036,488,992
https://api.github.com/repos/huggingface/datasets/issues/3167
https://github.com/huggingface/datasets/issues/3167
3,167
bookcorpusopen no longer works
closed
3
2021-10-26T16:06:15
2021-11-17T15:53:46
2021-11-17T15:53:46
lucadiliello
[ "bug" ]
## Describe the bug When using the latest version of datasets (1.14.0), I cannot use the `bookcorpusopen` dataset. The process blocks always around `9924 examples [00:06, 1439.61 examples/s]` when preparing the dataset. I also noticed that after half an hour the process is automatically killed because of the RAM usage (the machine has 1TB of RAM...). This did not happen with 1.4.1. I tried also `rm -rf ~/.cache/huggingface` but did not help. Changing python version between 3.7, 3.8 and 3.9 did not help too. ## Steps to reproduce the bug ```python import datasets d = datasets.load_dataset('bookcorpusopen') ``` ## Expected results A clear and concise description of the expected results. ## Actual results Specify the actual results or traceback. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.14.0 - Platform: Linux-5.4.0-1054-aws-x86_64-with-glibc2.27 - Python version: 3.9.7 - PyArrow version: 4.0.1
false
1,036,450,283
https://api.github.com/repos/huggingface/datasets/issues/3166
https://github.com/huggingface/datasets/pull/3166
3,166
Deprecate prepare_module
closed
1
2021-10-26T15:28:24
2021-11-05T09:27:37
2021-11-05T09:27:36
albertvillanova
[]
In version 1.13, `prepare_module` was deprecated. This PR adds a deprecation warning and removes it from all the library, using `dataset_module_factory` or `metric_module_factory` instead. Fix #3165.
true
1,036,448,998
https://api.github.com/repos/huggingface/datasets/issues/3165
https://github.com/huggingface/datasets/issues/3165
3,165
Deprecate prepare_module
closed
0
2021-10-26T15:27:15
2021-11-05T09:27:36
2021-11-05T09:27:36
albertvillanova
[]
In version 1.13, `prepare_module` was deprecated. Add deprecation warning and remove its usage from all the library.
false
1,035,662,830
https://api.github.com/repos/huggingface/datasets/issues/3164
https://github.com/huggingface/datasets/issues/3164
3,164
Add raw data files to the Hub with GitHub LFS for canonical dataset
closed
3
2021-10-25T23:28:21
2021-10-30T19:54:51
2021-10-30T19:54:51
zlucia
[]
I'm interested in sharing the CaseHOLD dataset (https://arxiv.org/abs/2104.08671) as a canonical dataset on the HuggingFace Hub and would like to add the raw data files to the Hub with GitHub LFS, since it seems like a more sustainable long term storage solution, compared to other storage solutions available to my team. From what I can tell, this option is not immediately supported if one follows the sharing steps detailed here: [https://huggingface.co/docs/datasets/share_dataset.html#sharing-a-canonical-dataset](https://huggingface.co/docs/datasets/share_dataset.html#sharing-a-canonical-dataset), since GitHub LFS is not supported for public forks. Is there a way to request this? Thanks!
false
1,035,475,061
https://api.github.com/repos/huggingface/datasets/issues/3163
https://github.com/huggingface/datasets/pull/3163
3,163
Add Image feature
closed
14
2021-10-25T19:07:48
2021-12-30T06:37:21
2021-12-06T17:49:02
mariosasko
[]
Adds the Image feature. This feature is heavily inspired by the recently added Audio feature (#2324). Currently, this PR is pretty simple. Some considerations that need further discussion: * I've decided to use `Pillow`/`PIL` as the image decoding library. Another candidate I considered is `torchvision`, mostly because of its `accimage` backend, which should be faster for loading `jpeg` images than `Pillow`. However, `torchvision`'s io module only supports png and jpeg images, has `torch` as a hard dependency, and requires magic to work with image bytes ( `torch.ByteTensor(torch.ByteStorage.from_buffer(image_bytes)))`). * Currently, I'm converting `PIL`'s `Image` type to `np.ndarray`. The vision models in Transformers such as ViT prefer the raw `Image` type and not the decoded tensors, so there is a small overhead due to [this conversion](https://github.com/huggingface/transformers/blob/3e8761ab8077e3bb243fe2f78b2a682bd2257cf1/src/transformers/image_utils.py#L62-L73). IMO this is justified to keep this part aligned with the Audio feature, which also returns `np.ndarray`. What do you think? * Still have to work on the channel decoding logic: * PyTorch prefers the channel-first ordering (C, H, W); TF and Flax the channel-last ordering (H, W, C). One cool feature would be adjusting the channel order based on the selected formatter (`torch`, `tf`, `jax`). * By default, `Image.open` returns images of shape (H, W, C). However, ViT's feature extractor expects the format (C, H, W) if the image is passed as an array (explained [here](https://huggingface.co/transformers/model_doc/vit.html#transformers.ViTFeatureExtractor.__call__)), so I'm more inclined to the format (C, H, W). Which one do you prefer, (C, H, W) or (H, W, C)? * Are there any options you'd like to see? (the user could change those via `cast_column`, such as `sampling_rate` in the Audio feature) TODOs: * [x] tests * in subsequent PRs: * docs - a section in the docs, which gives some additional info on the Image and Audio feature and compares them to `ArrayND` * streaming (waiting for #3129 and #3133 to get merged first) * update the image tasks and the datasets to use the new feature * Image/Audio formatting [Colab Notebook](https://colab.research.google.com/drive/1mIrTnqTVkWLJWoBzT1ABSe-LFelIep1c?usp=sharing) where you can play with this feature. I'm also adding a link to the [Image](https://github.com/tensorflow/datasets/blob/7ac7d506488d46038a5854961d068926b3f93c7f/tensorflow_datasets/core/features/image_feature.py#L155) feature in TFDS because one of our goals is to parse TFDS scripts eventually, so our Image feature has to (at least) support all the formats theirs does. Feel free to cc anyone who might be interested. P.S. Please ignore the changes in the `datasets/**/*.py` files 😄.
true
1,035,462,136
https://api.github.com/repos/huggingface/datasets/issues/3162
https://github.com/huggingface/datasets/issues/3162
3,162
`datasets-cli test` should work with datasets without scripts
open
5
2021-10-25T18:52:30
2021-11-25T16:04:29
null
sashavor
[ "enhancement" ]
It would be really useful to be able to run `datasets-cli test`for datasets that don't have scripts attached to them (whether the datasets are private or not). I wasn't able to run the script for a private test dataset that I had created on the hub (https://huggingface.co/datasets/huggingface/DataMeasurementsTest/tree/main) -- although @lhoestq came to save the day!
false
1,035,444,292
https://api.github.com/repos/huggingface/datasets/issues/3161
https://github.com/huggingface/datasets/pull/3161
3,161
Add riddle_sense dataset
closed
2
2021-10-25T18:30:56
2021-11-04T14:01:15
2021-11-04T14:01:15
ziyiwu9494
[]
Adding a new dataset for QA with riddles. I'm confused about the tagging process because it looks like the streamlit app loads data from the current repo, so is it something that should be done after merging or off my fork?
true
1,035,274,640
https://api.github.com/repos/huggingface/datasets/issues/3160
https://github.com/huggingface/datasets/pull/3160
3,160
Better error msg if `len(predictions)` doesn't match `len(references)` in metrics
closed
2
2021-10-25T15:25:05
2021-11-05T11:44:59
2021-11-05T09:31:02
mariosasko
[]
Improve the error message in `Metric.add_batch` if `len(predictions)` doesn't match `len(references)`. cc: @BramVanroy (feel free to test this code on your examples and review this PR)
true
1,035,174,560
https://api.github.com/repos/huggingface/datasets/issues/3159
https://github.com/huggingface/datasets/pull/3159
3,159
Make inspect.get_dataset_config_names always return a non-empty list
closed
4
2021-10-25T13:59:43
2021-10-29T13:14:37
2021-10-28T05:44:49
albertvillanova
[]
Make all named configs cases, so that no special unnamed config case needs to be handled differently. Fix #3135.
true
1,035,158,070
https://api.github.com/repos/huggingface/datasets/issues/3158
https://github.com/huggingface/datasets/pull/3158
3,158
Fix string encoding for Value type
closed
1
2021-10-25T13:44:13
2021-10-25T14:12:06
2021-10-25T14:12:05
lhoestq
[]
Some metrics have `string` features but currently it fails if users pass integers instead. Indeed feature encoding that handles the conversion of the user's objects to the right python type is missing a case for `string`, while it already works as expected for integers, floats and booleans Here is an example code that didn't work previously, but that works with this fix: ```python import datasets # Note that 'id' is an integer while the SQuAD metric uses strings predictions = [{'prediction_text': '1976', 'id': 5}] references = [{'answers': {'answer_start': [97], 'text': ['1976']}, 'id': 5}] squad_metric = datasets.load_metric("squad") squad_metric.add_batch(predictions=predictions, references=references) results = squad_metric.compute() # {'exact_match': 100.0, 'f1': 100.0} ``` cc @sgugger @philschmid
true
1,034,775,165
https://api.github.com/repos/huggingface/datasets/issues/3157
https://github.com/huggingface/datasets/pull/3157
3,157
Fixed: duplicate parameter and missing parameter in docstring
closed
0
2021-10-25T07:26:00
2021-10-25T14:02:19
2021-10-25T14:02:19
PanQiWei
[]
changing duplicate parameter `data_files` in `DatasetBuilder.__init__` to the missing parameter `data_dir`
true
1,034,468,757
https://api.github.com/repos/huggingface/datasets/issues/3155
https://github.com/huggingface/datasets/issues/3155
3,155
Illegal instruction (core dumped) at datasets import
closed
1
2021-10-24T17:21:36
2021-11-18T19:07:04
2021-11-18T19:07:03
hacobe
[ "bug" ]
## Describe the bug I install datasets using conda and when I import datasets I get: "Illegal instruction (core dumped)" ## Steps to reproduce the bug ``` conda create --prefix path/to/env conda activate path/to/env conda install -c huggingface -c conda-forge datasets # exits with output "Illegal instruction (core dumped)" python -m datasets ``` ## Environment info When I run "datasets-cli env", I also get "Illegal instruction (core dumped)" If I run the following commands: ``` conda create --prefix path/to/another/new/env conda activate path/to/another/new/env conda install -c huggingface transformers transformers-cli env ``` Then I get: - `transformers` version: 4.11.3 - Platform: Linux-5.4.0-67-generic-x86_64-with-glibc2.17 - Python version: 3.8.12 - PyTorch version (GPU?): not installed (NA) - Tensorflow version (GPU?): not installed (NA) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using GPU in script?: No - Using distributed or parallel set-up in script?: No Let me know what additional information you need in order to debug this issue. Thanks in advance!
false
1,034,361,806
https://api.github.com/repos/huggingface/datasets/issues/3154
https://github.com/huggingface/datasets/issues/3154
3,154
Sacrebleu unexpected behaviour/requirement for data format
closed
2
2021-10-24T08:55:33
2021-10-31T09:08:32
2021-10-31T09:08:31
BramVanroy
[ "bug" ]
## Describe the bug When comparing with the original `sacrebleu` implementation, the `datasets` implementation does some strange things that I do not quite understand. This issue was triggered when I was trying to implement TER and found the datasets implementation of BLEU [here](https://github.com/huggingface/datasets/pull/3153). In the below snippet, the original sacrebleu snippet works just fine whereas the datasets implementation throws an error. ## Steps to reproduce the bug ```python import sacrebleu import datasets refs = [ ['The dog bit the man.', 'It was not unexpected.', 'The man bit him first.'], ['The dog had bit the man.', 'No one was surprised.', 'The man had bitten the dog.'], ] hyps = ['The dog bit the man.', "It wasn't surprising.", 'The man had just bitten him.'] expected_bleu = 48.530827 ds_bleu = datasets.load_metric("sacrebleu") bleu_score_sb = sacrebleu.corpus_bleu(hyps, refs).score print(bleu_score_sb, expected_bleu) # works: 48.5308... bleu_score_ds = ds_bleu.compute(predictions=hyps, references=refs)["score"] print(bleu_score_ds, expected_bleu) # ValueError: Predictions and/or references don't match the expected format. ``` This seems to be related to how datasets forces the features format here: https://github.com/huggingface/datasets/blob/87c71b9c29a40958973004910f97e4892559dfed/metrics/sacrebleu/sacrebleu.py#L94-L99 and then manipulates the references during the compute stage here https://github.com/huggingface/datasets/blob/87c71b9c29a40958973004910f97e4892559dfed/metrics/sacrebleu/sacrebleu.py#L119-L122 I do not quite understand why that is required since sacrebleu handles argument parsing quite well [by itself](https://github.com/mjpost/sacrebleu/blob/2787185dd0f8d224c72ee5a831d163c2ac711a47/sacrebleu/metrics/base.py#L229). ## Actual results Traceback (most recent call last): File "C:\Users\bramv\AppData\Roaming\JetBrains\PyCharm2020.3\scratches\scratch_23.py", line 23, in <module> bleu_score_ds = ds_bleu.compute(predictions=hyps, references=refs)["score"] File "C:\dev\python\datasets\src\datasets\metric.py", line 392, in compute self.add_batch(predictions=predictions, references=references) File "C:\dev\python\datasets\src\datasets\metric.py", line 439, in add_batch raise ValueError( ValueError: Predictions and/or references don't match the expected format. Expected format: {'predictions': Value(dtype='string', id='sequence'), 'references': Sequence(feature=Value(dtype='string', id='sequence'), length=-1, id='references')}, Input predictions: ['The dog bit the man.', "It wasn't surprising.", 'The man had just bitten him.'], Input references: [['The dog bit the man.', 'It was not unexpected.', 'The man bit him first.'], ['The dog had bit the man.', 'No one was surprised.', 'The man had bitten the dog.']] ## Environment info - `datasets` version: 1.14.1.dev0 - Platform: Windows-10-10.0.19041-SP0 - Python version: 3.9.2 - PyArrow version: 4.0.1
false
1,034,179,198
https://api.github.com/repos/huggingface/datasets/issues/3153
https://github.com/huggingface/datasets/pull/3153
3,153
Add TER (as implemented in sacrebleu)
closed
1
2021-10-23T14:26:45
2021-11-02T11:04:11
2021-11-02T11:04:11
BramVanroy
[]
Implements TER (Translation Edit Rate) as per its implementation in sacrebleu. Sacrebleu for BLEU scores is already implemented in `datasets` so I thought this would be a nice addition. I started from the sacrebleu implementation, as the two metrics have a lot in common. Verified with sacrebleu's [testing suite](https://github.com/mjpost/sacrebleu/blob/078c440168c6adc89ba75fe6d63f0d922d42bcfe/test/test_ter.py) that this indeed works as intended. ```python import datasets test_cases = [ (['aaaa bbbb cccc dddd'], ['aaaa bbbb cccc dddd'], 0), # perfect match (['dddd eeee ffff'], ['aaaa bbbb cccc'], 1), # no overlap ([''], ['a'], 1), # corner case, empty hypothesis (['d e f g h a b c'], ['a b c d e f g h'], 1 / 8), # a single shift fixes MT ( [ 'wählen Sie " Bild neu berechnen , " um beim Ändern der Bildgröße Pixel hinzuzufügen oder zu entfernen , damit das Bild ungefähr dieselbe Größe aufweist wie die andere Größe .', 'wenn Sie alle Aufgaben im aktuellen Dokument aktualisieren möchten , wählen Sie im Menü des Aufgabenbedienfelds die Option " Alle Aufgaben aktualisieren . "', 'klicken Sie auf der Registerkarte " Optionen " auf die Schaltfläche " Benutzerdefiniert " und geben Sie Werte für " Fehlerkorrektur-Level " und " Y / X-Verhältnis " ein .', 'Sie können beispielsweise ein Dokument erstellen , das ein Auto über die Bühne enthält .', 'wählen Sie im Dialogfeld " Neu aus Vorlage " eine Vorlage aus und klicken Sie auf " Neu . "', ], [ 'wählen Sie " Bild neu berechnen , " um beim Ändern der Bildgröße Pixel hinzuzufügen oder zu entfernen , damit die Darstellung des Bildes in einer anderen Größe beibehalten wird .', 'wenn Sie alle Aufgaben im aktuellen Dokument aktualisieren möchten , wählen Sie im Menü des Aufgabenbedienfelds die Option " Alle Aufgaben aktualisieren . "', 'klicken Sie auf der Registerkarte " Optionen " auf die Schaltfläche " Benutzerdefiniert " und geben Sie für " Fehlerkorrektur-Level " und " Y / X-Verhältnis " niedrigere Werte ein .', 'Sie können beispielsweise ein Dokument erstellen , das ein Auto enthalt , das sich über die Bühne bewegt .', 'wählen Sie im Dialogfeld " Neu aus Vorlage " eine Vorlage aus und klicken Sie auf " Neu . "', ], 0.136 # realistic example from WMT dev data (2019) ), ] ter = datasets.load_metric(r"path\to\datasets\metrics\ter") predictions = ["hello there general kenobi", "foo bar foobar"] references = [["hello there general kenobi", "hello there !"], ["foo bar foobar", "foo bar foobar"]] print(ter.compute(predictions=predictions, references=references)) for hyp, ref, score in test_cases: # Note the reference transformation which is different from scarebleu's input format results = ter.compute(predictions=hyp, references=[[r] for r in ref]) assert 100*score == results["score"], f"expected {100*score}, got {results['score']}" ```
true
1,034,039,379
https://api.github.com/repos/huggingface/datasets/issues/3152
https://github.com/huggingface/datasets/pull/3152
3,152
Fix some typos in the documentation
closed
0
2021-10-23T01:38:35
2021-10-25T14:27:36
2021-10-25T14:03:48
h4iku
[]
null
true
1,033,890,501
https://api.github.com/repos/huggingface/datasets/issues/3151
https://github.com/huggingface/datasets/pull/3151
3,151
Re-add faiss to windows testing suite
closed
0
2021-10-22T19:34:29
2021-11-02T10:47:34
2021-11-02T10:06:03
BramVanroy
[]
In recent versions, `faiss-cpu` seems to be available for Windows as well. See the [PyPi page](https://pypi.org/project/faiss-cpu/#files) to confirm. We can therefore included it for Windows in the setup file. At first tests didn't pass due to problems with permissions as caused by `NamedTemporaryFile` on Windows. This built-in library is notoriously poor in playing nice on Windows. The required change isn't pretty, but it works. First set `delete=False` to not automatically try to delete the file on `exit`. Then, manually delete the file with `unlink`. It's weird, I know, but it works. ```python with tempfile.NamedTemporaryFile(delete=False) as tmp_file: # do stuff os.unlink(tmp_file.name) ``` closes #3150
true
1,033,831,530
https://api.github.com/repos/huggingface/datasets/issues/3150
https://github.com/huggingface/datasets/issues/3150
3,150
Faiss _is_ available on Windows
closed
1
2021-10-22T18:07:16
2021-11-02T10:06:03
2021-11-02T10:06:03
BramVanroy
[]
In the setup file, I find the following: https://github.com/huggingface/datasets/blob/87c71b9c29a40958973004910f97e4892559dfed/setup.py#L171 However, FAISS does install perfectly fine on Windows on my system. You can also confirm this on the [PyPi page](https://pypi.org/project/faiss-cpu/#files), where Windows wheels are available. Maybe this was true for older versions? For current versions, this can be removed I think. (This isn't really a bug but didn't know how else to tag.) If you agree I can do a quick PR and remove that line.
false
1,033,747,625
https://api.github.com/repos/huggingface/datasets/issues/3149
https://github.com/huggingface/datasets/pull/3149
3,149
Add CMU Hinglish DoG Dataset for MT
closed
2
2021-10-22T16:17:25
2021-11-15T11:36:42
2021-11-15T10:27:45
Ishan-Kumar2
[]
Address part of #2841 Added the CMU Hinglish DoG Dataset as in GLUECoS. Added it as a seperate dataset as unlike other tasks of GLUE CoS this can't be evaluated for a BERT like model. Consists of parallel dataset between Hinglish (Hindi-English) and English, can be used for Machine Translation between the two. The data processing part is inspired from the GLUECoS repo [here](https://github.com/microsoft/GLUECoS/blob/7fdc51653e37a32aee17505c47b7d1da364fa77e/Data/Preprocess_Scripts/preprocess_mt_en_hi.py) The dummy data part is not working properly, it shows ``` UnboundLocalError: local variable 'generator_splits' referenced before assignment ``` when I run without ``--auto_generate``. Please let me know how I can fix that. Thanks
true
1,033,685,208
https://api.github.com/repos/huggingface/datasets/issues/3148
https://github.com/huggingface/datasets/issues/3148
3,148
Streaming with num_workers != 0
closed
4
2021-10-22T15:07:17
2022-07-04T12:14:58
2022-07-04T12:14:58
justheuristic
[ "bug" ]
## Describe the bug When using dataset streaming with pytorch DataLoader, the setting num_workers to anything other than 0 causes the code to freeze forever before yielding the first batch. The code owner is likely @lhoestq ## Steps to reproduce the bug For your convenience, we've prepped a colab notebook that reproduces the bug https://colab.research.google.com/drive/1Mgl0oTZSNIE3UeGl_oX9wPCOIxRg19h1?usp=sharing ```python !pip install datasets==1.14.0 should_freeze_forever = True # ^-- set this to True in order to freeze forever, set to False in order to work normally import torch from datasets import load_dataset data = load_dataset("oscar", "unshuffled_deduplicated_bn", split="train", streaming=True) data = data.map(lambda x: {"text": x["text"], "orig": f"oscar[{x['id']}]"}, batched=True) data = data.shuffle(100, seed=1337) data = data.with_format("torch") loader = torch.utils.data.DataLoader(data, batch_size=2, num_workers=2 if should_freeze_forever else 0) # v-- the code should freeze forever at this line for i, row in enumerate(loader): print(row) if i > 10: break print("DONE!") ``` ## Expected results The code should not freeze forever with num_workers=2 ## Actual results The code freezes forever with num_workers=2 ## Environment info - `datasets` version: 1.14.0 (also found in previous versions) - Platform: google colab (also locally) - Python version: 3.7, (also 3.8) - PyArrow version: 3.0.0
false
1,033,607,659
https://api.github.com/repos/huggingface/datasets/issues/3147
https://github.com/huggingface/datasets/pull/3147
3,147
Fix CLI test to ignore verfications when saving infos
closed
0
2021-10-22T13:52:46
2021-10-27T08:01:50
2021-10-27T08:01:49
albertvillanova
[]
Fix #3146.
true
1,033,605,947
https://api.github.com/repos/huggingface/datasets/issues/3146
https://github.com/huggingface/datasets/issues/3146
3,146
CLI test command throws NonMatchingSplitsSizesError when saving infos
closed
0
2021-10-22T13:50:53
2021-10-27T08:01:49
2021-10-27T08:01:49
albertvillanova
[ "bug" ]
When trying to generate a datset JSON metadata, a `NonMatchingSplitsSizesError` is thrown: ``` $ datasets-cli test datasets/arabic_billion_words --save_infos --all_configs Testing builder 'Alittihad' (1/10) Downloading and preparing dataset arabic_billion_words/Alittihad (download: 332.13 MiB, generated: Unknown size, post-processed: Unknown size, total: 332.13 MiB) to .cache\arabic_billion_words\Alittihad\1.1.0\8175ff1c9714c6d5d15b1141b6042e5edf048276bb81a9c14e35e149a7a62ae4... Traceback (most recent call last): File "path\huggingface\datasets\.venv\Scripts\datasets-cli-script.py", line 33, in <module> sys.exit(load_entry_point('datasets', 'console_scripts', 'datasets-cli')()) File "path\huggingface\datasets\src\datasets\commands\datasets_cli.py", line 33, in main service.run() File "path\huggingface\datasets\src\datasets\commands\test.py", line 144, in run builder.download_and_prepare( File "path\huggingface\datasets\src\datasets\builder.py", line 607, in download_and_prepare self._download_and_prepare( File "path\huggingface\datasets\src\datasets\builder.py", line 709, in _download_and_prepare verify_splits(self.info.splits, split_dict) File "path\huggingface\datasets\src\datasets\utils\info_utils.py", line 74, in verify_splits raise NonMatchingSplitsSizesError(str(bad_splits)) datasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=0, num_examples=0, dataset_name='arabic_billion_words'), 'recorded': SplitInfo(name='train', num_bytes=1601790302, num_examples=349342, dataset_name='arabic_billion_words')}] ``` This is due because a previous run generated a wrong `dataset_info.json`. This error can be avoided by passing `--ignore_verifications`, but I think this should be assumed when passing `--save_infos`.
false
1,033,580,009
https://api.github.com/repos/huggingface/datasets/issues/3145
https://github.com/huggingface/datasets/issues/3145
3,145
[when Image type will exist] provide a way to get the data as binary + filename
closed
4
2021-10-22T13:23:49
2021-12-22T11:05:37
2021-12-22T11:05:36
severo
[ "enhancement", "dataset-viewer" ]
**Is your feature request related to a problem? Please describe.** When a dataset cell contains a value of type Image (be it from a remote URL, an Array2D/3D, or any other way to represent images), I want to be able to write the image to the disk, with the correct filename, and optionally to know its mimetype, in order to serve it on the web. Note: this issue would apply exactly the same for the `Audio` type. **Describe the solution you'd like** If a "cell" has the type `Image`, provide a way to get the binary content of the file, and the filename, eg as: ```python filename: str data: bytes ``` **Describe alternatives you've considered** A way to write the cell to the disk (passing a local directory), and then return the pathname, filename, and mimetype.
false
1,033,573,760
https://api.github.com/repos/huggingface/datasets/issues/3144
https://github.com/huggingface/datasets/issues/3144
3,144
Infer the features if missing
closed
1
2021-10-22T13:17:33
2022-09-08T08:23:10
2022-09-08T08:23:10
severo
[ "enhancement", "dataset-viewer" ]
**Is your feature request related to a problem? Please describe.** Some datasets, in particular community datasets, have no info file, thus no features. **Describe the solution you'd like** If a dataset has no features, the first loaded data (5-10 rows) could be used to infer the type. Related: `datasets` would provide a way to load the data, and get the rows AND the features as the result. **Describe alternatives you've considered** The HF hub could also provide some UI to help the dataset maintainers to explicit the types of their rows, or automatically infer them as an initial proposal.
false
1,033,569,655
https://api.github.com/repos/huggingface/datasets/issues/3143
https://github.com/huggingface/datasets/issues/3143
3,143
Provide a way to check if the features (in info) match with the data of a split
open
1
2021-10-22T13:13:36
2021-10-22T13:17:56
null
severo
[ "enhancement", "dataset-viewer" ]
**Is your feature request related to a problem? Please describe.** I understand that currently the data loaded has not always the type described in the info features **Describe the solution you'd like** Provide a way to check if the rows have the type described by info features **Describe alternatives you've considered** Always check it, and raise an error when loading the data if their type doesn't match the features.
false
1,033,566,034
https://api.github.com/repos/huggingface/datasets/issues/3142
https://github.com/huggingface/datasets/issues/3142
3,142
Provide a way to write a streamed dataset to the disk
open
2
2021-10-22T13:09:53
2024-01-12T07:26:43
null
severo
[ "enhancement", "dataset-viewer" ]
**Is your feature request related to a problem? Please describe.** The streaming mode allows to get the 100 first rows of a dataset very quickly. But it does not cache the answer, so a posterior call to get the same 100 rows will send a request to the server again and again. **Describe the solution you'd like** Provide a way to write the streamed rows of a dataset on the disk, and to load from it later. **Describe alternatives you've considered** Provide a third mode: `lazy`, which would use the local cache for the data that have already been fetched previously, and use streaming to get the rest of the requested data.
false
1,033,555,910
https://api.github.com/repos/huggingface/datasets/issues/3141
https://github.com/huggingface/datasets/pull/3141
3,141
Fix caching bugs
closed
0
2021-10-22T12:59:25
2021-10-22T20:52:08
2021-10-22T13:47:05
mariosasko
[]
This PR fixes some caching bugs (most likely introduced in the latest refactor): * remove ")" added by accident in the dataset dir name * correctly pass the namespace kwargs in `CachedDatasetModuleFactory` * improve the warning message if `HF_DATASETS_OFFLINE is `True`
true
1,033,524,079
https://api.github.com/repos/huggingface/datasets/issues/3139
https://github.com/huggingface/datasets/issues/3139
3,139
Fix file/directory deletion on Windows
open
0
2021-10-22T12:22:08
2021-10-22T12:22:08
null
mariosasko
[ "bug" ]
Currently, on Windows, some attempts to delete a dataset file/directory will fail with the `PerimissionError`. Examples: - download a dataset, then force redownload it in the same session while keeping a reference to the downloaded dataset ```python from datasets import load_dataset dset = load_dataset("sst", split="train") dset = load_dataset("sst", split="train", download_mode="force_redownload") ``` - try to clean up the cache files while keeping a reference to those files (via the mapped dataset): ```python from datasets import load_dataset dset = load_dataset("sst", split="train") dset_mapped = dset.map(lambda _: {"dummy_col": 1}) dset.cleanup_cache_files() ``` We should fix those.
false
1,033,379,997
https://api.github.com/repos/huggingface/datasets/issues/3138
https://github.com/huggingface/datasets/issues/3138
3,138
More fine-grained taxonomy of error types
open
1
2021-10-22T09:35:29
2022-09-20T13:04:42
null
severo
[ "enhancement", "dataset-viewer" ]
**Is your feature request related to a problem? Please describe.** Exceptions like `FileNotFoundError` can be raised by different parts of the code, and it's hard to detect which one did **Describe the solution you'd like** Give a specific exception type for every group of similar errors **Describe alternatives you've considered** Rely on the error message, using regex
false
1,033,363,652
https://api.github.com/repos/huggingface/datasets/issues/3137
https://github.com/huggingface/datasets/pull/3137
3,137
Fix numpy deprecation warning for ragged tensors
closed
1
2021-10-22T09:17:46
2021-10-22T16:04:15
2021-10-22T16:04:14
lhoestq
[]
Numpy shows a deprecation warning when we call `np.array` on a list of ragged tensors without specifying the `dtype`. If their shapes match, the tensors can be collated together, otherwise the resulting array should have `dtype=np.object`. Fix #3084 cc @Rocketknight1
true
1,033,360,396
https://api.github.com/repos/huggingface/datasets/issues/3136
https://github.com/huggingface/datasets/pull/3136
3,136
Fix script of Arabic Billion Words dataset to return all data
closed
0
2021-10-22T09:14:24
2021-10-22T13:28:41
2021-10-22T13:28:40
albertvillanova
[]
The script has a bug and only parses and generates a portion of the entire dataset. This PR fixes the loading script so that is properly parses the entire dataset. Current implementation generates the same number of examples as reported in the [original paper](https://arxiv.org/abs/1611.04033) for all configurations except for one: - For "Youm7" we generate more examples (1172136) than the ones reported by the paper (1025027) | | Number of examples | Number of examples according to the source | |:---------------|-------------------:|-----:| | Alittihad | 349342 |349342 | | Almasryalyoum | 291723 |291723 | | Almustaqbal | 446873 |446873 | | Alqabas | 817274 |817274 | | Echoroukonline | 139732 |139732 | | Ryiadh | 858188 | 858188 | | Sabanews | 92149 |92149 | | SaudiYoum | 888068 |888068 | | Techreen | 314597 |314597 | | Youm7 | 1172136 |1025027 | Fix #3126.
true
1,033,294,299
https://api.github.com/repos/huggingface/datasets/issues/3135
https://github.com/huggingface/datasets/issues/3135
3,135
Make inspect.get_dataset_config_names always return a non-empty list of configs
closed
2
2021-10-22T08:02:50
2021-10-28T05:44:49
2021-10-28T05:44:49
severo
[ "enhancement", "dataset-viewer" ]
**Is your feature request related to a problem? Please describe.** Currently, some datasets have a configuration, while others don't. It would be simpler for the user to always have configuration names to refer to **Describe the solution you'd like** In that sense inspect.get_dataset_config_names should always return at least one configuration name, be it `default` or `Check___region_1` (for community datasets like `Check/region_1`). https://github.com/huggingface/datasets/blob/c5747a5e1dde2670b7f2ca6e79e2ffd99dff85af/src/datasets/inspect.py#L161
false
1,033,251,755
https://api.github.com/repos/huggingface/datasets/issues/3134
https://github.com/huggingface/datasets/issues/3134
3,134
Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/1.11.0/metrics/rouge/rouge.py
closed
4
2021-10-22T07:07:52
2023-09-14T01:19:45
2022-01-19T14:02:31
yanan1116
[ "bug" ]
datasets version: 1.12.1 `metric = datasets.load_metric('rouge')` The error: > ConnectionError Traceback (most recent call last) > <ipython-input-3-dd10a0c5212f> in <module> > ----> 1 metric = datasets.load_metric('rouge') > > /usr/local/lib/python3.6/dist-packages/datasets/load.py in load_metric(path, config_name, process_id, num_process, cache_dir, experiment_id, keep_in_memory, download_config, download_mode, script_version, **metric_init_kwargs) > 613 download_config=download_config, > 614 download_mode=download_mode, > --> 615 dataset=False, > 616 ) > 617 metric_cls = import_main_class(module_path, dataset=False) > > /usr/local/lib/python3.6/dist-packages/datasets/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, dynamic_modules_path, return_resolved_file_path, **download_kwargs) > 328 file_path = hf_github_url(path=path, name=name, dataset=dataset, version=script_version) > 329 try: > --> 330 local_path = cached_path(file_path, download_config=download_config) > 331 except FileNotFoundError: > 332 if script_version is not None: > > /usr/local/lib/python3.6/dist-packages/datasets/utils/file_utils.py in cached_path(url_or_filename, download_config, **download_kwargs) > 296 use_etag=download_config.use_etag, > 297 max_retries=download_config.max_retries, > --> 298 use_auth_token=download_config.use_auth_token, > 299 ) > 300 elif os.path.exists(url_or_filename): > > /usr/local/lib/python3.6/dist-packages/datasets/utils/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, resume_download, user_agent, local_files_only, use_etag, max_retries, use_auth_token) > 603 raise FileNotFoundError("Couldn't find file at {}".format(url)) > 604 _raise_if_offline_mode_is_enabled(f"Tried to reach {url}") > --> 605 raise ConnectionError("Couldn't reach {}".format(url)) > 606 > 607 # Try a second time > > ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/1.11.0/metrics/rouge/rouge.py Is there any remedy to solve the connection issue ?
false
1,032,511,710
https://api.github.com/repos/huggingface/datasets/issues/3133
https://github.com/huggingface/datasets/pull/3133
3,133
Support Audio feature in streaming mode
closed
0
2021-10-21T13:37:57
2021-11-12T14:13:05
2021-11-12T14:13:04
albertvillanova
[]
Fix #3132.
true
1,032,505,430
https://api.github.com/repos/huggingface/datasets/issues/3132
https://github.com/huggingface/datasets/issues/3132
3,132
Support Audio feature in streaming mode
closed
0
2021-10-21T13:32:18
2021-11-12T14:13:04
2021-11-12T14:13:04
albertvillanova
[ "enhancement" ]
Currently, Audio feature is only supported for non-streaming datasets. Due to the large size of many speech datasets, we should also support Audio feature in streaming mode.
false
1,032,309,865
https://api.github.com/repos/huggingface/datasets/issues/3131
https://github.com/huggingface/datasets/issues/3131
3,131
Add ADE20k
closed
1
2021-10-21T10:13:09
2023-01-27T14:40:20
2023-01-27T14:40:20
NielsRogge
[ "dataset request", "vision" ]
## Adding a Dataset - **Name:** ADE20k (actually it's called the MIT Scene Parsing Benchmark, it's actually a subset of ADE20k but a lot of authors still call it ADE20k) - **Description:** A semantic segmentation dataset, consisting of 150 classes. - **Paper:** http://people.csail.mit.edu/bzhou/publication/scene-parse-camera-ready.pdf - **Data:** http://sceneparsing.csail.mit.edu/ - **Motivation:** I am currently adding Transformer-based semantic segmentation models that achieve SOTA on this dataset. It would be great to directly access this dataset using HuggingFace Datasets, in order to make example scripts in HuggingFace Transformers. Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
false
1,032,299,417
https://api.github.com/repos/huggingface/datasets/issues/3130
https://github.com/huggingface/datasets/pull/3130
3,130
Create SECURITY.md
closed
1
2021-10-21T10:03:03
2021-10-21T14:33:28
2021-10-21T14:31:50
zidingz
[]
To let the repository confirm feedback@huggingface.co as its security contact.
true
1,032,234,167
https://api.github.com/repos/huggingface/datasets/issues/3129
https://github.com/huggingface/datasets/pull/3129
3,129
Support Audio feature for TAR archives in sequential access
closed
7
2021-10-21T08:56:51
2021-11-17T17:42:08
2021-11-17T17:42:07
albertvillanova
[]
Add Audio feature support for TAR archived files in sequential access. Fix #3128.
true
1,032,201,870
https://api.github.com/repos/huggingface/datasets/issues/3128
https://github.com/huggingface/datasets/issues/3128
3,128
Support Audio feature for TAR archives in sequential access
closed
0
2021-10-21T08:23:01
2021-11-17T17:42:07
2021-11-17T17:42:07
albertvillanova
[ "enhancement" ]
Currently, Audio feature accesses each audio file by their file path. However, streamed TAR archive files do not allow random access to their archived files. Therefore, we should enhance the Audio feature to support TAR archived files in sequential access.
false
1,032,100,613
https://api.github.com/repos/huggingface/datasets/issues/3127
https://github.com/huggingface/datasets/issues/3127
3,127
datasets-cli: convertion of a tfds dataset to a huggingface one.
open
1
2021-10-21T06:14:27
2021-10-27T11:36:05
null
vitalyshalumov
[]
### Discussed in https://github.com/huggingface/datasets/discussions/3079 <div type='discussions-op-text'> <sup>Originally posted by **vitalyshalumov** October 14, 2021</sup> I'm trying to convert a tfds dataset to a huggingface one. I've tried: 1. datasets-cli convert --tfds_path ~/tensorflow_datasets/mnist/3.0.1/ --datasets_directory ~/.cache/huggingface/datasets/mnist/3.0.1/ 2. datasets-cli convert --tfds_path ~/tensorflow_datasets/mnist/3.0.1/ --datasets_directory ~/.cache/huggingface/datasets/ and other permutations. The script appears to be running and finishing without an error but when looking in the huggingface/datasets/ folder nothing is created. </div>
false
1,032,093,055
https://api.github.com/repos/huggingface/datasets/issues/3126
https://github.com/huggingface/datasets/issues/3126
3,126
"arabic_billion_words" dataset does not create the full dataset
closed
1
2021-10-21T06:02:38
2021-10-22T13:28:40
2021-10-22T13:28:40
vitalyshalumov
[ "bug" ]
## Describe the bug When running: raw_dataset = load_dataset('arabic_billion_words','Alittihad') the correct dataset file is pulled from the url. But, the generated dataset includes just a small portion of the data included in the file. This is true for all other portions of the "arabic_billion_words" dataset ('Almasryalyoum',.....) ## Steps to reproduce the bug ```python # Sample code to reproduce the bug raw_dataset = load_dataset('arabic_billion_words','Alittihad') #The screen message Downloading and preparing dataset arabic_billion_words/Alittihad (download: 332.13 MiB, generated: 20.62 MiB, post-processed: Unknown size, total: 352.74 MiB) ## Expected results over 100K sentences ## Actual results only 11K sentences ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.14.0 - Platform: Linux-5.8.0-63-generic-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 4.0.1
false
1,032,046,666
https://api.github.com/repos/huggingface/datasets/issues/3125
https://github.com/huggingface/datasets/pull/3125
3,125
Add SLR83 to OpenSLR
closed
0
2021-10-21T04:26:00
2021-10-22T20:10:05
2021-10-22T08:30:22
tyrius02
[]
The PR resolves #3119, adding SLR83 (UK and Ireland dialects) to the previously created OpenSLR dataset.
true
1,031,976,286
https://api.github.com/repos/huggingface/datasets/issues/3124
https://github.com/huggingface/datasets/pull/3124
3,124
More efficient nested features encoding
closed
2
2021-10-21T01:55:31
2021-11-02T15:07:13
2021-11-02T11:04:04
eladsegal
[]
Nested encoding of features wastes a lot of time on operations which are effectively doing nothing when lists are used. For example, if in the input we have a list of integers, `encoded_nested_example` will iterate over it and apply `encoded_nested_example` on every element even though it just return the int as is. A similar issue is handled at an earlier stage when casting pytorch/tensorflow/pandas objects to python lists/numpy arrays: https://github.com/huggingface/datasets/blob/c98c23c4260edadab00f997d1a5d66b7f2e93ce9/src/datasets/features/features.py#L149-L156 https://github.com/huggingface/datasets/blob/c98c23c4260edadab00f997d1a5d66b7f2e93ce9/src/datasets/features/features.py#L212-L228 In this pull request I suggest to use the same approach in `encoded_nested_example`. In my setup there was a major speedup with this change: loading the data was at least x4 faster.
true
1,031,793,207
https://api.github.com/repos/huggingface/datasets/issues/3123
https://github.com/huggingface/datasets/issues/3123
3,123
Segmentation fault when loading datasets from file
closed
2
2021-10-20T20:16:11
2021-11-02T14:57:07
2021-11-02T14:57:07
TevenLeScao
[ "bug" ]
## Describe the bug Custom dataset loading sometimes segfaults and kills the process if chunks contain a variety of features/ ## Steps to reproduce the bug Download an example file: ``` wget https://gist.githubusercontent.com/TevenLeScao/11e2184394b3fa47d693de2550942c6b/raw/4232704d08fbfcaf93e5b51def9e5051507651ad/tiny_kelm.jsonl ``` Then in Python: ``` import datasets tiny_kelm = datasets.load_dataset("json", data_files="tiny_kelm.jsonl", chunksize=100000) ``` ## Expected results a `tiny_kelm` functional dataset ## Actual results ☠️ `Segmentation fault (core dumped)` ☠️ ## Environment info - `datasets` version: 1.14.0 - Platform: Linux-5.11.0-38-generic-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 5.0.0
false
1,031,787,509
https://api.github.com/repos/huggingface/datasets/issues/3122
https://github.com/huggingface/datasets/issues/3122
3,122
OSError with a custom dataset loading script
closed
8
2021-10-20T20:08:39
2021-11-23T09:55:38
2021-11-23T09:55:38
suzanab
[ "bug" ]
## Describe the bug I am getting an OS error when trying to load the newly uploaded dataset classla/janes_tag. What puzzles me is that I have already uploaded a very similar dataset - classla/reldi_hr - with no issues. The loading scripts for the two datasets are almost identical and they have the same directory structure, yet I am only getting an error with janes_tag. ## Steps to reproduce the bug ```python dataset = datasets.load_dataset('classla/janes_tag', split='validation') ``` ## Expected results Dataset correctly loaded. ## Actual results Traceback (most recent call last): File "C:/mypath/test.py", line 91, in <module> load_and_print('janes_tag') File "C:/mypath/test.py", line 32, in load_and_print dataset = datasets.load_dataset('classla/{}'.format(ds_name), split='validation') File "C:\mypath\venv\lib\site-packages\datasets\load.py", line 1632, in load_dataset use_auth_token=use_auth_token, File "C:\mypath\venv\lib\site-packages\datasets\builder.py", line 608, in download_and_prepare dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "C:\mypath\venv\lib\site-packages\datasets\builder.py", line 704, in _download_and_prepare ) from None OSError: Cannot find data file. Original error: [Errno 2] No such file or directory: 'C:\\mypath\\.cache\\huggingface\\datasets\\downloads\\2c9996e44bdc5af9c89bffb9e6d7a3e42fdb2f56bacab45de13b20f3032ea7ca\\data\\train_all.conllup' ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.14.0 - Platform: Windows-10-10.0.19041-SP0 - Python version: 3.7.5 - PyArrow version: 3.0.0
false
1,031,673,115
https://api.github.com/repos/huggingface/datasets/issues/3121
https://github.com/huggingface/datasets/pull/3121
3,121
Use huggingface_hub.HfApi to list datasets/metrics
closed
0
2021-10-20T17:48:29
2021-11-05T11:45:08
2021-11-05T09:48:36
mariosasko
[]
Delete `datasets.inspect.HfApi` and use `huggingface_hub.HfApi` instead. WIP until https://github.com/huggingface/huggingface_hub/pull/429 is merged, then wait for the new release of `huggingface_hub`, update the `huggingface_hub` version in `setup.py` and merge this PR. cc: @lhoestq
true
1,031,574,511
https://api.github.com/repos/huggingface/datasets/issues/3120
https://github.com/huggingface/datasets/pull/3120
3,120
Correctly update metadata to preserve features when concatenating datasets with axis=1
closed
0
2021-10-20T15:54:58
2021-10-22T08:28:51
2021-10-21T14:50:21
mariosasko
[]
This PR correctly updates metadata to preserve higher-level feature types (e.g. `ClassLabel`) in `datasets.concatenate_datasets` when `axis=1`. Previously, we would delete the feature metadata in `datasets.concatenate_datasets` if `axis=1` and restore the feature types from the arrow table schema in `Dataset.__init__`. However, this approach only works for simple feature types (e.g. `Value`). Fixes #3111
true
1,031,328,044
https://api.github.com/repos/huggingface/datasets/issues/3119
https://github.com/huggingface/datasets/issues/3119
3,119
Add OpenSLR 83 - Crowdsourced high-quality UK and Ireland English Dialect speech
closed
1
2021-10-20T12:05:07
2021-10-22T19:00:52
2021-10-22T08:30:22
tyrius02
[ "dataset request" ]
## Adding a Dataset - **Name:** *openslr** - **Description:** *Data set which contains male and female recordings of English from various dialects of the UK and Ireland.* - **Paper:** *https://www.openslr.org/resources/83/about.html* - **Data:** *Eleven separate data files can be found via https://www.openslr.org/resources/83/* - **Motivation:** *Increase english ASR data with UK and Irish dialects* Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md). The *openslr* dataset already exists, this will add additional subset, *SLR83*.
false
1,031,309,549
https://api.github.com/repos/huggingface/datasets/issues/3118
https://github.com/huggingface/datasets/pull/3118
3,118
Fix CI error at each release commit
closed
0
2021-10-20T11:44:38
2021-10-20T13:02:36
2021-10-20T13:02:36
albertvillanova
[]
Fix test_load_dataset_canonical at release commit. Fix #3117.
true
1,031,308,083
https://api.github.com/repos/huggingface/datasets/issues/3117
https://github.com/huggingface/datasets/issues/3117
3,117
CI error at each release commit
closed
0
2021-10-20T11:42:53
2021-10-20T13:02:35
2021-10-20T13:02:35
albertvillanova
[ "bug" ]
After 1.12.0, there is a recurrent CI error at each release commit: https://app.circleci.com/pipelines/github/huggingface/datasets/8289/workflows/665d954d-e409-4602-8202-e678594d2946/jobs/51110 ``` ____________________ LoadTest.test_load_dataset_canonical _____________________ [gw0] win32 -- Python 3.6.8 C:\tools\miniconda3\python.exe self = <tests.test_load.LoadTest testMethod=test_load_dataset_canonical> def test_load_dataset_canonical(self): scripts_version = os.getenv("HF_SCRIPTS_VERSION", SCRIPTS_VERSION) with self.assertRaises(FileNotFoundError) as context: datasets.load_dataset("_dummy") self.assertIn( f"https://raw.githubusercontent.com/huggingface/datasets/{scripts_version}/datasets/_dummy/_dummy.py", > str(context.exception), ) E AssertionError: 'https://raw.githubusercontent.com/huggingface/datasets/1.14.0/datasets/_dummy/_dummy.py' not found in "Couldn't find a dataset script at C:\\Users\\circleci\\datasets\\_dummy\\_dummy.py or any data file in the same directory. Couldn't find '_dummy' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/master/datasets/_dummy/_dummy.py" tests\test_load.py:358: AssertionError ```
false
1,031,270,611
https://api.github.com/repos/huggingface/datasets/issues/3116
https://github.com/huggingface/datasets/pull/3116
3,116
Update doc links to point to new docs
closed
0
2021-10-20T11:00:47
2021-10-22T08:29:28
2021-10-22T08:26:45
mariosasko
[ "documentation" ]
This PR: * updates the README links and the ADD_NEW_DATASET template to point to the new docs (the new docs don't have a section with the list of all the possible features, so I added that info to the `Features` docstring, which is then referenced in the ADD_NEW_DATASET template) * fixes some broken links in the `.rst` files (fixed with the `make linkcheck` tool)
true
1,030,737,524
https://api.github.com/repos/huggingface/datasets/issues/3115
https://github.com/huggingface/datasets/pull/3115
3,115
Fill in dataset card for NCBI disease dataset
closed
0
2021-10-19T20:57:05
2021-10-22T08:25:07
2021-10-22T08:25:07
edugp
[]
null
true
1,030,693,130
https://api.github.com/repos/huggingface/datasets/issues/3114
https://github.com/huggingface/datasets/issues/3114
3,114
load_from_disk in DatasetsDict/Dataset not working with PyArrowHDFS wrapper implementing fsspec.spec.AbstractFileSystem
closed
2
2021-10-19T20:01:45
2022-02-14T14:00:28
2022-02-14T14:00:28
francisco-perez-sorrosal
[ "bug" ]
## Describe the bug Passing a PyArrowHDFS implementation of fsspec.spec.AbstractFileSystem (in the `fs` param required by `load_from_disk` methods in `DatasetDict` (in datasets_dict.py) and `Dataset` (in arrow_dataset.py) results in an error when calling the download method in the `fs` parameter. ## Steps to reproduce the bug The documentation for the `fs` parameter states: ``` fs (:class:`~filesystems.S3FileSystem` or ``fsspec.spec.AbstractFileSystem``, optional, default ``None``): Instance of the remote filesystem used to download the files from. ``` `PyArrowHDFS` from [fsspec](https://filesystem-spec.readthedocs.io/en/latest/_modules/fsspec/implementations/hdfs.html) implements `fsspec.spec.AbstractFileSystem`. However, when using it as shown below, I get an error. ```python from fsspec.implementations.hdfs import PyArrowHDFS ... transformed_corpus_path = "/user/my_user/clickbait/transformed_ds/" fs = PyArrowHDFS(host, port, user, kerb_ticket=kerb_ticket) dss = DatasetDict.load_from_disk(transformed_corpus_path, fs, True) ``` ## Expected results Previous to load from disk, I have managed to successfully store in HDFS the data and meta-information of a DatasetDict by doing: ```python transformed_corpus_path = "/user/my_user/clickbait/transformed_ds/" fs = PyArrowHDFS(host, port, user, kerb_ticket=kerb_ticket) my_datasets.save_to_disk(transformed_corpus_path, fs=fs) ``` As I have 3 datasets in the DatasetDict named `my_datasets`, the previous Python code creates the following contents in HDFS: ```sh $ hadoop fs -ls "/user/my_user/clickbait/transformed_ds/" Found 4 items -rw------- 3 my_user users 43 2021-10-19 03:08 /user/my_user/clickbait/transformed_ds/dataset_dict.json drwx------ - my_user users 0 2021-10-19 03:08 /user/my_user/clickbait/transformed_ds/test drwx------ - my_user users 0 2021-10-19 03:08 /user/my_user/clickbait/transformed_ds/train drwx------ - my_user users 0 2021-10-19 03:08 /user/my_user/clickbait/transformed_ds/validation ``` I would expect to recover on `dss` the Arrow-backed datasets I previously saved in HDFS calling the `save_to_disk` method on the `DatasetDict` object when invoking `DatasetDict.load_from_disk(...)` as described above. ## Actual results However, when trying to recover the saved datasets, I get this error: ``` ... File "/home/fperez/dev/neuromancer/neuromancer/corpus.py", line 186, in load_transformed_corpus_from_disk dss = DatasetDict.load_from_disk(transformed_corpus_path, fs, True) File "/home/fperez/anaconda3/envs/neuromancer/lib/python3.9/site-packages/datasets/dataset_dict.py", line 748, in load_from_disk dataset_dict[k] = Dataset.load_from_disk(dataset_dict_split_path, fs, keep_in_memory=keep_in_memory) File "/home/fperez/anaconda3/envs/neuromancer/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 1048, in load_from_disk fs.download(src_dataset_path, dataset_path.as_posix(), recursive=True) File "pyarrow/_hdfsio.pyx", line 438, in pyarrow._hdfsio.HadoopFileSystem.download TypeError: download() got an unexpected keyword argument 'recursive' ``` Examining the [signature of the download method in pyarrow 5.0.0](https://github.com/apache/arrow/blob/54d2bd89c99df72fa091b025452f85dd5d88e3cf/python/pyarrow/_hdfsio.pyx#L438) we can see that there's no download parameter: ```python def download(self, path, stream, buffer_size=None): with self.open(path, 'rb') as f: f.download(stream, buffer_size=buffer_size) ``` ## Environment info - `datasets` version: 1.13.3 - Platform: Linux-3.10.0-1160.15.2.el7.x86_64-x86_64-with-glibc2.33 - Python version: 3.9.7 - PyArrow version: 5.0.0
false
1,030,667,547
https://api.github.com/repos/huggingface/datasets/issues/3113
https://github.com/huggingface/datasets/issues/3113
3,113
Loading Data from HDF files
open
9
2021-10-19T19:26:46
2025-06-19T05:41:23
null
FeryET
[ "enhancement", "good second issue" ]
**Is your feature request related to a problem? Please describe.** More often than not I come along big HDF datasets, and currently there is no straight forward way to feed them to a dataset. **Describe the solution you'd like** I would love to see a `from_h5` method that gets an interface implemented by the user on how items are extracted from dataset (in case of multiple datasets containing elements like arrays and metadata and etc). **Describe alternatives you've considered** Currently I manually load hdf files using `h5py` and implement PyTorch dataset interface. For small h5 files I load them into a pandas dataframe and use `from_pandas` function in the `datasets` package to load them, but for big datasets this is not feasible. **Additional context** HDF files are widespread throughout different domains and are one of the go to's for many researchers/scientists/engineers who work with numerical data. Given `datasets`' usecases have outgrown NLP use cases, it will make a lot of sense focusing on things like supporting HDF files.
false
1,030,613,083
https://api.github.com/repos/huggingface/datasets/issues/3112
https://github.com/huggingface/datasets/issues/3112
3,112
OverflowError: There was an overflow in the <class 'pyarrow.lib.ListArray'>. Try to reduce writer_batch_size to have batches smaller than 2GB
open
4
2021-10-19T18:21:41
2021-10-19T18:52:29
null
BenoitDalFerro
[ "bug" ]
## Describe the bug Despite having batches way under 2Gb when running `datasets.map()`, after processing correctly the data of the first batch without fuss and irrespective of writer_batch_size (say 2,4,8,16,32,64 and 128 in my case), it returns the following error : > OverflowError: There was an overflow in the <class 'pyarrow.lib.ListArray'>. Try to reduce writer_batch_size to have batches smaller than 2GB Note that I always run `batch_size=writer_batch_size` : ## Steps to reproduce the bug ```python datasets.map(lambda example : {"column_name" : function(arguments)}, batched=False, remove_columns = datasets.column_names, batch_size=batch_size, writer_batch_size=batch_size, disable_nullable=True, num_proc=None, desc="blablabla") ``` ## Introspecting CUDA memory during bug Placed within `function(arguments)` the following statement to introspect memory usage, merely a little over 1/4 of 2Gb `print(torch.cuda.memory_summary(device=device, abbreviated=False))` > |===========================================================================| | PyTorch CUDA memory summary, device ID 0 | |---------------------------------------------------------------------------| | CUDA OOMs: 0 | cudaMalloc retries: 0 | |===========================================================================| | Metric | Cur Usage | Peak Usage | Tot Alloc | Tot Freed | |---------------------------------------------------------------------------| | Allocated memory | 541418 KB | 545725 KB | 555695 KB | 14276 KB | | from large pool | 540672 KB | 544431 KB | 544431 KB | 3759 KB | | from small pool | 746 KB | 1714 KB | 11264 KB | 10517 KB | |---------------------------------------------------------------------------| | Active memory | 541418 KB | 545725 KB | 555695 KB | 14276 KB | | from large pool | 540672 KB | 544431 KB | 544431 KB | 3759 KB | | from small pool | 746 KB | 1714 KB | 11264 KB | 10517 KB | |---------------------------------------------------------------------------| | GPU reserved memory | 598016 KB | 598016 KB | 598016 KB | 0 B | | from large pool | 595968 KB | 595968 KB | 595968 KB | 0 B | | from small pool | 2048 KB | 2048 KB | 2048 KB | 0 B | |---------------------------------------------------------------------------| | Non-releasable memory | 36117 KB | 52292 KB | 274275 KB | 238158 KB | | from large pool | 34816 KB | 51537 KB | 261713 KB | 226897 KB | | from small pool | 1301 KB | 2045 KB | 12562 KB | 11261 KB | |---------------------------------------------------------------------------| | Allocations | 198 | 224 | 478 | 280 | | from large pool | 74 | 75 | 75 | 1 | | from small pool | 124 | 150 | 403 | 279 | |---------------------------------------------------------------------------| | Active allocs | 198 | 224 | 478 | 280 | | from large pool | 74 | 75 | 75 | 1 | | from small pool | 124 | 150 | 403 | 279 | |---------------------------------------------------------------------------| | GPU reserved segments | 21 | 21 | 21 | 0 | | from large pool | 20 | 20 | 20 | 0 | | from small pool | 1 | 1 | 1 | 0 | |---------------------------------------------------------------------------| | Non-releasable allocs | 18 | 23 | 166 | 148 | | from large pool | 17 | 18 | 19 | 2 | | from small pool | 1 | 6 | 147 | 146 | |===========================================================================| ## Expected results Efficiently process the datasets and write it down to disk. ## Actual results -------------------------------------------------------------------------- OverflowError Traceback (most recent call last) ~\anaconda3\envs\xxx\lib\site-packages\datasets\arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, disable_tqdm, desc, cache_only) 2390 else: -> 2391 writer.write(example) 2392 else: ~\anaconda3\envs\xxx\lib\site-packages\datasets\arrow_writer.py in write(self, example, key, writer_batch_size) 367 --> 368 self.write_examples_on_file() 369 ~\anaconda3\envs\xxx\lib\site-packages\datasets\arrow_writer.py in write_examples_on_file(self) 316 if not isinstance(pa_array[0], pa.lib.FloatScalar): --> 317 raise OverflowError( 318 "There was an overflow in the {}. Try to reduce writer_batch_size to have batches smaller than 2GB".format( OverflowError: There was an overflow in the <class 'pyarrow.lib.ListArray'>. Try to reduce writer_batch_size to have batches smaller than 2GB During handling of the above exception, another exception occurred: OverflowError Traceback (most recent call last) ~\AppData\Local\Temp/ipykernel_16268/2456940807.py in <module> 3 #tracker = OfflineEmissionsTracker(country_iso_code="FRA", project_name='xxx'+time_stamp,output_dir='./codecarbon') 4 #tracker.start() ----> 5 process_datasets(source_datasets_paths, dataset_dir, LM_tokenizer, LMhead_model, datasets_selection=['wikipedia'], from_scratch=True, 6 clean_sentences=False, negative_sampling=False, translate=False, tokenize=False, generate_embeddings=True, concatenate_embeddings=False, 7 max_sample=10000, padding='do_not_pad', truncation=True, cpu_batch_size=1000, gpu_batch_size=2, cpu_writer_batch_size=1000, gpu_writer_batch_size=2, disable_nullable=True, num_proc=None) # ~\xxx\xxx.py in process_datasets(source_datasets_paths, dataset_dir, LM_tokenizer, LMhead_model, datasets_selection, from_scratch, clean_sentences, translate, negative_sampling, tokenize, generate_embeddings, concatenate_embeddings, max_sample, padding, truncation, cpu_batch_size, gpu_batch_size, cpu_writer_batch_size, gpu_writer_batch_size, disable_nullable, num_proc) 481 for column in tqdm(dataset.column_names, desc=f'Processing column', leave=False): 482 if "xxx_" in column: --> 483 dataset = dataset.map(lambda example : 484 {"embeddings_"+str(column).replace("translated_",""):function(input_ids=example[column], 485 token_type_ids=example[column.replace("input_ids","token_type_ids")], ~\anaconda3\envs\xxx\lib\site-packages\datasets\arrow_dataset.py in map(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc) 2034 2035 if num_proc is None or num_proc == 1: -> 2036 return self._map_single( 2037 function=function, 2038 with_indices=with_indices, ~\anaconda3\envs\xxx\lib\site-packages\datasets\arrow_dataset.py in wrapper(*args, **kwargs) 501 self: "Dataset" = kwargs.pop("self") 502 # apply actual function --> 503 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 504 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 505 for dataset in datasets: ~\anaconda3\envs\xxx\lib\site-packages\datasets\arrow_dataset.py in wrapper(*args, **kwargs) 468 } 469 # apply actual function --> 470 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 471 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 472 # re-apply format to the output ~\anaconda3\envs\xxx\lib\site-packages\datasets\fingerprint.py in wrapper(*args, **kwargs) 404 # Call actual function 405 --> 406 out = func(self, *args, **kwargs) 407 408 # Update fingerprint of in-place transforms + update in-place history of transforms ~\anaconda3\envs\xxx\lib\site-packages\datasets\arrow_dataset.py in _map_single(self, function, with_indices, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, new_fingerprint, rank, offset, disable_tqdm, desc, cache_only) 2425 if update_data: 2426 if writer is not None: -> 2427 writer.finalize() 2428 if tmp_file is not None: 2429 tmp_file.close() ~\anaconda3\envs\xxx\lib\site-packages\datasets\arrow_writer.py in finalize(self, close_stream) 440 # Re-intializing to empty list for next batch 441 self.hkey_record = [] --> 442 self.write_examples_on_file() 443 if self.pa_writer is None: 444 if self._schema is not None: ~\anaconda3\envs\xxx\lib\site-packages\datasets\arrow_writer.py in write_examples_on_file(self) 315 # This check fails with FloatArrays with nans, which is not what we want, so account for that: 316 if not isinstance(pa_array[0], pa.lib.FloatScalar): --> 317 raise OverflowError( 318 "There was an overflow in the {}. Try to reduce writer_batch_size to have batches smaller than 2GB".format( 319 type(pa_array) OverflowError: There was an overflow in the <class 'pyarrow.lib.ListArray'>. Try to reduce writer_batch_size to have batches smaller than 2GB ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.13.3 - Platform: Windows-10-10.0.19042-SP0 - Python version: 3.8.11 - PyArrow version: 3.0.0 ##Next steps Testing on Linux. @albertvillanova
false
1,030,598,983
https://api.github.com/repos/huggingface/datasets/issues/3111
https://github.com/huggingface/datasets/issues/3111
3,111
concatenate_datasets removes ClassLabel typing.
closed
1
2021-10-19T18:05:31
2021-10-21T14:50:21
2021-10-21T14:50:21
Dref360
[ "bug" ]
## Describe the bug When concatenating two datasets, we lose typing of ClassLabel columns. I can work on this if this is a legitimate bug, ## Steps to reproduce the bug ```python import datasets from datasets import Dataset, ClassLabel, Value, concatenate_datasets DS_LEN = 100 my_dataset = Dataset.from_dict( { "sentence": [f"{chr(i % 10)}" for i in range(DS_LEN)], "label": [i % 2 for i in range(DS_LEN)] } ) my_predictions = Dataset.from_dict( { "pred": [(i + 1) % 2 for i in range(DS_LEN)] } ) my_dataset = my_dataset.cast(datasets.Features({"sentence": Value("string"), "label": ClassLabel(2, names=["POS", "NEG"])})) print("Original") print(my_dataset) print(my_dataset.features) concat_ds = concatenate_datasets([my_dataset, my_predictions], axis=1) print("Concatenated") print(concat_ds) print(concat_ds.features) ``` ## Expected results The features of `concat_ds` should contain ClassLabel. ## Actual results On master, I get: ``` {'sentence': Value(dtype='string', id=None), 'label': Value(dtype='int64', id=None), 'pred': Value(dtype='int64', id=None)} ``` ## Environment info - `datasets` version: 1.14.1.dev0 - Platform: macOS-10.15.7-x86_64-i386-64bit - Python version: 3.8.11 - PyArrow version: 4.0.1
false
1,030,558,484
https://api.github.com/repos/huggingface/datasets/issues/3110
https://github.com/huggingface/datasets/pull/3110
3,110
Stream TAR-based dataset using iter_archive
closed
2
2021-10-19T17:16:24
2021-11-05T17:48:49
2021-11-05T17:48:48
lhoestq
[]
I converted all the dataset based on TAR archive to use iter_archive instead, so that they can be streamable. It means that around 80 datasets become streamable :)
true
1,030,543,284
https://api.github.com/repos/huggingface/datasets/issues/3109
https://github.com/huggingface/datasets/pull/3109
3,109
Update BibTeX entry
closed
0
2021-10-19T16:59:31
2021-10-19T17:13:28
2021-10-19T17:13:27
albertvillanova
[]
Update BibTeX entry.
true
1,030,405,618
https://api.github.com/repos/huggingface/datasets/issues/3108
https://github.com/huggingface/datasets/pull/3108
3,108
Add Google BLEU (aka GLEU) metric
closed
0
2021-10-19T14:48:38
2021-10-25T14:07:04
2021-10-25T14:07:04
slowwavesleep
[]
This PR adds the NLTK implementation of Google BLEU metric. This is also a part of an effort to resolve an unfortunate naming collision between GLEU for machine translation and GLEU for grammatical error correction. I used [this page](https://huggingface.co/docs/datasets/add_metric.html) for reference. Please, point me to the right direction if I missed anything.
true
1,030,357,527
https://api.github.com/repos/huggingface/datasets/issues/3107
https://github.com/huggingface/datasets/pull/3107
3,107
Add paper BibTeX citation
closed
0
2021-10-19T14:08:11
2021-10-19T14:26:22
2021-10-19T14:26:21
albertvillanova
[]
Add paper BibTeX citation to README file.
true
1,030,112,473
https://api.github.com/repos/huggingface/datasets/issues/3106
https://github.com/huggingface/datasets/pull/3106
3,106
Fix URLs in blog_authorship_corpus dataset
closed
0
2021-10-19T10:06:05
2021-10-19T12:50:40
2021-10-19T12:50:39
albertvillanova
[]
After contacting the authors of the paper "Effects of Age and Gender on Blogging", they confirmed: - the old URLs are no longer valid - there are alternative host URLs Fix #3091.
true
1,029,098,843
https://api.github.com/repos/huggingface/datasets/issues/3105
https://github.com/huggingface/datasets/issues/3105
3,105
download_mode=`force_redownload` does not work on removed datasets
open
0
2021-10-18T13:12:38
2021-10-22T09:36:10
null
severo
[ "bug", "dataset-viewer" ]
## Describe the bug If a cached dataset is removed from the library, I don't see how to delete it programmatically. I thought that using `force_redownload` would try to refresh the cache, then raise an exception, but it reuses the cache instead. ## Steps to reproduce the bug _requires to already have `wit` in the cache_: see https://github.com/huggingface/datasets/pull/2981 ```python import datasets as ds dataset = ds.load_dataset("wit", split="train", download_mode='force_redownload') ``` ## Expected results It should raise an exception, since the dataset does not exist anymore. ## Actual results It uses the cached result ``` Using the latest cached version of the module from /home/slesage/.cache/huggingface/modules/datasets_modules/datasets/wit/107afbffd48e058b19101bddc47fbee25fa68eb6d50a733e262875f1285a5171 (last modified on Wed Sep 29 08:21:10 2021) since it couldn't be found locally at wit, or remotely on the Hugging Face Hub. ``` ## Environment info - `datasets` version: 1.13.4.dev0 - Platform: Linux-5.11.0-1019-aws-x86_64-with-glibc2.31 - Python version: 3.9.6 - PyArrow version: 4.0.1
false
1,029,080,412
https://api.github.com/repos/huggingface/datasets/issues/3104
https://github.com/huggingface/datasets/issues/3104
3,104
Missing Zenodo 1.13.3 release
closed
1
2021-10-18T12:57:18
2021-10-22T13:22:25
2021-10-22T13:22:24
albertvillanova
[ "bug" ]
After `datasets` 1.13.3 release, this does not appear in Zenodo releases: https://zenodo.org/record/5570305 TODO: - [x] Contact Zenodo support - [x] Check it is fixed
false
1,029,069,310
https://api.github.com/repos/huggingface/datasets/issues/3103
https://github.com/huggingface/datasets/pull/3103
3,103
Fix project description in PyPI
closed
0
2021-10-18T12:47:29
2021-10-18T12:59:57
2021-10-18T12:59:56
albertvillanova
[]
Fix project description appearing in PyPI, so that it contains the content of the README.md file (like transformers). Currently, `datasets` project description appearing in PyPI shows the release instructions addressed to core maintainers: https://pypi.org/project/datasets/1.13.3/ Fix #3102.
true
1,029,067,062
https://api.github.com/repos/huggingface/datasets/issues/3102
https://github.com/huggingface/datasets/issues/3102
3,102
Unsuitable project description in PyPI
closed
0
2021-10-18T12:45:00
2021-10-18T12:59:56
2021-10-18T12:59:56
albertvillanova
[]
Currently, `datasets` project description appearing in PyPI shows the release instructions addressed to core maintainers: https://pypi.org/project/datasets/1.13.3/
false
1,028,966,968
https://api.github.com/repos/huggingface/datasets/issues/3101
https://github.com/huggingface/datasets/pull/3101
3,101
Update SUPERB to use Audio features
closed
1
2021-10-18T11:05:18
2021-10-18T12:33:54
2021-10-18T12:06:46
anton-l
[]
This is the same dataset refresh as the other Audio ones: https://github.com/huggingface/datasets/pull/3081 cc @patrickvonplaten
true
1,028,738,180
https://api.github.com/repos/huggingface/datasets/issues/3100
https://github.com/huggingface/datasets/pull/3100
3,100
Replace FSTimeoutError with parent TimeoutError
closed
0
2021-10-18T07:37:09
2021-10-18T07:51:55
2021-10-18T07:51:54
albertvillanova
[]
PR #3050 introduced a dependency on `fsspec.FSTiemoutError`. Note that this error only exists from `fsspec` version `2021.06.0` (June 2021). To fix #3097, there are 2 alternatives: - Either pinning `fsspec` to versions newer or equal to `2021.06.0` - Or replacing `fsspec.FSTimeoutError` wth its parent `asyncio.TimeoutError`, which exists from Python 3.8.0 (Sep 2018). This PR implements the second approach. Fix #3097.
true
1,028,338,078
https://api.github.com/repos/huggingface/datasets/issues/3099
https://github.com/huggingface/datasets/issues/3099
3,099
AttributeError: module 'huggingface_hub.hf_api' has no attribute 'DatasetInfo'
closed
6
2021-10-17T14:17:47
2021-11-09T16:42:29
2021-11-09T16:42:28
JTWang2000
[ "bug" ]
## Describe the bug When using `pip install datasets` or use `conda install -c huggingface -c conda-forge datasets` cannot install datasets ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("sst", "default") ``` ## Actual results --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) <ipython-input-5-fbe7981e6e21> in <module> 1 import torch 2 import transformers ----> 3 from datasets import load_dataset 4 5 dataset = load_dataset("sst", "default") ~/miniforge3/envs/actor/lib/python3.8/site-packages/datasets/__init__.py in <module> 35 from .arrow_reader import ArrowReader, ReadInstruction 36 from .arrow_writer import ArrowWriter ---> 37 from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder 38 from .combine import interleave_datasets 39 from .dataset_dict import DatasetDict, IterableDatasetDict ~/miniforge3/envs/actor/lib/python3.8/site-packages/datasets/builder.py in <module> 42 ) 43 from .arrow_writer import ArrowWriter, BeamWriter ---> 44 from .data_files import DataFilesDict, _sanitize_patterns 45 from .dataset_dict import DatasetDict, IterableDatasetDict 46 from .fingerprint import Hasher ~/miniforge3/envs/actor/lib/python3.8/site-packages/datasets/data_files.py in <module> 118 119 def _exec_patterns_in_dataset_repository( --> 120 dataset_info: huggingface_hub.hf_api.DatasetInfo, 121 patterns: List[str], 122 allowed_extensions: Optional[list] = None, AttributeError: module 'huggingface_hub.hf_api' has no attribute 'DatasetInfo' ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.13.3 - Platform: macOS-11.3.1-arm64-arm-64bit - Python version: 3.8.10 - PyArrow version: 5.0.0
false