id
int64
599M
3.48B
number
int64
1
7.8k
title
stringlengths
1
290
state
stringclasses
2 values
comments
listlengths
0
30
created_at
timestamp[s]date
2020-04-14 10:18:02
2025-10-05 06:37:50
updated_at
timestamp[s]date
2020-04-27 16:04:17
2025-10-05 10:32:43
closed_at
timestamp[s]date
2020-04-14 12:01:40
2025-10-01 13:56:03
body
stringlengths
0
228k
user
stringlengths
3
26
html_url
stringlengths
46
51
pull_request
dict
is_pull_request
bool
2 classes
1,070,426,462
3,374
NonMatchingChecksumError for the CLUE:cluewsc2020, chid, c3 and tnews
closed
[ "Seems like the issue still exists,:\r\n`Downloading and preparing dataset clue/chid (download: 127.15 MiB, generated: 259.71 MiB, post-processed: Unknown size, total: 386.86 MiB) to /mnt/cache/tanhaochen/.cache/huggingface/datasets/clue/chid/1.0.0/e55b490cb7809dcd8db31b9a87119f2e2ec87cdc060da8a9ac070b070ca3e379......
2021-12-03T10:10:54
2021-12-08T14:14:41
2021-12-08T14:14:41
Hi, it seems like there are updates in cluewsc2020, chid, c3 and tnews, since i could not load them due to the checksum error.
Namco0816
https://github.com/huggingface/datasets/issues/3374
null
false
1,070,406,391
3,373
Support streaming zipped CSV dataset repo by passing only repo name
closed
[]
2021-12-03T09:48:24
2021-12-16T18:03:31
2021-12-16T18:03:31
Given a community 🤗 dataset repository containing only a zipped CSV file (only raw data, no loading script), I would like to load it in streaming mode without passing `data_files`: ``` ds_name = "bigscience-catalogue-data/vietnamese_poetry_from_fsoft_ai_lab" ds = load_dataset(ds_name, split="train", streaming=True, use_auth_token=True) item = next(iter(ds)) ``` Currently, it gives a `FileNotFoundError` because there is no glob (no "\*" after "zip://": "zip://*") in the passed URL: ``` 'zip://::https://huggingface.co/datasets/bigscience-catalogue-data/vietnamese_poetry_from_fsoft_ai_lab/resolve/e5d45f1bd9a8a798cc14f0a45ebc1ce91907c792/poems_dataset.zip' ```
albertvillanova
https://github.com/huggingface/datasets/issues/3373
null
false
1,069,948,178
3,372
[SEO improvement] Add Dataset Metadata to make datasets indexable
closed
[]
2021-12-02T20:21:07
2022-03-18T09:36:48
2022-03-18T09:36:48
Some people who host datasets on github seem to include a table of metadata at the end of their README.md to make the dataset indexable by [Google Dataset Search](https://datasetsearch.research.google.com/) (See [here](https://github.com/google-research/google-research/tree/master/goemotions#dataset-metadata) and [here](https://github.com/cvdfoundation/google-landmark#dataset-metadata)). This could be a useful addition to canonical datasets; perhaps even community datasets. I'll include a screenshot (as opposed to markdown) as an example so as not to have a github issue indexed as a dataset: > ![image](https://user-images.githubusercontent.com/3664563/144496173-953428cf-633a-4571-b75b-f099c6b2ed65.png) **_PS: It might very well be the case that this is already covered by some other markdown magic I'm not aware of._**
cakiki
https://github.com/huggingface/datasets/issues/3372
null
false
1,069,821,335
3,371
New: Americas NLI dataset
closed
[]
2021-12-02T17:44:59
2021-12-08T13:58:12
2021-12-08T13:58:11
This PR adds the [Americas NLI](https://arxiv.org/abs/2104.08726) dataset, extension of XNLI to 10 low-resource indigenous languages spoken in the Americas: Ashaninka, Aymara, Bribri, Guarani, Nahuatl, Otomi, Quechua, Raramuri, Shipibo-Konibo, and Wixarika. One odd thing (not sure) is that I had to set `datasets-cli dummy_data ./datasets/americas_nli/ --auto_generate --n_lines 7500` `n_lines` very large to successfully generate the dummy files for all the subsets. Happy to get some guidance here. Otherwise, I hope everything is in order :) e: missed a step, onto fixing the tests e2: there you go -- hope it's ok to have added more languages with their ISO codes to `languages.json`, need those tests to pass :laughing:
fdschmidt93
https://github.com/huggingface/datasets/pull/3371
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3371", "html_url": "https://github.com/huggingface/datasets/pull/3371", "diff_url": "https://github.com/huggingface/datasets/pull/3371.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3371.patch", "merged_at": "2021-12-08T13:58:11" }
true
1,069,735,423
3,370
Document a training loop for streaming dataset
closed
[]
2021-12-02T16:17:00
2021-12-03T13:34:35
2021-12-03T13:34:34
I added some docs about streaming dataset. In particular I added two subsections: - one on how to use `map` for preprocessing - one on how to use a streaming dataset in a pytorch training loop cc @patrickvonplaten @stevhliu if you have some comments cc @Rocketknight1 later we can add the one for TF and I might need your help ^^'
lhoestq
https://github.com/huggingface/datasets/pull/3370
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3370", "html_url": "https://github.com/huggingface/datasets/pull/3370", "diff_url": "https://github.com/huggingface/datasets/pull/3370.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3370.patch", "merged_at": "2021-12-03T13:34:34" }
true
1,069,587,674
3,369
[Audio] Allow resampling for audio datasets in streaming mode
closed
[ "This requires implementing `cast_column` for iterable datasets, it could be a very nice addition !\r\n\r\n<s>It can also be useful to be able to disable the audio/image decoding for the dataset viewer (see PR https://github.com/huggingface/datasets/pull/3430) cc @severo </s>\r\nEDIT: actually following https://git...
2021-12-02T14:04:57
2021-12-16T15:55:19
2021-12-16T15:55:19
Many audio datasets like Common Voice always need to be resampled. This can very easily be done in non-streaming mode as follows: ```python from datasets import load_dataset ds = load_dataset("common_voice", "ab", split="test") ds = ds.cast_column("audio", Audio(sampling_rate=16_000)) ``` However in streaming mode it fails currently: ```python from datasets import load_dataset ds = load_dataset("common_voice", "ab", split="test", streaming=True) ds = ds.cast_column("audio", Audio(sampling_rate=16_000)) ``` with the following error: ``` AttributeError: 'IterableDataset' object has no attribute 'cast_column' ``` It would be great if we could add such a feature (I'm not 100% sure though how complex this would be)
patrickvonplaten
https://github.com/huggingface/datasets/issues/3369
null
false
1,069,403,624
3,368
Fix dict source_datasets tagset validator
closed
[]
2021-12-02T10:52:20
2021-12-02T15:48:38
2021-12-02T15:48:37
Currently, the `source_datasets` tag validation does not support passing a dict with configuration keys. This PR: - Extends `tagset_validator` to support regex tags - Uses `tagset_validator` to validate dict `source_datasets`
albertvillanova
https://github.com/huggingface/datasets/pull/3368
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3368", "html_url": "https://github.com/huggingface/datasets/pull/3368", "diff_url": "https://github.com/huggingface/datasets/pull/3368.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3368.patch", "merged_at": "2021-12-02T15:48:37" }
true
1,069,241,274
3,367
Fix typo in other-structured-to-text task tag
closed
[]
2021-12-02T08:02:27
2021-12-02T16:07:14
2021-12-02T16:07:13
Fix typo in task tag: - `other-stuctured-to-text` (before) - `other-structured-to-text` (now)
albertvillanova
https://github.com/huggingface/datasets/pull/3367
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3367", "html_url": "https://github.com/huggingface/datasets/pull/3367", "diff_url": "https://github.com/huggingface/datasets/pull/3367.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3367.patch", "merged_at": "2021-12-02T16:07:13" }
true
1,069,214,022
3,366
Add multimodal datasets
open
[]
2021-12-02T07:24:04
2023-02-28T16:29:22
null
Epic issue to track the addition of multimodal datasets: - [ ] #2526 - [x] #1842 - [ ] #1810 Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md). @VictorSanh feel free to add and sort by priority any interesting dataset. I have added the multimodal dataset requests which were already present as issues.
albertvillanova
https://github.com/huggingface/datasets/issues/3366
null
false
1,069,195,887
3,365
Add task tags for multimodal datasets
closed
[ "The Hub pulls these tags from [here](https://github.com/huggingface/hub-docs/blob/main/js/src/lib/interfaces/Types.ts) (allows multimodal tasks) now, so I'm closing this issue." ]
2021-12-02T06:58:20
2023-07-25T18:21:33
2023-07-25T18:21:32
## **Is your feature request related to a problem? Please describe.** Currently, task tags are either exclusively related to text or speech processing: - https://github.com/huggingface/datasets/blob/master/src/datasets/utils/resources/tasks.json ## **Describe the solution you'd like** We should also add tasks related to: - multimodality - image - video CC: @VictorSanh @lewtun @lhoestq @merveenoyan @SBrandeis
albertvillanova
https://github.com/huggingface/datasets/issues/3365
null
false
1,068,851,196
3,364
Use the Audio feature in the AutomaticSpeechRecognition template
closed
[ "Cool !\r\n\r\nI noticed that you removed the `audio_file_path_column` field of the template, note that you also have to update all the dataset_infos.json file that still contain this outdated field. For example in the common_voice you can find this:\r\n```\r\n\"task_templates\": [{\"task\": \"automatic-speech-reco...
2021-12-01T20:42:26
2022-03-24T14:34:09
2022-03-24T14:34:08
This updates the ASR template and all supported datasets to use the `Audio` feature
anton-l
https://github.com/huggingface/datasets/pull/3364
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3364", "html_url": "https://github.com/huggingface/datasets/pull/3364", "diff_url": "https://github.com/huggingface/datasets/pull/3364.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3364.patch", "merged_at": null }
true
1,068,824,340
3,363
Update URL of Jeopardy! dataset
closed
[ "Closing this PR in favor of #3266.", "I think you should also close this branch" ]
2021-12-01T20:08:10
2022-10-06T13:45:49
2021-12-03T12:35:01
Updates the URL of the Jeopardy! dataset. Fix #3361
mariosasko
https://github.com/huggingface/datasets/pull/3363
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3363", "html_url": "https://github.com/huggingface/datasets/pull/3363", "diff_url": "https://github.com/huggingface/datasets/pull/3363.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3363.patch", "merged_at": null }
true
1,068,809,768
3,362
Adapt image datasets
closed
[ "This PR can be merged after #3163 is merged (this PR is pretty big because I was working on the forked branch).\r\n\r\n@lhoestq @albertvillanova Could you please take a look at the changes in `src/datasets/utils/streaming_download_manager.py`? These changes were required to support streaming of the `cats_vs_dogs` ...
2021-12-01T19:52:01
2021-12-09T18:37:42
2021-12-09T18:37:41
This PR: * adapts the ImageClassification template to use the new Image feature * adapts the following datasets to use the new Image feature: * beans (+ fixes streaming) * cast_vs_dogs (+ fixes streaming) * cifar10 * cifar100 * fashion_mnist * mnist * head_qa cc @nateraw
mariosasko
https://github.com/huggingface/datasets/pull/3362
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3362", "html_url": "https://github.com/huggingface/datasets/pull/3362", "diff_url": "https://github.com/huggingface/datasets/pull/3362.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3362.patch", "merged_at": "2021-12-09T18:37:41" }
true
1,068,736,268
3,361
Jeopardy _URL access denied
closed
[ "Just a side note: duplicate #3264" ]
2021-12-01T18:21:33
2021-12-11T12:50:23
2021-12-06T11:16:31
## Describe the bug http://skeeto.s3.amazonaws.com/share/JEOPARDY_QUESTIONS1.json.gz returns Access Denied now. However, https://drive.google.com/file/d/0BwT5wj_P7BKXb2hfM3d2RHU1ckE/view?usp=sharing from the original Reddit post https://www.reddit.com/r/datasets/comments/1uyd0t/200000_jeopardy_questions_in_a_json_file/ may work. ## Steps to reproduce the bug ```shell > python Python 3.7.12 (default, Sep 5 2021, 08:34:29) [Clang 11.0.3 (clang-1103.0.32.62)] on darwin Type "help", "copyright", "credits" or "license" for more information. ``` ```python >>> from datasets import load_dataset >>> load_dataset("jeopardy") ``` ## Expected results The download completes. ## Actual results ```shell Downloading: 4.18kB [00:00, 1.60MB/s] Downloading: 2.03kB [00:00, 1.04MB/s] Using custom data configuration default Downloading and preparing dataset jeopardy/default (download: 12.13 MiB, generated: 34.46 MiB, post-processed: Unknown size, total: 46.59 MiB) to /Users/mike/.cache/huggingface/datasets/jeopardy/default/0.1.0/25ee3e4a73755e637b8810f6493fd36e4523dea3ca8a540529d0a6e24c7f9810... Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/load.py", line 1632, in load_dataset use_auth_token=use_auth_token, File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/builder.py", line 608, in download_and_prepare dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/builder.py", line 675, in _download_and_prepare split_generators = self._split_generators(dl_manager, **split_generators_kwargs) File "/Users/mike/.cache/huggingface/modules/datasets_modules/datasets/jeopardy/25ee3e4a73755e637b8810f6493fd36e4523dea3ca8a540529d0a6e24c7f9810/jeopardy.py", line 72, in _split_generators filepath = dl_manager.download_and_extract(_DATA_URL) File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 284, in download_and_extract return self.extract(self.download(url_or_urls)) File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 197, in download download_func, url_or_urls, map_tuple=True, num_proc=download_config.num_proc, disable_tqdm=False File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 197, in map_nested return function(data_struct) File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 217, in _download return cached_path(url_or_filename, download_config=download_config) File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 305, in cached_path use_auth_token=download_config.use_auth_token, File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 594, in get_from_cache raise ConnectionError("Couldn't reach {}".format(url)) ConnectionError: Couldn't reach http://skeeto.s3.amazonaws.com/share/JEOPARDY_QUESTIONS1.json.gz ``` --- ```shell > curl http://skeeto.s3.amazonaws.com/share/JEOPARDY_QUESTIONS1.json.gz ``` ```xml <?xml version="1.0" encoding="UTF-8"?> <Error><Code>AccessDenied</Code><Message>Access Denied</Message><RequestId>70Y9R36XNPEQXMGV</RequestId><HostId>G6F5AK4qo7JdaEdKGMtS0P6gdLPeFOdEfSEfvTOZEfk9km0/jAfp08QLfKSTFFj1oWIKoAoBehM=</HostId></Error> ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.14.0 - Platform: macOS Catalina 10.15.7 - Python version: 3.7.12 - PyArrow version: 6.0.1
tianjianjiang
https://github.com/huggingface/datasets/issues/3361
null
false
1,068,724,697
3,360
Add The Pile USPTO subset
closed
[]
2021-12-01T18:08:05
2021-12-03T11:45:29
2021-12-03T11:45:28
Add: - USPTO subset of The Pile: "uspto" config Close bigscience-workshop/data_tooling#297. CC: @StellaAthena
albertvillanova
https://github.com/huggingface/datasets/pull/3360
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3360", "html_url": "https://github.com/huggingface/datasets/pull/3360", "diff_url": "https://github.com/huggingface/datasets/pull/3360.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3360.patch", "merged_at": "2021-12-03T11:45:27" }
true
1,068,638,213
3,359
Add The Pile Free Law subset
closed
[ "@albertvillanova Is there a specific reason you’re adding the Pile under “the” instead of under “pile”? That does not appear to be consistent with other datasets.", "Hi @StellaAthena,\r\n\r\nI asked myself the same question, but at the end I decided to be consistent with previously added Pile subsets:\r\n- #2817...
2021-12-01T16:46:04
2021-12-06T10:12:17
2021-12-01T17:30:44
Add: - Free Law subset of The Pile: "free_law" config Close bigscience-workshop/data_tooling#75. CC: @StellaAthena
albertvillanova
https://github.com/huggingface/datasets/pull/3359
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3359", "html_url": "https://github.com/huggingface/datasets/pull/3359", "diff_url": "https://github.com/huggingface/datasets/pull/3359.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3359.patch", "merged_at": "2021-12-01T17:30:43" }
true
1,068,623,216
3,358
add new field, and get errors
closed
[ "Hi, \r\n\r\ncould you please post this question on our [Forum](https://discuss.huggingface.co/) as we keep issues for bugs and feature requests? ", "> Hi,\r\n> \r\n> could you please post this question on our [Forum](https://discuss.huggingface.co/) as we keep issues for bugs and feature requests?\r\n\r\nok." ]
2021-12-01T16:35:38
2021-12-02T02:26:22
2021-12-02T02:26:22
after adding new field **tokenized_examples["example_id"]**, and get errors below, I think it is due to changing data to tensor, and **tokenized_examples["example_id"]** is string list **all fields** ``` ***************** train_dataset 1: Dataset({ features: ['attention_mask', 'end_positions', 'example_id', 'input_ids', 'start_positions', 'token_type_ids'], num_rows: 87714 }) ``` **Errors** ``` Traceback (most recent call last): File "/usr/local/lib/python3.7/site-packages/transformers/tokenization_utils_base.py", line 705, in convert_to_tensors tensor = as_tensor(value) ValueError: too many dimensions 'str' ```
PatricYan
https://github.com/huggingface/datasets/issues/3358
null
false
1,068,607,382
3,357
Update languages in aeslc dataset card
closed
[]
2021-12-01T16:20:46
2022-09-23T13:16:49
2022-09-23T13:16:49
After having worked a bit with the dataset. As far as I know, it is solely in English (en-US). There are only a few mails in Spanish, French or German (less than a dozen I would estimate).
apergo-ai
https://github.com/huggingface/datasets/pull/3357
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3357", "html_url": "https://github.com/huggingface/datasets/pull/3357", "diff_url": "https://github.com/huggingface/datasets/pull/3357.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3357.patch", "merged_at": "2022-09-23T13:16:48" }
true
1,068,503,932
3,356
to_tf_dataset() refactor
closed
[ "Also, please don't merge yet - I need to make sure all the code samples and notebooks have a collate_fn specified, since we're removing the ability for this method to work without one!", "Hi @lhoestq @mariosasko, the other PRs this was depending on in Transformers and huggingface/notebooks are now merged, so thi...
2021-12-01T14:54:30
2021-12-09T10:26:53
2021-12-09T10:26:53
This is the promised cleanup to `to_tf_dataset()` now that the course is out of the way! The main changes are: - A collator is always required (there was way too much hackiness making things like labels work without it) - Lots of cleanup and a lot of code moved to `_get_output_signature` - Should now handle it gracefully when the data collator adds unexpected columns
Rocketknight1
https://github.com/huggingface/datasets/pull/3356
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3356", "html_url": "https://github.com/huggingface/datasets/pull/3356", "diff_url": "https://github.com/huggingface/datasets/pull/3356.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3356.patch", "merged_at": "2021-12-09T10:26:53" }
true
1,068,468,573
3,355
Extend support for streaming datasets that use pd.read_excel
closed
[ "TODO in the future: https://github.com/huggingface/datasets/pull/3355#discussion_r761138011\r\n- If we finally find a use case where the `pd.read_excel()` can work in streaming mode (using fsspec), that is, without using the `.read()`, I propose to try this first, catch the ValueError and then try with `.read`, bu...
2021-12-01T14:22:43
2021-12-17T07:24:19
2021-12-17T07:24:18
This PR fixes error: ``` ValueError: Cannot seek streaming HTTP file ``` CC: @severo
albertvillanova
https://github.com/huggingface/datasets/pull/3355
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3355", "html_url": "https://github.com/huggingface/datasets/pull/3355", "diff_url": "https://github.com/huggingface/datasets/pull/3355.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3355.patch", "merged_at": "2021-12-17T07:24:18" }
true
1,068,307,271
3,354
Remove duplicate name from dataset cards
closed
[]
2021-12-01T11:45:40
2021-12-01T13:14:30
2021-12-01T13:14:29
Remove duplicate name from dataset card for: - ajgt_twitter_ar - emotone_ar
albertvillanova
https://github.com/huggingface/datasets/pull/3354
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3354", "html_url": "https://github.com/huggingface/datasets/pull/3354", "diff_url": "https://github.com/huggingface/datasets/pull/3354.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3354.patch", "merged_at": "2021-12-01T13:14:29" }
true
1,068,173,783
3,353
add one field "example_id", but I can't see it in the "comput_loss" function
closed
[ "Hi ! Your function looks fine, I used to map `squad` locally and it indeed added the `example_id` field correctly.\r\n\r\nHowever I think that in the `compute_loss` method only a subset of the fields are available: the model inputs. Since `example_id` is not a model input (it's not passed as a parameter to the mod...
2021-12-01T09:35:09
2021-12-01T16:02:39
2021-12-01T16:02:39
Hi, I add one field **example_id**, but I can't see it in the **comput_loss** function, how can I do this? below is the information of inputs ``` *********************** inputs: {'attention_mask': tensor([[1, 1, 1, ..., 0, 0, 0], [1, 1, 1, ..., 0, 0, 0], [1, 1, 1, ..., 0, 0, 0], ..., [1, 1, 1, ..., 0, 0, 0], [1, 1, 1, ..., 0, 0, 0], [1, 1, 1, ..., 0, 0, 0]], device='cuda:0'), 'end_positions': tensor([ 25, 97, 93, 44, 25, 112, 109, 134], device='cuda:0'), 'input_ids': tensor([[ 101, 2054, 2390, ..., 0, 0, 0], [ 101, 2054, 2515, ..., 0, 0, 0], [ 101, 2054, 2106, ..., 0, 0, 0], ..., [ 101, 2339, 2001, ..., 0, 0, 0], [ 101, 2054, 2515, ..., 0, 0, 0], [ 101, 2054, 2003, ..., 0, 0, 0]], device='cuda:0'), 'start_positions': tensor([ 20, 90, 89, 41, 25, 96, 106, 132], device='cuda:0'), 'token_type_ids': tensor([[0, 0, 0, ..., 0, 0, 0], [0, 0, 0, ..., 0, 0, 0], [0, 0, 0, ..., 0, 0, 0], ..., [0, 0, 0, ..., 0, 0, 0], [0, 0, 0, ..., 0, 0, 0], [0, 0, 0, ..., 0, 0, 0]], device='cuda:0')} ``` ``` # This function preprocesses a question answering dataset, tokenizing the question and context text # and finding the right offsets for the answer spans in the tokenized context (to use as labels). # Adapted from https://github.com/huggingface/transformers/blob/master/examples/pytorch/question-answering/run_qa.py def prepare_train_dataset_qa(examples, tokenizer, max_seq_length=None): questions = [q.lstrip() for q in examples["question"]] max_seq_length = tokenizer.model_max_length # tokenize both questions and the corresponding context # if the context length is longer than max_length, we split it to several # chunks of max_length tokenized_examples = tokenizer( questions, examples["context"], truncation="only_second", max_length=max_seq_length, stride=min(max_seq_length // 2, 128), return_overflowing_tokens=True, return_offsets_mapping=True, padding="max_length" ) # Since one example might give us several features if it has a long context, # we need a map from a feature to its corresponding example. sample_mapping = tokenized_examples.pop("overflow_to_sample_mapping") # The offset mappings will give us a map from token to character position # in the original context. This will help us compute the start_positions # and end_positions to get the final answer string. offset_mapping = tokenized_examples.pop("offset_mapping") tokenized_examples["start_positions"] = [] tokenized_examples["end_positions"] = [] tokenized_examples["example_id"] = [] for i, offsets in enumerate(offset_mapping): input_ids = tokenized_examples["input_ids"][i] # We will label features not containing the answer the index of the CLS token. cls_index = input_ids.index(tokenizer.cls_token_id) sequence_ids = tokenized_examples.sequence_ids(i) # from the feature idx to sample idx sample_index = sample_mapping[i] # get the answer for a feature answers = examples["answers"][sample_index] tokenized_examples["example_id"].append(examples["id"][sample_index]) if len(answers["answer_start"]) == 0: tokenized_examples["start_positions"].append(cls_index) tokenized_examples["end_positions"].append(cls_index) else: # Start/end character index of the answer in the text. start_char = answers["answer_start"][0] end_char = start_char + len(answers["text"][0]) # Start token index of the current span in the text. token_start_index = 0 while sequence_ids[token_start_index] != 1: token_start_index += 1 # End token index of the current span in the text. token_end_index = len(input_ids) - 1 while sequence_ids[token_end_index] != 1: token_end_index -= 1 # Detect if the answer is out of the span (in which case this feature is labeled with the CLS index). if not (offsets[token_start_index][0] <= start_char and offsets[token_end_index][1] >= end_char): tokenized_examples["start_positions"].append(cls_index) tokenized_examples["end_positions"].append(cls_index) else: # Otherwise move the token_start_index and token_end_index to the two ends of the answer. # Note: we could go after the last offset if the answer is the last word (edge case). while token_start_index < len(offsets) and \ offsets[token_start_index][0] <= start_char: token_start_index += 1 tokenized_examples["start_positions"].append( token_start_index - 1) while offsets[token_end_index][1] >= end_char: token_end_index -= 1 tokenized_examples["end_positions"].append(token_end_index + 1) return tokenized_examples ``` _Originally posted by @yanllearnn in https://github.com/huggingface/datasets/issues/3333#issuecomment-983457161_
PatricYan
https://github.com/huggingface/datasets/issues/3353
null
false
1,068,102,994
3,352
Make LABR dataset streamable
closed
[]
2021-12-01T08:22:27
2021-12-01T10:49:02
2021-12-01T10:49:01
Fix LABR dataset to make it streamable. Related to: #3350.
albertvillanova
https://github.com/huggingface/datasets/pull/3352
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3352", "html_url": "https://github.com/huggingface/datasets/pull/3352", "diff_url": "https://github.com/huggingface/datasets/pull/3352.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3352.patch", "merged_at": "2021-12-01T10:49:01" }
true
1,068,094,873
3,351
Add VCTK dataset
closed
[ "Hello @patrickvonplaten, I hope it's okay to ping you with a (dumb) question!\r\n\r\nI've been trying to get `dl_manager.download_and_extract(_DL_URL)` to work with no avail. I verified that this is a problem on two different machines (lab server, GCP), so I doubt it's an issue with network connectivity. Here is t...
2021-12-01T08:13:17
2022-02-28T09:22:03
2021-12-28T15:05:08
Fixes #1837.
jaketae
https://github.com/huggingface/datasets/pull/3351
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3351", "html_url": "https://github.com/huggingface/datasets/pull/3351", "diff_url": "https://github.com/huggingface/datasets/pull/3351.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3351.patch", "merged_at": "2021-12-28T15:05:07" }
true
1,068,078,160
3,350
Avoid content-encoding issue while streaming datasets
closed
[]
2021-12-01T07:56:48
2021-12-01T08:15:01
2021-12-01T08:15:00
This PR will fix streaming of datasets served with gzip content-encoding: ``` ClientPayloadError: 400, message='Can not decode content-encoding: gzip' ``` Fix #2918. CC: @severo
albertvillanova
https://github.com/huggingface/datasets/pull/3350
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3350", "html_url": "https://github.com/huggingface/datasets/pull/3350", "diff_url": "https://github.com/huggingface/datasets/pull/3350.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3350.patch", "merged_at": "2021-12-01T08:15:00" }
true
1,067,853,601
3,349
raise exception instead of using assertions.
closed
[ "@mariosasko - Thanks for the review & suggestions. Updated as per the suggestions. ", "@mariosasko - Hello, Are there any additional changes required from my end??. Wondering if this PR can be merged or still pending on additional steps.", "@mariosasko - The approved changes in the PR now has conflicts with th...
2021-12-01T01:37:51
2021-12-20T16:07:27
2021-12-20T16:07:27
fix for the remaining files https://github.com/huggingface/datasets/issues/3171
manisnesan
https://github.com/huggingface/datasets/pull/3349
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3349", "html_url": "https://github.com/huggingface/datasets/pull/3349", "diff_url": "https://github.com/huggingface/datasets/pull/3349.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3349.patch", "merged_at": "2021-12-20T16:07:27" }
true
1,067,831,113
3,348
BLEURT: Match key names to correspond with filename
closed
[ "Thanks for the suggestion! I think the current checked-in `CHECKPOINT_URLS` is already not working. I believe anyone who tried using the new ckpts (`BLEURT-20-X`) can't unless this fix is in. The zip file from bleurt side unzips to directory name matching the filename (capitalized for new ones). For example withou...
2021-12-01T01:01:18
2021-12-07T16:06:57
2021-12-07T16:06:57
In order to properly locate downloaded ckpt files key name needs to match filename. Correcting change introduced in #3235
jaehlee
https://github.com/huggingface/datasets/pull/3348
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3348", "html_url": "https://github.com/huggingface/datasets/pull/3348", "diff_url": "https://github.com/huggingface/datasets/pull/3348.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3348.patch", "merged_at": "2021-12-07T16:06:57" }
true
1,067,738,902
3,347
iter_archive for zip files
closed
[ "And also don't always try streaming with Google Drive - it can have issues because of how Google Drive works (with quotas, restrictions, etc.) and it can indeed cause `BlockSizeError`.\r\n\r\nFeel free to host your test data elsewhere, such as in a dataset repository on https://huggingface.co (see [here](https://h...
2021-11-30T22:34:17
2021-12-04T00:22:22
2021-12-04T00:22:11
* In this PR, I added the option to iterate through zipfiles for `download_manager.py` only. * Next PR will be the same applied to `streaming_download_manager.py`. * Related issue #3272. ## Comments : * There is no `.isreg()` equivalent in zipfile library to check if file is Regular so I used `.is_dir()` instead to skip directories. * For now I got `streaming_download_manager.py` working for local zip files, but not for urls. I get the following error when I test it on an archive in google drive, so still working on it. `BlockSizeError: Got more bytes so far (>2112) than requested (22)` ## Tasks : - [x] download_manager.py - [ ] streaming_download_manager.py
Mehdi2402
https://github.com/huggingface/datasets/pull/3347
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3347", "html_url": "https://github.com/huggingface/datasets/pull/3347", "diff_url": "https://github.com/huggingface/datasets/pull/3347.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3347.patch", "merged_at": null }
true
1,067,632,365
3,346
Failed to convert `string` with pyarrow for QED since 1.15.0
closed
[ "Scratch that, probably the old and incompatible usage of dataset builder from promptsource.", "Actually, re-opening this issue cause the error persists\r\n\r\n```python\r\n>>> load_dataset(\"qed\")\r\nDownloading and preparing dataset qed/qed (download: 13.43 MiB, generated: 9.70 MiB, post-processed: Unknown siz...
2021-11-30T20:11:42
2021-12-14T14:39:05
2021-12-14T14:39:05
## Describe the bug Loading QED was fine until 1.15.0. related: bigscience-workshop/promptsource#659, bigscience-workshop/promptsource#670 Not sure where the root cause is, but here are some candidates: - #3158 - #3120 - #3196 - #2891 ## Steps to reproduce the bug ```python load_dataset("qed") ``` ## Expected results Loading completed. ## Actual results ```shell ArrowInvalid: Could not convert in with type str: tried to convert to boolean Traceback: File "/Users/s0s0cr3/Library/Python/3.9/lib/python/site-packages/streamlit/script_runner.py", line 354, in _run_script exec(code, module.__dict__) File "/Users/s0s0cr3/Documents/GitHub/promptsource/promptsource/app.py", line 260, in <module> dataset = get_dataset(dataset_key, str(conf_option.name) if conf_option else None) File "/Users/s0s0cr3/Library/Python/3.9/lib/python/site-packages/streamlit/caching.py", line 543, in wrapped_func return get_or_create_cached_value() File "/Users/s0s0cr3/Library/Python/3.9/lib/python/site-packages/streamlit/caching.py", line 527, in get_or_create_cached_value return_value = func(*args, **kwargs) File "/Users/s0s0cr3/Documents/GitHub/promptsource/promptsource/utils.py", line 49, in get_dataset builder_instance.download_and_prepare() File "/Users/s0s0cr3/Library/Python/3.9/lib/python/site-packages/datasets/builder.py", line 607, in download_and_prepare self._download_and_prepare( File "/Users/s0s0cr3/Library/Python/3.9/lib/python/site-packages/datasets/builder.py", line 697, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/Users/s0s0cr3/Library/Python/3.9/lib/python/site-packages/datasets/builder.py", line 1106, in _prepare_split num_examples, num_bytes = writer.finalize() File "/Users/s0s0cr3/Library/Python/3.9/lib/python/site-packages/datasets/arrow_writer.py", line 456, in finalize self.write_examples_on_file() File "/Users/s0s0cr3/Library/Python/3.9/lib/python/site-packages/datasets/arrow_writer.py", line 325, in write_examples_on_file pa_array = pa.array(typed_sequence) File "pyarrow/array.pxi", line 222, in pyarrow.lib.array File "pyarrow/array.pxi", line 110, in pyarrow.lib._handle_arrow_array_protocol File "/Users/s0s0cr3/Library/Python/3.9/lib/python/site-packages/datasets/arrow_writer.py", line 121, in __arrow_array__ out = pa.array(cast_to_python_objects(self.data, only_1d_for_numpy=True), type=type) File "pyarrow/array.pxi", line 305, in pyarrow.lib.array File "pyarrow/array.pxi", line 39, in pyarrow.lib._sequence_to_array File "pyarrow/error.pxi", line 122, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 84, in pyarrow.lib.check_status ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.15.0, 1.16.1 - Platform: macOS 1.15.7 or above - Python version: 3.7.12 and 3.9 - PyArrow version: 3.0.0, 5.0.0, 6.0.1
tianjianjiang
https://github.com/huggingface/datasets/issues/3346
null
false
1,067,622,951
3,345
Failed to download species_800 from Google Drive zip file
closed
[ "Hi,\r\n\r\nthe dataset is downloaded normally on my machine. Maybe the URL was down at the time of your download. Could you try again?", "> Hi,\r\n> \r\n> the dataset is downloaded normally on my machine. Maybe the URL was down at the time of your download. Could you try again?\r\n\r\nI have tried that many time...
2021-11-30T20:00:28
2021-12-01T17:53:15
2021-12-01T17:53:15
## Describe the bug One can manually download the zip file on Google Drive, but `load_dataset()` cannot. related: #3248 ## Steps to reproduce the bug ```shell > python Python 3.7.12 (default, Sep 5 2021, 08:34:29) [Clang 11.0.3 (clang-1103.0.32.62)] on darwin Type "help", "copyright", "credits" or "license" for more information. ``` ```python >>> from datasets import load_dataset >>> s800 = load_dataset("species_800") ``` ## Expected results species_800 downloaded. ## Actual results ```shell Downloading: 5.68kB [00:00, 1.22MB/s] Downloading: 2.70kB [00:00, 691kB/s] Downloading and preparing dataset species800/species_800 (download: 17.36 MiB, generated: 3.53 MiB, post-processed: Unknown size, total: 20.89 MiB) to /Users/mike/.cache/huggingface/datasets/species800/species_800/1.0.0/532167f0bb8fbc0d77d6d03c4fd642c8c55527b9c5f2b1da77f3d00b0e559976... 0%| | 0/1 [00:00<?, ?it/s]Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/load.py", line 1632, in load_dataset use_auth_token=use_auth_token, File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/builder.py", line 608, in download_and_prepare dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/builder.py", line 675, in _download_and_prepare split_generators = self._split_generators(dl_manager, **split_generators_kwargs) File "/Users/mike/.cache/huggingface/modules/datasets_modules/datasets/species_800/532167f0bb8fbc0d77d6d03c4fd642c8c55527b9c5f2b1da77f3d00b0e559976/species_800.py", line 104, in _split_generators downloaded_files = dl_manager.download_and_extract(urls_to_download) File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 284, in download_and_extract return self.extract(self.download(url_or_urls)) File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 197, in download download_func, url_or_urls, map_tuple=True, num_proc=download_config.num_proc, disable_tqdm=False File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 209, in map_nested for obj in utils.tqdm(iterable, disable=disable_tqdm) File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 209, in <listcomp> for obj in utils.tqdm(iterable, disable=disable_tqdm) File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/py_utils.py", line 143, in _single_map_nested return function(data_struct) File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/download_manager.py", line 217, in _download return cached_path(url_or_filename, download_config=download_config) File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 305, in cached_path use_auth_token=download_config.use_auth_token, File "/Users/mike/Library/Caches/pypoetry/virtualenvs/promptsource-hsdAcWsQ-py3.7/lib/python3.7/site-packages/datasets/utils/file_utils.py", line 594, in get_from_cache raise ConnectionError("Couldn't reach {}".format(url)) ConnectionError: Couldn't reach https://drive.google.com/u/0/uc?id=1OletxmPYNkz2ltOr9pyT0b0iBtUWxslh&export=download/ ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.14,0 1.15.0, 1.16.1 - Platform: macOS Catalina 10.15.7 - Python version: 3.7.12 - PyArrow version: 6.0.1
tianjianjiang
https://github.com/huggingface/datasets/issues/3345
null
false
1,067,567,603
3,344
Add ArrayXD docs
closed
[]
2021-11-30T18:53:31
2021-12-01T20:16:03
2021-12-01T19:35:32
Documents support for dynamic first dimension in `ArrayXD` from #2891, and explain the `ArrayXD` feature in general. Let me know if I'm missing anything @lhoestq :)
stevhliu
https://github.com/huggingface/datasets/pull/3344
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3344", "html_url": "https://github.com/huggingface/datasets/pull/3344", "diff_url": "https://github.com/huggingface/datasets/pull/3344.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3344.patch", "merged_at": "2021-12-01T19:35:32" }
true
1,067,505,507
3,343
Better error message when download fails
closed
[]
2021-11-30T17:38:50
2021-12-01T11:27:59
2021-12-01T11:27:58
From our discussions in https://github.com/huggingface/datasets/issues/3269 and https://github.com/huggingface/datasets/issues/3282 it would be nice to have better messages if a download fails. In particular the error now shows: - the error from the HEAD request if there's one - otherwise the response code of the HEAD request I also added an error to tell users to pass `use_auth_token` when the Hugging Face Hub returns 401 (Unauthorized). While paying around with this I also fixed a minor issue with the `force_download` parameter that was not always taken into account
lhoestq
https://github.com/huggingface/datasets/pull/3343
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3343", "html_url": "https://github.com/huggingface/datasets/pull/3343", "diff_url": "https://github.com/huggingface/datasets/pull/3343.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3343.patch", "merged_at": "2021-12-01T11:27:58" }
true
1,067,481,390
3,342
Fix ASSET dataset data URLs
closed
[ "> Hi @tianjianjiang, thanks for the fix.\r\n> The links should also be updated in the `dataset_infos.json` file.\r\n> The failing tests are due to the missing tag in the header of the `README.md` file:\r\n\r\nHi @albertvillanova, thank you for the info! My apologies for the messy PR.\r\n" ]
2021-11-30T17:13:30
2021-12-14T14:50:00
2021-12-14T14:50:00
Change the branch name "master" to "main" in the data URLs, since facebookresearch has changed that.
tianjianjiang
https://github.com/huggingface/datasets/pull/3342
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3342", "html_url": "https://github.com/huggingface/datasets/pull/3342", "diff_url": "https://github.com/huggingface/datasets/pull/3342.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3342.patch", "merged_at": "2021-12-14T14:50:00" }
true
1,067,449,569
3,341
Mirror the canonical datasets to the Hugging Face Hub
closed
[ "I created a GitHub project to keep track of what needs to be done:\r\nhttps://github.com/huggingface/datasets/projects/3\r\n\r\nI also store my code in a (private for now) repository at https://github.com/huggingface/mirror_canonical_datasets_on_hub", "I understand that the datasets are mirrored on the Hub now, ...
2021-11-30T16:42:05
2022-01-26T14:47:37
2022-01-26T14:47:37
- [ ] create a repo on https://hf.co/datasets for every canonical dataset - [ ] on every commit related to a dataset, update the hf.co repo See https://github.com/huggingface/moon-landing/pull/1562 @SBrandeis: I let you edit this description if needed to precise the intent.
severo
https://github.com/huggingface/datasets/issues/3341
null
false
1,067,292,636
3,340
Fix JSON ClassLabel casting for integers
closed
[]
2021-11-30T14:19:54
2021-12-01T11:27:30
2021-12-01T11:27:30
Loading a JSON dataset with ClassLabel feature types currently fails if the JSON data already has integers. Indeed currently it tries to convert the strings to integers without even checking if the data are not integers already. For example this currently fails: ```python from datasets import load_dataset, Features, ClassLabel path = "data.json" f = Features({"a": ClassLabel(names=["neg", "pos"])}) d = load_dataset("json", data_files=path, features=f) ``` data.json ```json {"a": 0} {"a": 1} ``` I fixed that by adding a line that checks the type of the JSON data before trying to convert them cc @albertvillanova let me know if it sounds good to you
lhoestq
https://github.com/huggingface/datasets/pull/3340
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3340", "html_url": "https://github.com/huggingface/datasets/pull/3340", "diff_url": "https://github.com/huggingface/datasets/pull/3340.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3340.patch", "merged_at": "2021-12-01T11:27:30" }
true
1,066,662,477
3,339
to_tf_dataset fails on TPU
open
[ "This might be related to https://github.com/tensorflow/tensorflow/issues/38762 , what do you think @Rocketknight1 ?\r\n> Dataset.from_generator is expected to not work with TPUs as it uses py_function underneath which is incompatible with Cloud TPU 2VM setup. If you would like to read from large datasets, maybe tr...
2021-11-30T00:50:52
2021-12-02T14:21:27
null
Using `to_tf_dataset` to create a dataset and then putting it in `model.fit` results in an internal error on TPUs. I've only tried on Colab and Kaggle TPUs, not GCP TPUs. ## Steps to reproduce the bug I made a colab to show the error. https://colab.research.google.com/drive/12x_PFKzGouFxqD4OuWfnycW_1TaT276z?usp=sharing ## Expected results dataset from `to_tf_dataset` works in `model.fit` Right below the first error in the colab I use `tf.data.Dataset.from_tensor_slices` and `model.fit` works just fine. This is the desired outcome. ## Actual results ``` InternalError: 5 root error(s) found. (0) INTERNAL: {{function_node __inference_train_function_30558}} failed to connect to all addresses Additional GRPC error information from remote target /job:localhost/replica:0/task:0/device:CPU:0: :{"created":"@1638231897.932218653","description":"Failed to pick subchannel","file":"third_party/grpc/src/core/ext/filters/client_channel/client_channel.cc","file_line":3151,"referenced_errors":[{"created":"@1638231897.932216754","description":"failed to connect to all addresses","file":"third_party/grpc/src/core/lib/transport/error_utils.cc","file_line":161,"grpc_status":14}]} [[{{node StatefulPartitionedCall}}]] [[MultiDeviceIteratorGetNextFromShard]] Executing non-communication op <MultiDeviceIteratorGetNextFromShard> originally returned UnavailableError, and was replaced by InternalError to avoid invoking TF network error handling logic. [[RemoteCall]] [[IteratorGetNextAsOptional]] [[tpu_compile_succeeded_assert/_14023832043698465348/_7/_439]] ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.16.1 - Platform: Linux-5.4.104+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.12 - PyArrow version: 3.0.0 - Tensorflow 2.7.0 - `transformers` 4.12.5
nbroad1881
https://github.com/huggingface/datasets/issues/3339
null
false
1,066,371,235
3,338
[WIP] Add doctests for tutorials
closed
[ "I manage to remove the mentions of ellipsis in the code by launching the command as follows:\r\n\r\n```\r\npython -m doctest -v docs/source/load_hub.rst -o=ELLIPSIS\r\n```\r\n\r\nThe way you put your ellipsis will only work on mac, I've adapted it for linux as well with the following:\r\n\r\n```diff\r\n >>> fro...
2021-11-29T18:40:46
2023-05-05T17:18:20
2023-05-05T17:18:15
Opening a PR as discussed with @LysandreJik for some help with doctest issues. The goal is to add doctests for each of the tutorials in the documentation to make sure the code samples work as shown. ### Issues A doctest has been added in the docstring of the `load_dataset_builder` function in `load.py` to handle variable outputs with the `ELLIPSIS` directive. When I run doctest on the `load_hub.rst` file, doctest should recognize the expected output from the docstring, and the corresponding code sample in `load_hub.rst` should pass. I am having the same issue with handling tracebacks in the `load_dataset` function. From the docstring: ``` >>> dataset_builder.cache_dir #doctest: +ELLIPSIS /Users/.../.cache/huggingface/datasets/imdb/plain_text/1.0.0/... ``` Test result: ``` Failed example: dataset_builder.cache_dir Expected: /Users/.../.cache/huggingface/datasets/imdb/plain_text/1.0.0/... Got: /Users/steven/.cache/huggingface/datasets/imdb/plain_text/1.0.0/2fdd8b9bcadd6e7055e742a706876ba43f19faee861df134affd7a3f60fc38a1 ``` I am able to get the doctest to pass by adding the doctest directives (`ELLIPSIS` and `NORMALIZE_WHITESPACE`) to the code samples in the `rst` file directly. But my understanding is that these directives should also work in the docstrings of the functions. I am running the test from the root of the directory: ``` python -m doctest -v docs/source/load_hub.rst ```
stevhliu
https://github.com/huggingface/datasets/pull/3338
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3338", "html_url": "https://github.com/huggingface/datasets/pull/3338", "diff_url": "https://github.com/huggingface/datasets/pull/3338.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3338.patch", "merged_at": null }
true
1,066,232,936
3,337
Typing of Dataset.__getitem__ could be improved.
closed
[ "Hi ! Thanks for the suggestion, I didn't know about this decorator.\r\n\r\nIf you are interesting in contributing, feel free to open a pull request to add the overload methods for each typing combination :) To assign you to this issue, you can comment `#self-assign` in this thread.\r\n\r\n`Dataset.__getitem__` is ...
2021-11-29T16:20:11
2021-12-14T10:28:54
2021-12-14T10:28:54
## Describe the bug The newly added typing for Dataset.__getitem__ is Union[Dict, List]. This makes tools like mypy a bit awkward to use as we need to check the type manually. We could use type overloading to make this easier. [Documentation](https://docs.python.org/3/library/typing.html#typing.overload) ## Steps to reproduce the bug Let's have a file `test.py` ```python from typing import List, Dict, Any from datasets import Dataset ds = Dataset.from_dict({ 'a': [1,2,3], 'b': ["1", "2", "3"] }) one_colum: List[str] = ds['a'] some_index: Dict[Any, Any] = ds[1] ``` ## Expected results Running `mypy test.py` should not give any error. ## Actual results ``` test.py:10: error: Incompatible types in assignment (expression has type "Union[Dict[Any, Any], List[Any]]", variable has type "List[str]") test.py:11: error: Incompatible types in assignment (expression has type "Union[Dict[Any, Any], List[Any]]", variable has type "Dict[Any, Any]") Found 2 errors in 1 file (checked 1 source file) ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.13.3 - Platform: macOS-10.16-x86_64-i386-64bit - Python version: 3.8.8 - PyArrow version: 6.0.1
Dref360
https://github.com/huggingface/datasets/issues/3337
null
false
1,066,208,436
3,336
Add support for multiple dynamic dimensions and to_pandas conversion for dynamic arrays
closed
[]
2021-11-29T15:58:59
2023-09-24T09:53:52
2023-05-16T18:24:46
Add support for multiple dynamic dimensions (e.g. `(None, None, 3)` for arbitrary sized images) and `to_pandas()` conversion for dynamic arrays. TODOs: * [ ] Cleaner code * [ ] Formatting issues (if NumPy doesn't allow broadcasting even though dtype is np.object) * [ ] Fix some issues with zero-dim tensors * [ ] Tests
mariosasko
https://github.com/huggingface/datasets/pull/3336
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3336", "html_url": "https://github.com/huggingface/datasets/pull/3336", "diff_url": "https://github.com/huggingface/datasets/pull/3336.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3336.patch", "merged_at": null }
true
1,066,064,126
3,335
add Speech commands dataset
closed
[ "@anton-l ping", "@lhoestq \r\nHi Quentin! Thank you for your feedback and suggestions! 🤗\r\n\r\nYes, that was actually what I wanted to do next - I mean the steaming stuff :)\r\nAlso, I need to make some changes to the readme (to account for the updated features set).\r\n\r\nHopefully, I will be done by tomorro...
2021-11-29T13:52:47
2021-12-10T10:37:21
2021-12-10T10:30:15
closes #3283
polinaeterna
https://github.com/huggingface/datasets/pull/3335
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3335", "html_url": "https://github.com/huggingface/datasets/pull/3335", "diff_url": "https://github.com/huggingface/datasets/pull/3335.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3335.patch", "merged_at": "2021-12-10T10:30:15" }
true
1,065,983,923
3,334
Integrate Polars library
closed
[ "If possible, a neat API could be something like `Dataset.to_polars()`, as well as `Dataset.set_format(\"polars\")`", "Note they use a \"custom\" implementation of Arrow: [Arrow2](https://github.com/jorgecarleitao/arrow2).", "Polars has grown rapidly in popularity over the last year - could you consider integra...
2021-11-29T12:31:54
2024-08-31T05:31:28
2024-08-31T05:31:27
Check potential integration of the Polars library: https://github.com/pola-rs/polars - Benchmark: https://h2oai.github.io/db-benchmark/ CC: @thomwolf @lewtun
albertvillanova
https://github.com/huggingface/datasets/issues/3334
null
false
1,065,346,919
3,333
load JSON files, get the errors
closed
[ "Hi ! The message you're getting is not an error. It simply says that your JSON dataset is being prepared to a location in `/root/.cache/huggingface/datasets`", "> \r\n\r\nbut I want to load local JSON file by command\r\n`python3 run.py --do_train --task qa --dataset squad-retrain-data/train-v2.0.json --output_di...
2021-11-28T14:29:58
2021-12-01T09:34:31
2021-12-01T03:57:48
Hi, does this bug be fixed? when I load JSON files, I get the same errors by the command `!python3 run.py --do_train --task qa --dataset squad-retrain-data/train-v2.0.json --output_dir ./re_trained_model/` change the dateset to load json by refering to https://huggingface.co/docs/datasets/loading.html `dataset = datasets.load_dataset('json', data_files=args.dataset)` Errors: `Downloading and preparing dataset json/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to /root/.cache/huggingface/datasets/json/default-c1e124ad488911b8/0.0.0/45636811569ec4a6630521c18235dfbbab83b7ab572e3393c5ba68ccabe98264... ` _Originally posted by @yanllearnn in https://github.com/huggingface/datasets/issues/730#issuecomment-981095050_
PatricYan
https://github.com/huggingface/datasets/issues/3333
null
false
1,065,345,853
3,332
Fix error message and add extension fallback
closed
[]
2021-11-28T14:25:29
2021-11-29T13:34:15
2021-11-29T13:34:14
Fix the error message raised if `infered_module_name` is `None` in `CommunityDatasetModuleFactoryWithoutScript.get_module` and make `infer_module_for_data_files` more robust. In the linked issue, `infer_module_for_data_files` returns `None` because `json` is the second most common extension due to the suffix ordering. Now, we go from the most common to the least common extension and try to map it or return `None`. Fix #3331
mariosasko
https://github.com/huggingface/datasets/pull/3332
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3332", "html_url": "https://github.com/huggingface/datasets/pull/3332", "diff_url": "https://github.com/huggingface/datasets/pull/3332.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3332.patch", "merged_at": "2021-11-29T13:34:14" }
true
1,065,275,896
3,331
AttributeError: 'CommunityDatasetModuleFactoryWithoutScript' object has no attribute 'path'
closed
[ "Hi,\r\n\r\nthe fix was merged and will be available in the next release of `datasets`.\r\nIn the meantime, you can use it by installing `datasets` directly from master as follows:\r\n```\r\npip install git+https://github.com/huggingface/datasets.git\r\n```" ]
2021-11-28T08:54:05
2021-11-29T13:49:44
2021-11-29T13:34:14
## Describe the bug I add a new question answering dataset to huggingface datasets manually. Here is the link: [luozhouyang/question-answering-datasets](https://huggingface.co/datasets/luozhouyang/question-answering-datasets) But when I load the dataset, an error raised: ```bash AttributeError: 'CommunityDatasetModuleFactoryWithoutScript' object has no attribute 'path' ``` ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("luozhouyang/question-answering-datasets", data_files=["dureader_robust.train.json"]) ``` ## Expected results Load dataset successfully without any error. ## Actual results ```bash Traceback (most recent call last): File "/mnt/home/zhouyang.lzy/github/naivenlp/naivenlp/tests/question_answering_tests/dataset_test.py", line 89, in test_load_dataset_with_hf data_files=["dureader_robust.train.json"], File "/mnt/home/zhouyang.lzy/.conda/envs/naivenlp/lib/python3.6/site-packages/datasets/load.py", line 1616, in load_dataset **config_kwargs, File "/mnt/home/zhouyang.lzy/.conda/envs/naivenlp/lib/python3.6/site-packages/datasets/load.py", line 1443, in load_dataset_builder path, revision=revision, download_config=download_config, download_mode=download_mode, data_files=data_files File "/mnt/home/zhouyang.lzy/.conda/envs/naivenlp/lib/python3.6/site-packages/datasets/load.py", line 1157, in dataset_module_factory raise e1 from None File "/mnt/home/zhouyang.lzy/.conda/envs/naivenlp/lib/python3.6/site-packages/datasets/load.py", line 1144, in dataset_module_factory download_mode=download_mode, File "/mnt/home/zhouyang.lzy/.conda/envs/naivenlp/lib/python3.6/site-packages/datasets/load.py", line 798, in get_module raise FileNotFoundError(f"No data files or dataset script found in {self.path}") AttributeError: 'CommunityDatasetModuleFactoryWithoutScript' object has no attribute 'path' ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.15.1 - Platform: linux - Python version: 3.6.13 - PyArrow version: 6.0.1
luozhouyang
https://github.com/huggingface/datasets/issues/3331
null
false
1,065,176,619
3,330
Change TriviaQA license (#3313)
closed
[]
2021-11-28T03:26:45
2021-11-29T11:24:21
2021-11-29T11:24:21
Fixes (#3313)
avinashsai
https://github.com/huggingface/datasets/pull/3330
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3330", "html_url": "https://github.com/huggingface/datasets/pull/3330", "diff_url": "https://github.com/huggingface/datasets/pull/3330.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3330.patch", "merged_at": "2021-11-29T11:24:21" }
true
1,065,096,971
3,329
Map function: Type error on iter #999
closed
[ "Hi, thanks for reporting.\r\n\r\nIt would be really helpful if you could provide the actual code of the `text_numbers_to_int` function so we can reproduce the error.", "```\r\ndef text_numbers_to_int(text, column=\"\"):\r\n \"\"\"\r\n Convert text numbers to int.\r\n\r\n :param text: text numbers\r\n ...
2021-11-27T17:53:05
2021-11-29T20:40:15
2021-11-29T20:40:15
## Describe the bug Using the map function, it throws a type error on iter #999 Here is the code I am calling: ``` dataset = datasets.load_dataset('squad') dataset['validation'].map(text_numbers_to_int, input_columns=['context'], fn_kwargs={'column': 'context'}) ``` text_numbers_to_int returns the input text with numbers replaced in the format {'context': text} It happens at ` File "C:\Users\lonek\anaconda3\envs\ai\Lib\site-packages\datasets\arrow_writer.py", line 289, in <listcomp> [row[0][col] for row in self.current_examples], type=col_type, try_type=col_try_type, col=col ` The issue is that the list comprehension expects self.current_examples to be type tuple(dict, str), but for some reason 26 out of 1000 of the sefl.current_examples are type tuple(str, str) Here is an example of what self.current_examples should be ({'context': 'Super Bowl 50 was an...merals 50.'}, '') Here is an example of what self.current_examples are when it throws the error: ('The Panthers used th... Marriott.', '')
josephkready666
https://github.com/huggingface/datasets/issues/3329
null
false
1,065,015,262
3,328
Quick fix error formatting
closed
[]
2021-11-27T11:47:48
2021-11-29T13:32:42
2021-11-29T13:32:42
While working on a dataset, I got the error ``` TypeError: Provided `function` which is applied to all elements of table returns a `dict` of types {[type(x) for x in processed_inputs.values()]}. When using `batched=True`, make sure provided `function` returns a `dict` of types like `{allowed_batch_return_types}`. ``` This PR should fix the formatting of this error
NouamaneTazi
https://github.com/huggingface/datasets/pull/3328
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3328", "html_url": "https://github.com/huggingface/datasets/pull/3328", "diff_url": "https://github.com/huggingface/datasets/pull/3328.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3328.patch", "merged_at": "2021-11-29T13:32:42" }
true
1,064,675,888
3,327
"Shape of query is incorrect, it has to be either a 1D array or 2D (1, N)"
closed
[ "#3323 " ]
2021-11-26T16:26:36
2021-11-26T16:44:11
2021-11-26T16:44:11
## Describe the bug Passing a correctly shaped Numpy-Array to get_nearest_examples leads to the Exception "Shape of query is incorrect, it has to be either a 1D array or 2D (1, N)" Probably the reason for this is a wrongly converted assertion. 1.15.1: `assert len(query.shape) == 1 or (len(query.shape) == 2 and query.shape[0] == 1)` 1.16.1: ``` if len(query.shape) != 1 or (len(query.shape) == 2 and query.shape[0] != 1): raise ValueError("Shape of query is incorrect, it has to be either a 1D array or 2D (1, N)") ``` ## Steps to reproduce the bug follow the steps described here: https://huggingface.co/course/chapter5/6?fw=tf ```python question_embedding.shape # (1, 768) scores, samples = embeddings_dataset.get_nearest_examples( "embeddings", question_embedding, k=5 # Error ) # "Shape of query is incorrect, it has to be either a 1D array or 2D (1, N)" ``` ## Expected results Should work without exception ## Actual results Throws exception ## Environment info - `datasets` version: 1.15.1 - Platform: Darwin-20.6.0-x86_64-i386-64bit - Python version: 3.7.12 - PyArrow version: 6.0.
eliasws
https://github.com/huggingface/datasets/issues/3327
null
false
1,064,664,479
3,326
Fix import `datasets` on python 3.10
closed
[]
2021-11-26T16:10:00
2021-11-26T16:31:23
2021-11-26T16:31:23
In python 3.10 it's no longer possible to use `functools.wraps` on a method decorated with `classmethod`. To fix this I inverted the order of the `inject_arrow_table_documentation` and `classmethod` decorators Fix #3324
lhoestq
https://github.com/huggingface/datasets/pull/3326
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3326", "html_url": "https://github.com/huggingface/datasets/pull/3326", "diff_url": "https://github.com/huggingface/datasets/pull/3326.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3326.patch", "merged_at": "2021-11-26T16:31:23" }
true
1,064,663,075
3,325
Update conda dependencies
closed
[]
2021-11-26T16:08:07
2021-11-26T16:20:37
2021-11-26T16:20:36
Some dependencies minimum versions were outdated. For example `pyarrow` and `huggingface_hub`
lhoestq
https://github.com/huggingface/datasets/pull/3325
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3325", "html_url": "https://github.com/huggingface/datasets/pull/3325", "diff_url": "https://github.com/huggingface/datasets/pull/3325.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3325.patch", "merged_at": "2021-11-26T16:20:36" }
true
1,064,661,212
3,324
Can't import `datasets` in python 3.10
closed
[]
2021-11-26T16:06:14
2021-11-26T16:31:23
2021-11-26T16:31:23
When importing `datasets` I'm getting this error in python 3.10: ```python Traceback (most recent call last): File "<string>", line 1, in <module> File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/__init__.py", line 34, in <module> from .arrow_dataset import Dataset, concatenate_datasets File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/arrow_dataset.py", line 47, in <module> from .arrow_reader import ArrowReader File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/arrow_reader.py", line 33, in <module> from .table import InMemoryTable, MemoryMappedTable, Table, concat_tables File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/table.py", line 334, in <module> class InMemoryTable(TableBlock): File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/table.py", line 361, in InMemoryTable def from_pandas(cls, *args, **kwargs): File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/table.py", line 24, in wrapper out = wraps(arrow_table_method)(method) File "/Users/quentinlhoest/.pyenv/versions/3.10.0/lib/python3.10/functools.py", line 61, in update_wrapper wrapper.__wrapped__ = wrapped AttributeError: readonly attribute ``` This makes the conda build fail. I'm opening a PR to fix this and do a patch release 1.16.1
lhoestq
https://github.com/huggingface/datasets/issues/3324
null
false
1,064,660,452
3,323
Fix wrongly converted assert
closed
[ "Closes #3327 " ]
2021-11-26T16:05:39
2021-11-26T16:44:12
2021-11-26T16:44:11
Seems like this assertion was replaced by an exception but the condition got wrongly converted.
eliasws
https://github.com/huggingface/datasets/pull/3323
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3323", "html_url": "https://github.com/huggingface/datasets/pull/3323", "diff_url": "https://github.com/huggingface/datasets/pull/3323.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3323.patch", "merged_at": "2021-11-26T16:44:11" }
true
1,064,429,705
3,322
Add missing tags to XTREME
closed
[]
2021-11-26T12:37:05
2021-11-29T13:40:07
2021-11-29T13:40:06
Add missing tags to the XTREME benchmark for better discoverability.
mariosasko
https://github.com/huggingface/datasets/pull/3322
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3322", "html_url": "https://github.com/huggingface/datasets/pull/3322", "diff_url": "https://github.com/huggingface/datasets/pull/3322.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3322.patch", "merged_at": "2021-11-29T13:40:06" }
true
1,063,858,386
3,321
Update URL of tatoeba subset of xtreme
closed
[ "<s>To be more precise: `os.path.join` is replaced on-the-fly by `xjoin` anyway with patching, to extend it to remote files</s>", "Oh actually just ignore what I said: they were used to concatenate URLs, which is not recommended. Let me fix that again by appending using `+`" ]
2021-11-25T18:42:31
2021-11-26T10:30:30
2021-11-26T10:30:30
Updates the URL of the tatoeba subset of xtreme. Additionally, replaces `os.path.join` with `xjoin` to correctly join the URL segments on Windows. Fix #3320
mariosasko
https://github.com/huggingface/datasets/pull/3321
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3321", "html_url": "https://github.com/huggingface/datasets/pull/3321", "diff_url": "https://github.com/huggingface/datasets/pull/3321.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3321.patch", "merged_at": "2021-11-26T10:30:29" }
true
1,063,531,992
3,320
Can't get tatoeba.rus dataset
closed
[]
2021-11-25T12:31:11
2021-11-26T10:30:29
2021-11-26T10:30:29
## Describe the bug It gives an error. > FileNotFoundError: Couldn't find file at https://github.com/facebookresearch/LASER/raw/master/data/tatoeba/v1/tatoeba.rus-eng.rus ## Steps to reproduce the bug ```python data=load_dataset("xtreme","tatoeba.rus", split="validation") ``` ## Solution The library tries to access the **master** branch. In the github repo of facebookresearch, it is in the **main** branch.
mmg10
https://github.com/huggingface/datasets/issues/3320
null
false
1,062,749,654
3,319
Add push_to_hub docs
closed
[ "Looks good to me! :)\r\n\r\nMaybe we can mention that users can also set the `private` argument if they want to keep their dataset private? It would lead nicely into the next section on Privacy.", "Thanks for your comments, I fixed the capitalization for consistency and added an passage to mention the `private` ...
2021-11-24T18:21:11
2021-11-25T14:47:46
2021-11-25T14:47:46
Since #3098 it's now possible to upload a dataset on the Hub directly from python using the `push_to_hub` method. I just added a section in the "Upload a dataset to the Hub" tutorial. I kept the section quite simple but let me know if it sounds good to you @LysandreJik @stevhliu :)
lhoestq
https://github.com/huggingface/datasets/pull/3319
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3319", "html_url": "https://github.com/huggingface/datasets/pull/3319", "diff_url": "https://github.com/huggingface/datasets/pull/3319.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3319.patch", "merged_at": "2021-11-25T14:47:46" }
true
1,062,369,717
3,318
Finish transition to PyArrow 3.0.0
closed
[]
2021-11-24T12:30:14
2021-11-24T15:35:05
2021-11-24T15:35:04
Finish transition to PyArrow 3.0.0 that was started in #3098.
mariosasko
https://github.com/huggingface/datasets/pull/3318
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3318", "html_url": "https://github.com/huggingface/datasets/pull/3318", "diff_url": "https://github.com/huggingface/datasets/pull/3318.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3318.patch", "merged_at": "2021-11-24T15:35:04" }
true
1,062,284,447
3,317
Add desc parameter to Dataset filter method
closed
[ "Hi,\r\n\r\n`Dataset.map` allows more generic transforms compared to `Dataset.filter`, which purpose is very specific (to filter examples based on a condition). That's why I don't think we need the `desc` parameter there for consistency. #3196 has added descriptions to the `Dataset` methods that call `.map` intern...
2021-11-24T11:01:36
2022-01-05T18:31:24
2022-01-05T18:31:24
**Is your feature request related to a problem? Please describe.** As I was filtering very large datasets I noticed the filter method doesn't have the desc parameter which is available in the map method. Why don't we add a desc parameter to the filter method both for consistency and it's nice to give some feedback to users during long operations on Datasets? **Describe the solution you'd like** Add desc parameter to Dataset filter method **Describe alternatives you've considered** N/A **Additional context** N/A
vblagoje
https://github.com/huggingface/datasets/issues/3317
null
false
1,062,185,822
3,316
Add RedCaps dataset
closed
[]
2021-11-24T09:23:02
2022-01-12T14:13:15
2022-01-12T14:13:15
## Adding a Dataset - **Name:** RedCaps - **Description:** Web-curated image-text data created by the people, for the people - **Paper:** https://arxiv.org/abs/2111.11431 - **Data:** https://redcaps.xyz/ - **Motivation:** Multimodal image-text dataset: 12M+ Image-text pairs Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md). Proposed by @patil-suraj
albertvillanova
https://github.com/huggingface/datasets/issues/3316
null
false
1,061,678,452
3,315
Removing query params for dynamic URL caching
closed
[ "IMO it makes more sense to have `ignore_url_params` as an attribute of `DownloadConfig` to avoid defining a new argument in `DownloadManger`'s methods.", "@mariosasko that would make sense to me too, but it seems like `DownloadConfig` wasn't intended to be modified from a dataset loading script. @lhoestq wdyt?",...
2021-11-23T20:24:12
2021-11-25T14:44:32
2021-11-25T14:44:31
The main use case for this is to make dynamically generated private URLs (like the ones returned by CommonVoice API) compatible with the datasets' caching logic. Usage example: ```python import datasets class CommonVoice(datasets.GeneratorBasedBuilder): def _info(self): return datasets.DatasetInfo() def _split_generators(self, dl_manager): dl_manager.download_config.ignore_url_params = True HUGE_URL = "https://mozilla-common-voice-datasets.s3.dualstack.us-west-2.amazonaws.com/cv-corpus-7.0-2021-07-21/cv-corpus-7.0-2021-07-21-ab.tar.gz?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=ASIAQ3GQRTO3IU5JYB5K%2F20211125%2Fus-west-2%2Fs3%2Faws4_request&X-Amz-Date=20211125T131423Z&X-Amz-Expires=43200&X-Amz-Security-Token=FwoGZXIvYXdzEL7%2F%2F%2F%2F%2F%2F%2F%2F%2F%2FwEaDLsZw7Nj0d9h4rgheyKSBJJ6bxo1JdWLXAUhLMrUB8AXfhP8Ge4F8dtjwXmvGJgkIvdMT7P4YOEE1pS3mW8AyKsz7Z7IRVCIGQrOH1AbxGVVcDoCMMswXEOqL3nJFihKLf99%2F6l8iJVZdzftRUNgMhX5Hz0xSIL%2BzRDpH5nYa7C6YpEdOdW81CFVXybx7WUrX13wc8X4ZlUj7zrWcWf5p2VEIU5Utb7YHVi0Y5TQQiZSDoedQl0j4VmMuFkDzoobIO%2BvilgGeE2kIX0E62X423mEGNu4uQV5JsOuLAtv3GVlemsqEH3ZYrXDuxLmnvGj5HfMtySwI4vKv%2BlnnirD29o7hxvtidXiA8JMWhp93aP%2Fw7sod%2BPPbb5EqP%2B4Qb2GJ1myClOKcLEY0cqoy7XWm8NeVljLJojnFJVS5mNFBAzCCTJ%2FidxNsj8fflzkRoAzYaaPBuOTL1dgtZCdslK3FAuEvw0cik7P9A7IYiULV33otSHKMPcVfNHFsWQljs03gDztsIUWxaXvu6ck5vCcGULsHbfe6xoMPm2bR9jtKLONsslPcnzWIf7%2Fch2w%2F%2BjtTCd9IxaH4kytyJ6mIjpV%2FA%2F2h9qeDnDFsCphnMjAzPQn6tqCgTtPcyJ2b8c94ncgUnE4mepx%2FDa%2FanAEsrg9RPdmbdoPswzHn1IClh91IfSN74u95DZUxlPeZrHG5HxVCN3dKO6j%2Ft1xd20L0hEtazDdKOr8%2FYwGMirp8rp%2BII0pYOwQOrYHqH%2FREX2dRJctJtwE86Qj1eU8BAdXuFIkLC4NWXw%3D&X-Amz-Signature=1b8108d29b0e9c2bf6c7246e58ca8d5749a83de0704757ad8e8a44d78194691f&X-Amz-SignedHeaders=host" dl_path = dl_manager.download_and_extract(HUGE_URL) print(dl_path) HUGE_URL += "&some_new_or_changed_param=12345" dl_path = dl_manager.download_and_extract(HUGE_URL) print(dl_path) dl_manager = datasets.DownloadManager(dataset_name="common_voice") CommonVoice()._split_generators(dl_manager) ``` Output: ``` /home/user/.cache/huggingface/datasets/downloads/6ef2a377398ff3309554be040caa78414e6562d623dbd0ce8fc262459a7f8ec6 /home/user/.cache/huggingface/datasets/downloads/6ef2a377398ff3309554be040caa78414e6562d623dbd0ce8fc262459a7f8ec6 ```
anton-l
https://github.com/huggingface/datasets/pull/3315
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3315", "html_url": "https://github.com/huggingface/datasets/pull/3315", "diff_url": "https://github.com/huggingface/datasets/pull/3315.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3315.patch", "merged_at": "2021-11-25T14:44:31" }
true
1,061,448,227
3,314
Adding arg to pass process rank to `map`
closed
[ "Some commits seem to be there twice (made the mistake of rebasing because I wasn't sure whether the doc had changed), is this an issue @lhoestq ?" ]
2021-11-23T15:55:21
2021-11-24T11:54:13
2021-11-24T11:54:13
This PR adds a `with_rank` argument to `map` that gives the user the possibility to pass the rank of each process to their function. This is mostly designed for multi-GPU map (each process can be sent to a different device thanks to the rank). I've also added tests. I'm putting the PR up so you can check the code, I'll add a multi-GPU example to the doc (+ write a bit in the doc for the new arg)
TevenLeScao
https://github.com/huggingface/datasets/pull/3314
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3314", "html_url": "https://github.com/huggingface/datasets/pull/3314", "diff_url": "https://github.com/huggingface/datasets/pull/3314.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3314.patch", "merged_at": "2021-11-24T11:54:13" }
true
1,060,933,392
3,313
TriviaQA License Mismatch
closed
[ "Hi ! You're completely right, this must be mentioned in the dataset card.\r\nIf you're interesting in contributing, feel free to open a pull request to mention this in the `trivia_qa` dataset card in the \"Licensing Information\" section at https://github.com/huggingface/datasets/blob/master/datasets/trivia_qa/REA...
2021-11-23T08:00:15
2021-11-29T11:24:21
2021-11-29T11:24:21
## Describe the bug TriviaQA Webpage at http://nlp.cs.washington.edu/triviaqa/ says they do not own the copyright to the data. However, Huggingface datasets at https://huggingface.co/datasets/trivia_qa mentions that the dataset is released under Apache License Is the License Information on HuggingFace correct?
akhilkedia
https://github.com/huggingface/datasets/issues/3313
null
false
1,060,440,346
3,312
add bl books genre dataset
closed
[ "To fix the CI, feel free to run the `make style` command to format the code.\r\n\r\nThen it also looks like the dummy_data.zip archives are all empty, which makes the tests fail. Can you try regenerating them ? They should have one file inside which is a dummy version of the file at https://bl.iro.bl.uk/downloads/...
2021-11-22T17:54:50
2021-12-02T16:10:29
2021-12-02T16:07:47
First of all thanks for the fantastic library/collection of datasets 🤗 This pull request adds a dataset of metadata from digitised (mostly 19th Century) books from the British Library The [data](https://bl.iro.bl.uk/concern/datasets/1e1ccb46-65b4-4481-b6f8-b8129d5da053) contains various metadata about the books. In addition, a subset of the data includes 'genre' information which can be used for supervised text classification tasks. I hope that this offers easier access to a dataset for doing text classification on GLAM (galleries, libraries, archives and museums) data. I have tried to create three configurations that provide both an 'easy' version of the dataset if you want to use it for training a genre classification model and a more 'raw' version of the data for other potential use cases for the data. I am open to suggestions if this doesn't make sense. Similarly, for some of the arrow datatypes, I have had to fall back to strings since there are missing values for some fields/rows but I may have missed a more elegant way of dealing with it.
davanstrien
https://github.com/huggingface/datasets/pull/3312
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3312", "html_url": "https://github.com/huggingface/datasets/pull/3312", "diff_url": "https://github.com/huggingface/datasets/pull/3312.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3312.patch", "merged_at": "2021-12-02T16:07:47" }
true
1,060,387,957
3,311
Add WebSRC
open
[]
2021-11-22T16:58:33
2021-11-22T16:58:33
null
## Adding a Dataset - **Name:** WebSRC - **Description:** WebSRC is a novel Web-based Structural Reading Comprehension dataset. It consists of 0.44M question-answer pairs, which are collected from 6.5K web pages with corresponding HTML source code, screenshots and metadata. - **Paper:** https://arxiv.org/abs/2101.09465 - **Data:** https://x-lance.github.io/WebSRC/dashboard.html# - **Motivation:** Currently adding MarkupLM to HuggingFace Transformers, which achieves SOTA on this dataset. Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
NielsRogge
https://github.com/huggingface/datasets/issues/3311
null
false
1,060,098,104
3,310
Fatal error condition occurred in aws-c-io
closed
[ "Hi ! Are you having this issue only with this specific dataset, or it also happens with other ones like `squad` ?", "@lhoestq It happens also on `squad`. It successfully downloads the whole dataset and then crashes on: \r\n\r\n```\r\nFatal error condition occurred in D:\\bld\\aws-c-io_1633633258269\\work\\source...
2021-11-22T12:27:54
2023-02-08T10:31:05
2021-11-29T22:22:37
## Describe the bug Fatal error when using the library ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset('wikiann', 'en') ``` ## Expected results No fatal errors ## Actual results ``` Fatal error condition occurred in D:\bld\aws-c-io_1633633258269\work\source\event_loop.c:74: aws_thread_launch(&cleanup_thread, s_event_loop_destroy_async_thread_fn, el_group, &thread_options) == AWS_OP_SUCCESS Exiting Application ``` ## Environment info - `datasets` version: 1.15.2.dev0 - Platform: Windows-10-10.0.22504-SP0 - Python version: 3.8.12 - PyArrow version: 6.0.0
Crabzmatic
https://github.com/huggingface/datasets/issues/3310
null
false
1,059,496,154
3,309
fix: files counted twice in inferred structure
closed
[ "I see it creates some errors in the tests.\r\n\r\nAnother solution if needed is to add something like `data_files = list(set(data_files))` after [this line](https://github.com/huggingface/datasets/blob/8555197a3fe826e98bd0206c2d031c4488c53c5c/src/datasets/data_files.py#L511)", "Hi ! Thanks for the correction :)\...
2021-11-21T21:50:38
2021-11-23T17:00:58
2021-11-23T17:00:58
Files were counted twice in a structure like: ``` my_dataset_local_path/ ├── README.md └── data/ ├── train/ │ ├── shard_0.csv │ ├── shard_1.csv │ ├── shard_2.csv │ └── shard_3.csv └── valid/ ├── shard_0.csv └── shard_1.csv ``` The reason is that they were matching both `*train*/*` and `*train*/**/*`. This PR fixes it. @lhoestq
borisdayma
https://github.com/huggingface/datasets/pull/3309
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3309", "html_url": "https://github.com/huggingface/datasets/pull/3309", "diff_url": "https://github.com/huggingface/datasets/pull/3309.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3309.patch", "merged_at": "2021-11-23T17:00:58" }
true
1,059,255,705
3,308
"dataset_infos.json" missing for chr_en and mc4
open
[ "Hi ! Thanks for reporting :) \r\nWe can easily add the metadata for `chr_en` IMO, but for mC4 it will take more time, since it requires to count the number of examples in each language", "No problem. I am trying to do some analysis on the metadata of all available datasets. Is reading `metadata_infos.json` for e...
2021-11-21T00:07:22
2022-01-19T13:55:32
null
## Describe the bug In the repository, every dataset has its metadata in a file called`dataset_infos.json`. But, this file is missing for two datasets: `chr_en` and `mc4`. ## Steps to reproduce the bug Check [chr_en](https://github.com/huggingface/datasets/tree/master/datasets/chr_en) and [mc4](https://github.com/huggingface/datasets/tree/master/datasets/mc4)
amitness
https://github.com/huggingface/datasets/issues/3308
null
false
1,059,226,297
3,307
Add IndoNLI dataset
closed
[ "@lhoestq thanks for the review! I've modified the labels to follow other NLI datasets.\r\nPlease review my change and let me know if I miss anything." ]
2021-11-20T20:46:03
2021-11-25T14:51:48
2021-11-25T14:51:48
This PR adds IndoNLI dataset, from https://aclanthology.org/2021.emnlp-main.821/
afaji
https://github.com/huggingface/datasets/pull/3307
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3307", "html_url": "https://github.com/huggingface/datasets/pull/3307", "diff_url": "https://github.com/huggingface/datasets/pull/3307.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3307.patch", "merged_at": "2021-11-25T14:51:48" }
true
1,059,185,860
3,306
nested sequence feature won't encode example if the first item of the outside sequence is an empty list
closed
[ "knock knock", "Hi, thanks for reporting! I've linked a PR that should fix the issue.", "I've checked the PR and it looks great, thanks a lot!" ]
2021-11-20T16:57:54
2021-12-08T13:02:15
2021-12-08T13:02:15
## Describe the bug As the title, nested sequence feature won't encode example if the first item of the outside sequence is an empty list. ## Steps to reproduce the bug ```python from datasets import Features, Sequence, ClassLabel features = Features({ 'x': Sequence(Sequence(ClassLabel(names=['a', 'b']))), }) print(features.encode_batch({ 'x': [ [['a'], ['b']], [[], ['b']], ] })) ``` ## Expected results print `{'x': [[[0], [1]], [[], ['1']]]}` ## Actual results print `{'x': [[[0], [1]], [[], ['b']]]}` ## Environment info - `datasets` version: 1.15.1 - Platform: Linux-5.13.0-21-generic-x86_64-with-glibc2.34 - Python version: 3.9.7 - PyArrow version: 6.0.0 ## Additional information I think the issue stems from [here](https://github.com/huggingface/datasets/blob/8555197a3fe826e98bd0206c2d031c4488c53c5c/src/datasets/features/features.py#L847-L848).
function2-llx
https://github.com/huggingface/datasets/issues/3306
null
false
1,059,161,000
3,305
asserts replaced with exception for ``fingerprint.py``, ``search.py``, ``arrow_writer.py`` and ``metric.py``
closed
[]
2021-11-20T14:51:23
2021-11-22T18:24:32
2021-11-22T17:08:13
Addresses #3171 Fixes exception for ``fingerprint.py``, ``search.py``, ``arrow_writer.py`` and ``metric.py`` and modified tests
Ishan-Kumar2
https://github.com/huggingface/datasets/pull/3305
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3305", "html_url": "https://github.com/huggingface/datasets/pull/3305", "diff_url": "https://github.com/huggingface/datasets/pull/3305.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3305.patch", "merged_at": "2021-11-22T17:08:13" }
true
1,059,130,494
3,304
Dataset object has no attribute `to_tf_dataset`
closed
[ "The issue is due to the older version of transformers and datasets. It has been resolved by upgrading their versions.\r\n\r\n```\r\n# upgrade transformers and datasets to latest versions\r\n!pip install --upgrade transformers\r\n!pip install --upgrade datasets\r\n```\r\n\r\nRegards!" ]
2021-11-20T12:03:59
2021-11-21T07:07:25
2021-11-21T07:07:25
I am following HuggingFace Course. I am at Fine-tuning a model. Link: https://huggingface.co/course/chapter3/2?fw=tf I use tokenize_function and `map` as mentioned in the course to process data. `# define a tokenize function` `def Tokenize_function(example):` ` return tokenizer(example['sentence'], truncation=True)` `# tokenize entire data` `tokenized_data = raw_data.map(Tokenize_function, batched=True)` I get Dataset object at this point. When I try converting this to a TF dataset object as mentioned in the course, it throws the following error. `# convert to TF dataset` `train_data = tokenized_data["train"].to_tf_dataset( ` ` columns = ['attention_mask', 'input_ids', 'token_type_ids'], ` ` label_cols = ['label'], ` ` shuffle = True, ` ` collate_fn = data_collator, ` ` batch_size = 8 ` `)` Output: `---------------------------------------------------------------------------` `AttributeError Traceback (most recent call last)` `/tmp/ipykernel_42/103099799.py in <module>` ` 1 # convert to TF dataset` `----> 2 train_data = tokenized_data["train"].to_tf_dataset( \` ` 3 columns = ['attention_mask', 'input_ids', 'token_type_ids'], \` ` 4 label_cols = ['label'], \` ` 5 shuffle = True, \` `AttributeError: 'Dataset' object has no attribute 'to_tf_dataset'` When I look for `dir(tokenized_data["train"])`, there is no method or attribute in the name of `to_tf_dataset`. Why do I get this error? And how to clear this? Please help me.
RajkumarGalaxy
https://github.com/huggingface/datasets/issues/3304
null
false
1,059,129,732
3,303
DataCollatorWithPadding: TypeError
closed
[ "\r\n> \r\n> Input:\r\n> \r\n> ```\r\n> tokenizer = AutoTokenizer.from_pretrained(checkpoint)\r\n> data_collator = DataCollatorWithPadding(tokenizer=tokenizer, return_tensors=\"tf\")\r\n> ```\r\n> \r\n> Output:\r\n> \r\n> ```\r\n> TypeError Traceback (most recent call last)\r\n> /tmp...
2021-11-20T11:59:55
2021-11-21T07:05:37
2021-11-21T07:05:37
Hi, I am following the HuggingFace course. I am now at Fine-tuning [https://huggingface.co/course/chapter3/3?fw=tf](https://huggingface.co/course/chapter3/3?fw=tf). When I set up `DataCollatorWithPadding` as following I got an error while trying to reproduce the course code in Kaggle. This error occurs with either a CPU-only-device or a GPU-device. Input: ```checkpoint = 'bert-base-uncased' tokenizer = AutoTokenizer.from_pretrained(checkpoint) data_collator = DataCollatorWithPadding(tokenizer=tokenizer, return_tensors="tf") ``` Output: ```--------------------------------------------------------------------------- TypeError Traceback (most recent call last) /tmp/ipykernel_42/1563280798.py in <module> 1 checkpoint = 'bert-base-uncased' 2 tokenizer = AutoTokenizer.from_pretrained(checkpoint) ----> 3 data_collator = DataCollatorWithPadding(tokenizer=tokenizer, return_tensors="pt") TypeError: __init__() got an unexpected keyword argument 'return_tensors' ``` When I call `help` method, it too confirms that there is no argument `return_tensors`. Input: ``` help(DataCollatorWithPadding.__init__) ``` Output: ``` Help on function __init__ in module transformers.data.data_collator: __init__(self, tokenizer: transformers.tokenization_utils_base.PreTrainedTokenizerBase, padding: Union[bool, str, transformers.file_utils.PaddingStrategy] = True, max_length: Union[int, NoneType] = None, pad_to_multiple_of: Union[int, NoneType] = None) -> None ``` But, the source file *[Data Collator - docs](https://huggingface.co/transformers/main_classes/data_collator.html#datacollatorwithpadding)* says that there is such an argument. By default, it returns Pytorch tensors while I need TF tensors. Where do I miss? Please help me.
RajkumarGalaxy
https://github.com/huggingface/datasets/issues/3303
null
false
1,058,907,168
3,302
fix old_val typo in f-string
closed
[]
2021-11-19T20:51:08
2021-11-25T22:14:43
2021-11-22T17:04:19
This PR is to correct a typo in #3277 that @Carlosbogo revieled in a comment. Related closed issue : #3257 Sorry about that 😅.
Mehdi2402
https://github.com/huggingface/datasets/pull/3302
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3302", "html_url": "https://github.com/huggingface/datasets/pull/3302", "diff_url": "https://github.com/huggingface/datasets/pull/3302.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3302.patch", "merged_at": "2021-11-22T17:04:19" }
true
1,058,718,957
3,301
Add wikipedia tags
closed
[]
2021-11-19T16:39:25
2021-11-19T16:49:30
2021-11-19T16:49:29
Add the missing tags to the wikipedia dataset card. I also added the missing languages code in our language codes list. This should also fix the code snippet that is presented on the Hub to load the dataset: fix https://github.com/huggingface/datasets/issues/3292
lhoestq
https://github.com/huggingface/datasets/pull/3301
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3301", "html_url": "https://github.com/huggingface/datasets/pull/3301", "diff_url": "https://github.com/huggingface/datasets/pull/3301.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3301.patch", "merged_at": "2021-11-19T16:49:29" }
true
1,058,644,459
3,300
❓ Dataset loading script from Hugging Face Hub
closed
[ "Hi ! In the next version of `datasets`, your train and test splits will be correctly separated (changes from #3027) if you create a dataset repository with only your CSV files.\r\n\r\nAlso it seems that you overwrite the `data_files` and `data_dir` arguments in your code, when you instantiate the AGNewsConfig obje...
2021-11-19T15:20:52
2021-12-22T10:57:56
2021-12-22T10:57:56
Hi there, I am trying to add my custom `ag_news` with its own loading script on the Hugging Face datasets hub. In particular, I would like to test the addition of a second configuration to the existing `ag_news` dataset. Once it works in my hub, I plan to make a PR to the original dataset. However, in trying to do so I have encountered certain problems as detailed below. Issues I have encountered: - Without a loading script, the train and test files are loaded together into a unique `dataset.Dataset` -> so I wrote a loading script. Also, I need a loading script otherwise I cannot specify multiple configurations - Once my loading script is working locally, I do not manage to make it work on the hub. In particular, I would like to be able to load the dataset like this ```python load_dataset("pietrolesci/ag_news", name="my_configuration") ``` Apparently, the `load_dataset` is able to pick up the loading script from the hub and run it. However, it errors because it is unable to find the files. The structure of my hub repo is the following ``` ag_news.py train.csv test.csv ``` and the loading script I specify `data_dir=Path(__file__).parent` and `data_files=DataFilesDict({"train": "train.csv", "test": "test.csv"})`. In the documentation I could not find info regarding loading a dataset from the hub using a loading script present on the hub. Any suggestion is very much appreciated. Best, Pietro Link to the hub repo: https://huggingface.co/datasets/pietrolesci/ag_news BONUS: how can I make the data viewer work in this specific case? :)
pietrolesci
https://github.com/huggingface/datasets/issues/3300
null
false
1,058,518,213
3,299
Add option to find unique elements in nested sequences when calling `Dataset.unique`
open
[ "Hi @mariosasko!\r\n\r\nHas this been patched into any of the releases?", "Hi! Not yet, would you be interested in contributing a PR? I can give you some pointers if needed. ", "@mariosasko did this ever get implemented? Willing to help if you are still up for it.", "@dcruiz01 No, but here is an example of ho...
2021-11-19T13:16:06
2023-05-19T14:45:40
null
It would be nice to have an option to flatten nested sequences to find unique elements stored in them when calling `Dataset.unique`. ~~Currently, `Dataset.unique` only supports finding unique sequences and not unique elements in that situation.~~
mariosasko
https://github.com/huggingface/datasets/issues/3299
null
false
1,058,420,201
3,298
Agnews dataset viewer is not working
closed
[ "Hi ! Thanks for reporting\r\nWe've already fixed the code that generates the preview for this dataset, we'll release the fix soon :)", "Hi @lhoestq, thanks for your feedback!", "Fixed in the viewer.\r\n\r\nhttps://huggingface.co/datasets/ag_news" ]
2021-11-19T11:18:59
2021-12-21T16:24:05
2021-12-21T16:24:05
## Dataset viewer issue for '*name of the dataset*' **Link:** https://huggingface.co/datasets/ag_news Hi there, the `ag_news` dataset viewer is not working. Am I the one who added this dataset? No
pietrolesci
https://github.com/huggingface/datasets/issues/3298
null
false
1,058,263,859
3,297
.map() cache is wrongfully reused - only happens when the mapping function is imported
open
[ "Hi ! Thanks for reporting. Indeed this is a current limitation of the usage we have of `dill` in `datasets`. I'd suggest you use your workaround for now until we find a way to fix this. Maybe functions that are not coming from a module not installed with pip should be dumped completely, rather than only taking the...
2021-11-19T08:18:36
2025-09-08T10:12:16
null
## Describe the bug When `.map` is used with a mapping function that is imported, the cache is reused even if the mapping function has been modified. The reason for this is that `dill` that is used for creating the fingerprint [pickles imported functions by reference](https://stackoverflow.com/a/67851411). I guess it is not a widespread case, but it can still lead to unwanted results unnoticeably. ## Steps to reproduce the bug Create files `a.py` and `b.py`: ```python # a.py from datasets import load_dataset def main(): squad = load_dataset("squad") squad.map(mapping_func, batched=True) def mapping_func(examples): ID_LENGTH = 4 examples["id"] = [id_[:ID_LENGTH] for id_ in examples["id"]] return examples if __name__ == "__main__": main() ``` ```python # b.py from datasets import load_dataset from a import mapping_func def main(): squad = load_dataset("squad") squad.map(mapping_func, batched=True) if __name__ == "__main__": main() ``` Run `python b.py` twice: In the first run you will see tqdm bars showing that the data is processed, and in the second run you will see "Loading cached processed dataset at...". Now change `ID_LENGTH` to another number in order to change the mapping function, and run `python b.py` again. You'll see that `.map` loads from the cache the result of the previous mapping function. ## Expected results Run `python a.py` twice: In the first run you will see tqdm bars showing that the data is processed, and in the second run you will see "Loading cached processed dataset at...". Now change `ID_LENGTH` to another number in order to change the mapping function, and run `python a.py` again. You'll see that the dataset is being processed and that there's no reuse of the previous mapping function result. ## Workaround Put the mapping function inside a dummy class as a static method: ```python # a.py class MappingFuncClass: @staticmethod def mapping_func(examples): ID_LENGTH = 4 examples["id"] = [id_[:ID_LENGTH] for id_ in examples["id"]] return examples ``` ```python # b.py from datasets import load_dataset from a import MappingFuncClass def main(): squad = load_dataset("squad") squad.map(MappingFuncClass.mapping_func, batched=True) if __name__ == "__main__": main() ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.15.1 - Platform: Linux-4.4.0-19041-Microsoft-x86_64-with-glibc2.17 - Python version: 3.8.10 - PyArrow version: 4.0.1
eladsegal
https://github.com/huggingface/datasets/issues/3297
null
false
1,057,970,638
3,296
Fix temporary dataset_path creation for URIs related to remote fs
closed
[ "Hi ! Thanks for the fix :) \r\n\r\nI think this should be `extract_path_from_uri` 's responsibility to strip the extra `/` from a badly formatted path like `hdfs:///absolute/path` (or raise an error). Do you think you can simply do the changes in `extract_path_from_uri` ? This way this fix will be available for al...
2021-11-18T23:32:45
2021-12-06T10:45:04
2021-12-06T10:45:04
This aims to close #3295
francisco-perez-sorrosal
https://github.com/huggingface/datasets/pull/3296
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3296", "html_url": "https://github.com/huggingface/datasets/pull/3296", "diff_url": "https://github.com/huggingface/datasets/pull/3296.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3296.patch", "merged_at": "2021-12-06T10:45:03" }
true
1,057,954,892
3,295
Temporary dataset_path for remote fs URIs not built properly in arrow_dataset.py::load_from_disk
closed
[ "Hi ! Good catch and thanks for opening a PR :)\r\n\r\nI just responded in your PR" ]
2021-11-18T23:24:02
2021-12-06T10:45:04
2021-12-06T10:45:04
## Describe the bug When trying to build a temporary dataset path from a remote URI in this block of code: https://github.com/huggingface/datasets/blob/42f6b1d18a4a1b6009b6e62d115491be16dfca22/src/datasets/arrow_dataset.py#L1038-L1042 the result is not the expected when passing an absolute path in an URI like `hdfs:///absolute/path`. ## Steps to reproduce the bug ```python dataset_path = "hdfs:///absolute/path" src_dataset_path = extract_path_from_uri(dataset_path) tmp_dir = get_temporary_cache_files_directory() dataset_path = Path(tmp_dir, src_dataset_path) print(dataset_path) ``` ## Expected results With the code above, we would expect a value in `dataset_path` similar to: `/tmp/tmpnwxyvao5/absolute/path` ## Actual results However, we get a `dataset_path` value like: `/absolute/path` This is because this line here: https://github.com/huggingface/datasets/blob/42f6b1d18a4a1b6009b6e62d115491be16dfca22/src/datasets/arrow_dataset.py#L1041 returns the last absolute path when two absolute paths (the one in `tmp_dir` and the one extracted from the URI in `src_dataset_path`) are passed as arguments. ## Environment info - `datasets` version: 1.13.3 - Platform: Linux-3.10.0-1160.15.2.el7.x86_64-x86_64-with-glibc2.33 - Python version: 3.9.7 - PyArrow version: 5.0.0
francisco-perez-sorrosal
https://github.com/huggingface/datasets/issues/3295
null
false
1,057,495,473
3,294
Add Natural Adversarial Objects dataset
open
[]
2021-11-18T15:34:44
2021-12-08T12:00:02
null
## Adding a Dataset - **Name:** Natural Adversarial Objects (NAO) - **Description:** Natural Adversarial Objects (NAO) is a new dataset to evaluate the robustness of object detection models. NAO contains 7,934 images and 9,943 objects that are unmodified and representative of real-world scenarios, but cause state-of-the-art detection models to misclassify with high confidence. - **Paper:** https://arxiv.org/abs/2111.04204v1 - **Data:** https://drive.google.com/drive/folders/15P8sOWoJku6SSEiHLEts86ORfytGezi8 - **Motivation:** interesting object detection dataset useful for miscclassifications cc @NielsRogge Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
osanseviero
https://github.com/huggingface/datasets/issues/3294
null
false
1,057,004,431
3,293
Pin version exclusion for Markdown
closed
[]
2021-11-18T06:56:01
2021-11-18T10:28:05
2021-11-18T10:28:04
As Markdown version 3.3.5 has a bug, it is better to exclude it in case the users have it previously installed in their environment. Related to #3289, #3286.
albertvillanova
https://github.com/huggingface/datasets/pull/3293
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3293", "html_url": "https://github.com/huggingface/datasets/pull/3293", "diff_url": "https://github.com/huggingface/datasets/pull/3293.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3293.patch", "merged_at": "2021-11-18T10:28:04" }
true
1,056,962,554
3,292
Not able to load 'wikipedia' dataset
closed
[ "Hi ! Indeed it looks like the code snippet on the Hugging face Hub doesn't show the second parameter\r\n\r\n![image](https://user-images.githubusercontent.com/42851186/142649237-45ba55c5-1a64-4c30-8692-2c8120572f92.png)\r\n\r\nThanks for reporting, I'm taking a look\r\n" ]
2021-11-18T05:41:18
2021-11-19T16:49:29
2021-11-19T16:49:29
## Describe the bug I am following the instruction for loading the wikipedia dataset using datasets. However getting the below error. ## Steps to reproduce the bug from datasets import load_dataset dataset = load_dataset("wikipedia") ``` ## Expected results A clear and concise description of the expected results. ## Actual results ~/anaconda3/envs/pytorch_p36/lib/python3.6/site-packages/datasets/builder.py in _create_builder_config(self, name, custom_features, **config_kwargs) 339 "Config name is missing." 340 "\nPlease pick one among the available configs: %s" % list(self.builder_configs.keys()) --> 341 + "\nExample of usage:\n\t`{}`".format(example_of_usage) 342 ) 343 builder_config = self.BUILDER_CONFIGS[0] ValueError: Config name is missing. Please pick one among the available configs: ['20200501.aa', '20200501.ab', '20200501.ace', '20200501.ady', '20200501.af', '20200501.ak', '20200501.als', '20200501.am', '20200501.an', '20200501.ang', '20200501.ar', '20200501.arc', '20200501.arz', '20200501.as', '20200501.ast', '20200501.atj', '20200501.av', '20200501.ay', '20200501.az', '20200501.azb', '20200501.ba', '20200501.bar', '20200501.bat-smg', '20200501.bcl', '20200501.be', '20200501.be-x-old', '20200501.bg', '20200501.bh', '20200501.bi', '20200501.bjn', '20200501.bm', '20200501.bn', '20200501.bo', '20200501.bpy', '20200501.br', '20200501.bs', '20200501.bug', '20200501.bxr', '20200501.ca', '20200501.cbk-zam', '20200501.cdo', '20200501.ce', '20200501.ceb', '20200501.ch', '20200501.cho', '20200501.chr', '20200501.chy', '20200501.ckb', '20200501.co', '20200501.cr', '20200501.crh', '20200501.cs', '20200501.csb', '20200501.cu', '20200501.cv', '20200501.cy', '20200501.da', '20200501.de', '20200501.din', '20200501.diq', '20200501.dsb', '20200501.dty', '20200501.dv', '20200501.dz', '20200501.ee', '20200501.el', '20200501.eml', '20200501.en', '20200501.eo', '20200501.es', '20200501.et', '20200501.eu', '20200501.ext', '20200501.fa', '20200501.ff', '20200501.fi', '20200501.fiu-vro', '20200501.fj', '20200501.fo', '20200501.fr', '20200501.frp', '20200501.frr', '20200501.fur', '20200501.fy', '20200501.ga', '20200501.gag', '20200501.gan', '20200501.gd', '20200501.gl', '20200501.glk', '20200501.gn', '20200501.gom', '20200501.gor', '20200501.got', '20200501.gu', '20200501.gv', '20200501.ha', '20200501.hak', '20200501.haw', '20200501.he', '20200501.hi', '20200501.hif', '20200501.ho', '20200501.hr', '20200501.hsb', '20200501.ht', '20200501.hu', '20200501.hy', '20200501.ia', '20200501.id', '20200501.ie', '20200501.ig', '20200501.ii', '20200501.ik', '20200501.ilo', '20200501.inh', '20200501.io', '20200501.is', '20200501.it', '20200501.iu', '20200501.ja', '20200501.jam', '20200501.jbo', '20200501.jv', '20200501.ka', '20200501.kaa', '20200501.kab', '20200501.kbd', '20200501.kbp', '20200501.kg', '20200501.ki', '20200501.kj', '20200501.kk', '20200501.kl', '20200501.km', '20200501.kn', '20200501.ko', '20200501.koi', '20200501.krc', '20200501.ks', '20200501.ksh', '20200501.ku', '20200501.kv', '20200501.kw', '20200501.ky', '20200501.la', '20200501.lad', '20200501.lb', '20200501.lbe', '20200501.lez', '20200501.lfn', '20200501.lg', '20200501.li', '20200501.lij', '20200501.lmo', '20200501.ln', '20200501.lo', '20200501.lrc', '20200501.lt', '20200501.ltg', '20200501.lv', '20200501.mai', '20200501.map-bms', '20200501.mdf', '20200501.mg', '20200501.mh', '20200501.mhr', '20200501.mi', '20200501.min', '20200501.mk', '20200501.ml', '20200501.mn', '20200501.mr', '20200501.mrj', '20200501.ms', '20200501.mt', '20200501.mus', '20200501.mwl', '20200501.my', '20200501.myv', '20200501.mzn', '20200501.na', '20200501.nah', '20200501.nap', '20200501.nds', '20200501.nds-nl', '20200501.ne', '20200501.new', '20200501.ng', '20200501.nl', '20200501.nn', '20200501.no', '20200501.nov', '20200501.nrm', '20200501.nso', '20200501.nv', '20200501.ny', '20200501.oc', '20200501.olo', '20200501.om', '20200501.or', '20200501.os', '20200501.pa', '20200501.pag', '20200501.pam', '20200501.pap', '20200501.pcd', '20200501.pdc', '20200501.pfl', '20200501.pi', '20200501.pih', '20200501.pl', '20200501.pms', '20200501.pnb', '20200501.pnt', '20200501.ps', '20200501.pt', '20200501.qu', '20200501.rm', '20200501.rmy', '20200501.rn', '20200501.ro', '20200501.roa-rup', '20200501.roa-tara', '20200501.ru', '20200501.rue', '20200501.rw', '20200501.sa', '20200501.sah', '20200501.sat', '20200501.sc', '20200501.scn', '20200501.sco', '20200501.sd', '20200501.se', '20200501.sg', '20200501.sh', '20200501.si', '20200501.simple', '20200501.sk', '20200501.sl', '20200501.sm', '20200501.sn', '20200501.so', '20200501.sq', '20200501.sr', '20200501.srn', '20200501.ss', '20200501.st', '20200501.stq', '20200501.su', '20200501.sv', '20200501.sw', '20200501.szl', '20200501.ta', '20200501.tcy', '20200501.te', '20200501.tet', '20200501.tg', '20200501.th', '20200501.ti', '20200501.tk', '20200501.tl', '20200501.tn', '20200501.to', '20200501.tpi', '20200501.tr', '20200501.ts', '20200501.tt', '20200501.tum', '20200501.tw', '20200501.ty', '20200501.tyv', '20200501.udm', '20200501.ug', '20200501.uk', '20200501.ur', '20200501.uz', '20200501.ve', '20200501.vec', '20200501.vep', '20200501.vi', '20200501.vls', '20200501.vo', '20200501.wa', '20200501.war', '20200501.wo', '20200501.wuu', '20200501.xal', '20200501.xh', '20200501.xmf', '20200501.yi', '20200501.yo', '20200501.za', '20200501.zea', '20200501.zh', '20200501.zh-classical', '20200501.zh-min-nan', '20200501.zh-yue', '20200501.zu'] Example of usage: `load_dataset('wikipedia', '20200501.aa')` I think the other parameter is missing in the load_dataset function that is not shown in the instruction.
abhibisht89
https://github.com/huggingface/datasets/issues/3292
null
false
1,056,689,876
3,291
Use f-strings in the dataset scripts
closed
[]
2021-11-17T22:20:19
2021-11-22T16:40:16
2021-11-22T16:40:16
Uses f-strings to format the .py files in the dataset folder
Carlosbogo
https://github.com/huggingface/datasets/pull/3291
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3291", "html_url": "https://github.com/huggingface/datasets/pull/3291", "diff_url": "https://github.com/huggingface/datasets/pull/3291.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3291.patch", "merged_at": "2021-11-22T16:40:16" }
true
1,056,414,856
3,290
Make several audio datasets streamable
closed
[ "Reading FLAC (for `librispeech_asr`) works OK for me (`soundfile` version: `0.10.3`):\r\n```python\r\nIn [2]: ds = load_dataset(\"datasets/librispeech_asr/librispeech_asr.py\", \"clean\", streaming=True, split=\"train.100\")\r\n\r\nIn [3]: item = next(iter(ds))\r\n\r\nIn [4]: item.keys()\r\nOut[4]: dict_keys(['fil...
2021-11-17T17:43:41
2022-02-01T21:00:52
2021-11-19T15:08:57
<s>Needs https://github.com/huggingface/datasets/pull/3129 to be merged first</s> Make those audio datasets streamable: - [x] common_voice - [x] openslr - [x] vivos - [x] librispeech_asr <s>(still has some issues to read FLAC)</s> *actually it's ok* - [ ] <s>multilingual_librispeech (yet to be converted)</S> *TODO in a separate PR*
lhoestq
https://github.com/huggingface/datasets/pull/3290
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3290", "html_url": "https://github.com/huggingface/datasets/pull/3290", "diff_url": "https://github.com/huggingface/datasets/pull/3290.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3290.patch", "merged_at": "2021-11-19T15:08:57" }
true
1,056,323,715
3,289
Unpin markdown for build_docs now that it's fixed
closed
[]
2021-11-17T16:22:53
2021-11-17T16:23:09
2021-11-17T16:23:08
`markdown`'s bug has been fixed, so this PR reverts #3286
lhoestq
https://github.com/huggingface/datasets/pull/3289
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3289", "html_url": "https://github.com/huggingface/datasets/pull/3289", "diff_url": "https://github.com/huggingface/datasets/pull/3289.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3289.patch", "merged_at": "2021-11-17T16:23:08" }
true
1,056,145,703
3,288
Allow datasets with indices table when concatenating along axis=1
closed
[]
2021-11-17T13:41:28
2021-11-17T15:41:12
2021-11-17T15:41:11
Calls `flatten_indices` on the datasets with indices table in `concatenate_datasets` to fix issues when concatenating along `axis=1`. cc @lhoestq: I decided to flatten all the datasets instead of flattening all the datasets except the largest one in the end. The latter approach fails on the following example: ```python a = Dataset.from_dict({"a": [10, 20, 30, 40]}) b = Dataset.from_dict({"b": [10, 20, 30, 40, 50, 60]}) # largest dataset a = a.select([1, 2, 3]) b = b.select([1, 2, 3]) concatenate_datasets([a, b], axis=1) # fails at line concat_tables(...) because the real length of b's data is 6 and a's length is 3 after flattening (was 4 before flattening) ``` Also, it requires additional re-ordering of indices to prepare them for working with the indices table of the largest dataset. IMO not worth when we save only one `flatten_indices` call. (feel free to check the code of that approach at https://github.com/huggingface/datasets/commit/6acd10481c70950dcfdbfd2bab0bf0c74ad80bcb if you are interested) Fixes #3273
mariosasko
https://github.com/huggingface/datasets/pull/3288
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3288", "html_url": "https://github.com/huggingface/datasets/pull/3288", "diff_url": "https://github.com/huggingface/datasets/pull/3288.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3288.patch", "merged_at": "2021-11-17T15:41:11" }
true
1,056,079,724
3,287
Add The Pile dataset and PubMed Central subset
closed
[]
2021-11-17T12:35:58
2021-12-01T15:29:08
2021-12-01T15:29:07
Add: - The complete final version of The Pile dataset: "all" config - PubMed Central subset of The Pile: "pubmed_central" config Close #1675, close bigscience-workshop/data_tooling#74. CC: @StellaAthena, @lewtun
albertvillanova
https://github.com/huggingface/datasets/pull/3287
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3287", "html_url": "https://github.com/huggingface/datasets/pull/3287", "diff_url": "https://github.com/huggingface/datasets/pull/3287.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3287.patch", "merged_at": "2021-12-01T15:29:06" }
true
1,056,008,586
3,286
Fix build_docs CI
closed
[]
2021-11-17T11:18:56
2021-11-17T11:19:20
2021-11-17T11:19:19
Because of https://github.com/Python-Markdown/markdown/issues/1196 we have to temporarily pin `markdown` to 3.3.4 for the docs to build without issues
lhoestq
https://github.com/huggingface/datasets/pull/3286
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3286", "html_url": "https://github.com/huggingface/datasets/pull/3286", "diff_url": "https://github.com/huggingface/datasets/pull/3286.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3286.patch", "merged_at": "2021-11-17T11:19:19" }
true
1,055,506,730
3,285
Add IEMOCAP dataset
open
[ "The IEMOCAP dataset is private and available only on request.\r\n```\r\nTo obtain the IEMOCAP data you just need to fill out an electronic release form below.\r\n```\r\n\r\n- [Request form](https://sail.usc.edu/iemocap/release_form.php)\r\n- [License ](https://sail.usc.edu/iemocap/Data_Release_Form_IEMOCAP.pdf)\r\...
2021-11-16T22:47:20
2023-06-10T08:14:52
null
## Adding a Dataset - **Name:** IEMOCAP - **Description:** acted, multimodal and multispeaker database - **Paper:** https://sail.usc.edu/iemocap/Busso_2008_iemocap.pdf - **Data:** https://sail.usc.edu/iemocap/index.html - **Motivation:** Useful multimodal dataset cc @anton-l Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
osanseviero
https://github.com/huggingface/datasets/issues/3285
null
false
1,055,502,909
3,284
Add VoxLingua107 dataset
open
[ "#self-assign" ]
2021-11-16T22:44:08
2021-12-06T09:49:45
null
## Adding a Dataset - **Name:** VoxLingua107 - **Description:** VoxLingua107 is a speech dataset for training spoken language identification models. The dataset consists of short speech segments automatically extracted from YouTube videos and labeled according the language of the video title and description, with some post-processing steps to filter out false positives. - **Paper:** https://arxiv.org/abs/2011.12998 - **Data:** http://bark.phon.ioc.ee/voxlingua107/ - **Motivation:** Nice audio classification dataset cc @anton-l Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
osanseviero
https://github.com/huggingface/datasets/issues/3284
null
false
1,055,495,874
3,283
Add Speech Commands dataset
closed
[ "#self-assign" ]
2021-11-16T22:39:56
2021-12-10T10:30:15
2021-12-10T10:30:15
## Adding a Dataset - **Name:** Speech commands - **Description:** A Dataset for Limited-Vocabulary Speech Recognition - **Paper:** https://arxiv.org/abs/1804.03209 - **Data:** https://www.tensorflow.org/datasets/catalog/speech_commands, Available: http://download.tensorflow.org/data/speech_commands_v0.02.tar.gz - **Motivation:** Nice dataset for audio classification training cc @anton-l Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
osanseviero
https://github.com/huggingface/datasets/issues/3283
null
false
1,055,054,898
3,282
ConnectionError: Couldn't reach https://huggingface.co/datasets/oscar-corpus/OSCAR-2109/resolve/main/OSCAR-2109.py
closed
[ "Hi ! Thanks for reporting :)\r\nI think this is because the dataset is behind an access page. We can fix the dataset viewer\r\n\r\nIf you also have this error when you use the `datasets` library in python, you should probably pass `use_auth_token=True` to the `load_dataset()` function to use your account to access...
2021-11-16T16:05:19
2022-04-12T11:57:43
2022-04-12T11:57:43
## Dataset viewer issue for '*oscar-corpus/OSCAR-2109*' **Link:** *[link to the dataset viewer page](https://huggingface.co/datasets/oscar-corpus/OSCAR-2109)* *The dataset library cannot download any language from the oscar-corpus/OSCAR-2109 dataset. By entering the URL in your browser I can access the file.* ``` raise ConnectionError("Couldn't reach {}".format(url)) ConnectionError: Couldn't reach https://huggingface.co/datasets/oscar-corpus/OSCAR-2109/resolve/main/OSCAR-2109.py ``` Am I the one who added this dataset ? No Using the older version of [OSCAR](https://huggingface.co/datasets/oscar) I don't have any issues downloading languages with the dataset library.
MinionAttack
https://github.com/huggingface/datasets/issues/3282
null
false
1,055,018,876
3,281
[Datasets] Improve Covost 2
closed
[ "I am trying to use `load_dataset` with the French dataset(common voice corpus 1) which is downloaded from a common voice site and the target language is English (using colab)\r\n\r\nSteps I have followed:\r\n\r\n**1. untar:**\r\n`!tar xvzf fr.tar -C data_dir`\r\n\r\n**2. load data:**\r\n`load_dataset('covost2', 'f...
2021-11-16T15:32:19
2022-01-26T16:17:06
2021-11-18T10:44:04
It's currently quite confusing to understand the manual data download instruction of Covost and not very user-friendly. Currenty the user has to: 1. Go on Common Voice website 2. Find the correct dataset which is **not** mentioned in the error message 3. Download it 4. Untar it 5. Create a language id folder (why? this folder does not exist in the `.tar` downloaded file) 6. pass the folder containing the created language id folder This PR improves this to: 1. Go on Common Voice website 2. Find the correct dataset which **is** mentioned in the error message 3. Download it 4. Untar it 5. pass the untared folder **Note**: This PR is not at all time-critical
patrickvonplaten
https://github.com/huggingface/datasets/pull/3281
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3281", "html_url": "https://github.com/huggingface/datasets/pull/3281", "diff_url": "https://github.com/huggingface/datasets/pull/3281.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3281.patch", "merged_at": "2021-11-18T10:44:04" }
true
1,054,766,828
3,280
Fix bookcorpusopen RAM usage
closed
[]
2021-11-16T11:27:52
2021-11-17T15:53:28
2021-11-16T13:34:30
Each document is a full book, so the default arrow writer batch size of 10,000 is too big, and it can fill up RAM quickly before flushing the first batch on disk. I changed its batch size to 256 to use maximum 100MB of memory Fix #3167.
lhoestq
https://github.com/huggingface/datasets/pull/3280
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3280", "html_url": "https://github.com/huggingface/datasets/pull/3280", "diff_url": "https://github.com/huggingface/datasets/pull/3280.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3280.patch", "merged_at": "2021-11-16T13:34:30" }
true
1,054,711,852
3,279
Minor Typo Fix - Precision to Recall
closed
[]
2021-11-16T10:32:22
2021-11-16T11:18:03
2021-11-16T11:18:02
null
SebastinSanty
https://github.com/huggingface/datasets/pull/3279
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3279", "html_url": "https://github.com/huggingface/datasets/pull/3279", "diff_url": "https://github.com/huggingface/datasets/pull/3279.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3279.patch", "merged_at": "2021-11-16T11:18:02" }
true
1,054,249,463
3,278
Proposed update to the documentation for WER
closed
[]
2021-11-15T23:28:31
2021-11-16T11:19:37
2021-11-16T11:19:37
I wanted to submit a minor update to the description of WER for your consideration. Because of the possibility of insertions, the numerator in the WER formula can be larger than N, so the value of WER can be greater than 1.0: ``` >>> from datasets import load_metric >>> metric = load_metric("wer") >>> metric.compute(predictions=["hello how are you"], references=["hello"]) 3.0 ``` and similarly from the underlying jiwer module's `wer` function: ``` >>> from jiwer import wer >>> wer("hello", "hello how are you") 3.0 ```
wooters
https://github.com/huggingface/datasets/pull/3278
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3278", "html_url": "https://github.com/huggingface/datasets/pull/3278", "diff_url": "https://github.com/huggingface/datasets/pull/3278.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3278.patch", "merged_at": "2021-11-16T11:19:37" }
true
1,054,122,656
3,277
f-string formatting
closed
[ "Hello @lhoestq, ```make style``` is applied as asked. :)" ]
2021-11-15T21:37:05
2021-11-19T20:40:08
2021-11-17T16:18:38
**Fix #3257** Replaced _.format()_ and _%_ by f-strings in the following modules : - [x] **tests** - [x] **metrics** - [x] **benchmarks** - [x] **utils** - [x] **templates** - [x] **src/Datasets/\*.py** Modules in **_src/Datasets/_**: - [x] **commands** - [x] **features** - [x] **formatting** - [x] **io** - [x] **tasks** - [x] **utils** Module **datasets** will not be edited as asked by @mariosasko -A correction of the first PR (#3267)-
Mehdi2402
https://github.com/huggingface/datasets/pull/3277
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3277", "html_url": "https://github.com/huggingface/datasets/pull/3277", "diff_url": "https://github.com/huggingface/datasets/pull/3277.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3277.patch", "merged_at": "2021-11-17T16:18:38" }
true
1,053,793,063
3,276
Update KILT metadata JSON
closed
[]
2021-11-15T15:25:25
2021-11-16T11:21:59
2021-11-16T11:21:58
Fix #3265.
albertvillanova
https://github.com/huggingface/datasets/pull/3276
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3276", "html_url": "https://github.com/huggingface/datasets/pull/3276", "diff_url": "https://github.com/huggingface/datasets/pull/3276.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3276.patch", "merged_at": "2021-11-16T11:21:58" }
true
1,053,698,898
3,275
Force data files extraction if download_mode='force_redownload'
closed
[]
2021-11-15T14:00:24
2021-11-15T14:45:23
2021-11-15T14:45:23
Avoids weird issues when redownloading a dataset due to cached data not being fully updated. With this change, issues #3122 and https://github.com/huggingface/datasets/issues/2956 (not a fix, but a workaround) can be fixed as follows: ```python dset = load_dataset(..., download_mode="force_redownload") ```
mariosasko
https://github.com/huggingface/datasets/pull/3275
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3275", "html_url": "https://github.com/huggingface/datasets/pull/3275", "diff_url": "https://github.com/huggingface/datasets/pull/3275.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3275.patch", "merged_at": "2021-11-15T14:45:23" }
true