url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1.83B
node_id
stringlengths
18
32
number
int64
1
6.09k
title
stringlengths
1
290
labels
list
state
stringclasses
2 values
locked
bool
1 class
milestone
dict
comments
int64
0
54
created_at
stringlengths
20
20
updated_at
stringlengths
20
20
closed_at
stringlengths
20
20
active_lock_reason
null
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
state_reason
stringclasses
3 values
draft
bool
2 classes
pull_request
dict
is_pull_request
bool
2 classes
comments_text
list
https://api.github.com/repos/huggingface/datasets/issues/6028
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6028/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6028/comments
https://api.github.com/repos/huggingface/datasets/issues/6028/events
https://github.com/huggingface/datasets/pull/6028
1,803,294,981
PR_kwDODunzps5Vb3LJ
6,028
Use new hffs
[]
closed
false
null
13
2023-07-13T15:41:44Z
2023-07-17T17:09:39Z
2023-07-17T17:01:00Z
null
Thanks to @janineguo 's work in https://github.com/huggingface/datasets/pull/5919 which was needed to support HfFileSystem. Switching to `HfFileSystem` will help implementing optimization in data files resolution ## Implementation details I replaced all the from_hf_repo and from_local_or_remote in data_files.py to only use a new `from_patterns` which works for any fsspec path, including hf:// paths, https:// URLs and local paths. This simplifies the codebase since there is no logic duplication anymore when it comes to data files resolution. I added `_prepare_path_and_storage_options` which returns the right storage_options to use given a path and a `DownloadConfig`. This is the only place where the logic depends on the filesystem type that must be used. I also removed the `get_metadata_data_files_list ` and `get_patterns_and_data_files` functions added recently, since data files resolution is now handled using a common interface. ## New features hf:// paths are now supported in data_files ## Breaking changes DataFilesList and DataFilesDict: - use `str` paths instead of `Union[Path, Url]` - require posix paths for windows paths close https://github.com/huggingface/datasets/issues/6017
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6028/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6028/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/6028.diff", "html_url": "https://github.com/huggingface/datasets/pull/6028", "merged_at": "2023-07-17T17:01:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/6028.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6028" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
https://api.github.com/repos/huggingface/datasets/issues/4551
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4551/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4551/comments
https://api.github.com/repos/huggingface/datasets/issues/4551/events
https://github.com/huggingface/datasets/pull/4551
1,282,534,807
PR_kwDODunzps46QAV-
4,551
Perform hidden file check on relative data file path
[]
closed
false
null
5
2022-06-23T14:49:11Z
2022-06-30T14:49:20Z
2022-06-30T14:38:18Z
null
Fix #4549
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4551/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4551/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4551.diff", "html_url": "https://github.com/huggingface/datasets/pull/4551", "merged_at": "2022-06-30T14:38:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/4551.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4551" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "I'm aware of this behavior, which is tricky to solve due to fsspec's hidden file handling (see https://github.com/huggingface/datasets/issues/4115#issuecomment-1108819538). I've tested some regex patterns to address this, and they se...
https://api.github.com/repos/huggingface/datasets/issues/1269
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1269/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1269/comments
https://api.github.com/repos/huggingface/datasets/issues/1269/events
https://github.com/huggingface/datasets/pull/1269
758,886,174
MDExOlB1bGxSZXF1ZXN0NTMzOTc3MTE2
1,269
Adding OneStopEnglish corpus dataset
[]
closed
false
null
1
2020-12-07T22:05:11Z
2020-12-09T18:43:38Z
2020-12-09T15:33:53Z
null
This PR adds OneStopEnglish Corpus containing texts classified into reading levels (elementary, intermediate, advance) for automatic readability assessment and text simplification. Link to the paper: https://www.aclweb.org/anthology/W18-0535.pdf
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1269/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1269/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1269.diff", "html_url": "https://github.com/huggingface/datasets/pull/1269", "merged_at": "2020-12-09T15:33:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/1269.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1269" }
true
[ "Hi @lhoestq, thanks for the review.\r\nI have made all the changes, PTAL! :) " ]
https://api.github.com/repos/huggingface/datasets/issues/3186
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3186/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3186/comments
https://api.github.com/repos/huggingface/datasets/issues/3186/events
https://github.com/huggingface/datasets/issues/3186
1,040,369,397
I_kwDODunzps4-Asb1
3,186
Dataset viewer for nli_tr
[ { "color": "fef2c0", "default": false, "description": "", "id": 3287858981, "name": "streaming", "node_id": "MDU6TGFiZWwzMjg3ODU4OTgx", "url": "https://api.github.com/repos/huggingface/datasets/labels/streaming" } ]
closed
false
null
6
2021-10-31T03:56:33Z
2022-09-12T09:15:34Z
2022-09-12T08:43:09Z
null
## Dataset viewer issue for '*nli_tr*' **Link:** https://huggingface.co/datasets/nli_tr Hello, Thank you for the new dataset preview feature that will help the users to view the datasets online. We just noticed that the dataset viewer widget in the `nli_tr` dataset shows the error below. The error must be due to a temporary problem that may have blocked access to the dataset through the dataset viewer. But the dataset is currently accessible through the link in the error message. May we kindly ask if it would be possible to rerun the job so that it can access the dataset for the dataset viewer function? Thank you. Emrah ------------------------------------------ Server Error Status code: 404 Exception: FileNotFoundError Message: [Errno 2] No such file or directory: 'zip://snli_tr_1.0_train.jsonl::https://tabilab.cmpe.boun.edu.tr/datasets/nli_datasets/snli_tr_1.0.zip ------------------------------------------ Am I the one who added this dataset ? Yes
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3186/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3186/timeline
null
completed
null
null
false
[ "It's an issue with the streaming mode:\r\n\r\n```python\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('nli_tr', name='snli_tr',split='test', streaming=True)\r\n>>> next(iter(dataset))\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/slesage/hf/datas...
https://api.github.com/repos/huggingface/datasets/issues/4565
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4565/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4565/comments
https://api.github.com/repos/huggingface/datasets/issues/4565/events
https://github.com/huggingface/datasets/issues/4565
1,284,141,666
I_kwDODunzps5MinJi
4,565
Add UFSC OCPap dataset
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
null
1
2022-06-24T20:07:54Z
2022-07-06T19:03:02Z
2022-07-06T19:03:02Z
null
## Adding a Dataset - **Name:** UFSC OCPap: Papanicolaou Stained Oral Cytology Dataset (v4) - **Description:** The UFSC OCPap dataset comprises 9,797 labeled images of 1200x1600 pixels acquired from 5 slides of cancer diagnosed and 3 healthy of oral brush samples, from distinct patients. - **Paper:** https://dx.doi.org/10.2139/ssrn.4119212 - **Data:** https://data.mendeley.com/datasets/dr7ydy9xbk/1 - **Motivation:** real data of pap stained oral cytology samples Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4565/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4565/timeline
null
completed
null
null
false
[ "I will add this directly on the hub (same as #4486)—in https://huggingface.co/lapix" ]
https://api.github.com/repos/huggingface/datasets/issues/5661
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5661/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5661/comments
https://api.github.com/repos/huggingface/datasets/issues/5661/events
https://github.com/huggingface/datasets/issues/5661
1,637,129,445
I_kwDODunzps5hlJzl
5,661
CI is broken: Unnecessary `dict` comprehension
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
0
2023-03-23T09:13:01Z
2023-03-23T09:37:51Z
2023-03-23T09:37:51Z
null
CI check_code_quality is broken: ``` src/datasets/arrow_dataset.py:3267:35: C416 [*] Unnecessary `dict` comprehension (rewrite using `dict()`) Found 1 error. ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5661/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5661/timeline
null
completed
null
null
false
[]
https://api.github.com/repos/huggingface/datasets/issues/4543
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4543/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4543/comments
https://api.github.com/repos/huggingface/datasets/issues/4543/events
https://github.com/huggingface/datasets/pull/4543
1,280,379,781
PR_kwDODunzps46IiEp
4,543
[CI] Fix upstream hub test url
[]
closed
false
null
2
2022-06-22T15:34:27Z
2022-06-22T16:37:40Z
2022-06-22T16:27:37Z
null
Some tests were still using moon-stagign instead of hub-ci. I also updated the token to use one dedicated to `datasets`
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4543/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4543/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4543.diff", "html_url": "https://github.com/huggingface/datasets/pull/4543", "merged_at": "2022-06-22T16:27:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/4543.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4543" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "Remaining CI failures are unrelated to this fix, merging" ]
https://api.github.com/repos/huggingface/datasets/issues/858
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/858/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/858/comments
https://api.github.com/repos/huggingface/datasets/issues/858/events
https://github.com/huggingface/datasets/pull/858
743,904,516
MDExOlB1bGxSZXF1ZXN0NTIxNzE3ODQ4
858
Add SemEval-2010 task 8
[]
closed
false
null
1
2020-11-16T14:57:57Z
2020-11-26T17:28:55Z
2020-11-26T17:28:55Z
null
Hi, I don't know how to add dummy data, since I create the validation set out of the last 1000 examples of the train set. If you have a suggestion, I am happy to implement it. Cheers, Joel
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/858/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/858/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/858.diff", "html_url": "https://github.com/huggingface/datasets/pull/858", "merged_at": "2020-11-26T17:28:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/858.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/858" }
true
[ "Added dummy data and encoding to open(). Now everything should be fine, hopefully :)" ]
https://api.github.com/repos/huggingface/datasets/issues/3882
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3882/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3882/comments
https://api.github.com/repos/huggingface/datasets/issues/3882/events
https://github.com/huggingface/datasets/pull/3882
1,164,595,388
PR_kwDODunzps40NKz7
3,882
Image process doc
[ { "color": "0075ca", "default": true, "description": "Improvements or additions to documentation", "id": 1935892861, "name": "documentation", "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation" } ]
closed
false
null
1
2022-03-10T00:32:10Z
2022-03-15T15:24:16Z
2022-03-15T15:24:09Z
null
This PR is a first draft of how to process image data. It adds: - Load an image dataset with `image` and `path` (adds tip about `decode=False` param to access the path and bytes, thanks to @mariosasko). - Load an image using the `ImageFolder` builder. I know there is an [example](https://huggingface.co/docs/datasets/master/en/loading#image-folders) of this already, but I also wanted to add it here so users don't miss it. This doc seems important for centralizing all of the image-related things so far. Datasets has grown so quickly 🚀 now that I think maybe splitting up the How-to guides by modality may be better since working with vision/audio data is slightly different from what users have seen up until now. This way we can continue to scale the docs to better accommodate vision/audio things. - Add a data augmentation with `set_transform`. There is only 1 example here so far, but we can certainly add more. Todo: - [x] Couldn't figure out why my augmentation function works with `set_transform` but not `map` 🥲. Working with @mariosasko on this!
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 2, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/3882/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3882/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/3882.diff", "html_url": "https://github.com/huggingface/datasets/pull/3882", "merged_at": "2022-03-15T15:24:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/3882.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3882" }
true
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_3882). All of your documentation changes will be reflected on that endpoint." ]
https://api.github.com/repos/huggingface/datasets/issues/1318
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1318/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1318/comments
https://api.github.com/repos/huggingface/datasets/issues/1318/events
https://github.com/huggingface/datasets/pull/1318
759,565,629
MDExOlB1bGxSZXF1ZXN0NTM0NTQ5NjE3
1,318
ethos first commit
[]
closed
false
null
3
2020-12-08T15:59:47Z
2020-12-10T14:45:57Z
2020-12-10T14:45:57Z
null
Ethos passed all the tests except from this one: RUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_configs_<your-dataset-name> with this error: E OSError: Cannot find data file. E Original error: E [Errno 2] No such file or directory:
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1318/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1318/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1318.diff", "html_url": "https://github.com/huggingface/datasets/pull/1318", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1318.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1318" }
true
[ "> Nice thanks !\r\n> \r\n> I left a few comments\r\n> \r\n> Also it looks like this PR includes changes about other files than the ones for ethos\r\n> \r\n> Can you create another branch and another PR please ?\r\n\r\n@lhoestq Should I close this PR? The new one is the: #1453", "You can create another PR and clo...
https://api.github.com/repos/huggingface/datasets/issues/93
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/93/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/93/comments
https://api.github.com/repos/huggingface/datasets/issues/93/events
https://github.com/huggingface/datasets/pull/93
617,522,029
MDExOlB1bGxSZXF1ZXN0NDE3NDIxODUy
93
Cleanup notebooks and various fixes
[]
closed
false
null
0
2020-05-13T14:58:58Z
2020-05-13T15:01:48Z
2020-05-13T15:01:47Z
null
Fixes on dataset (more flexible) metrics (fix) and general clean ups
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/93/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/93/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/93.diff", "html_url": "https://github.com/huggingface/datasets/pull/93", "merged_at": "2020-05-13T15:01:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/93.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/93" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/5942
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5942/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5942/comments
https://api.github.com/repos/huggingface/datasets/issues/5942/events
https://github.com/huggingface/datasets/pull/5942
1,752,021,681
PR_kwDODunzps5Su-V4
5,942
Pass datasets-cli additional args as kwargs to DatasetBuilder in `run_beam.py`
[]
open
false
null
0
2023-06-12T06:50:50Z
2023-06-30T09:15:00Z
null
null
Hi, Following this <https://discuss.huggingface.co/t/how-to-preprocess-a-wikipedia-dataset-using-dataflowrunner/41991/3>, here is a simple PR to pass any additional args to datasets-cli as kwargs in the DatasetBuilder in `run_beam.py`. I also took the liberty to add missing setup steps to the `beam.mdx` docs in order to help everyone. @lhoestq
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5942/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5942/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/5942.diff", "html_url": "https://github.com/huggingface/datasets/pull/5942", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5942.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5942" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/4476
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4476/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4476/comments
https://api.github.com/repos/huggingface/datasets/issues/4476/events
https://github.com/huggingface/datasets/issues/4476
1,267,987,499
I_kwDODunzps5Lk_Qr
4,476
`to_pandas` doesn't take into account format.
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
4
2022-06-10T20:25:31Z
2022-06-15T17:41:41Z
2022-06-15T17:41:41Z
null
**Is your feature request related to a problem? Please describe.** I have a large dataset that I need to convert part of to pandas to do some further analysis. Calling `to_pandas` directly on it is expensive. So I thought I could simply select the columns that I want and then call `to_pandas`. **Describe the solution you'd like** ```python from datasets import Dataset ds = Dataset.from_dict({'a': [1,2,3], 'b': [5,6,7], 'c': [8,9,10]}) pandas_df = ds.with_format(columns=['a', 'b']).to_pandas() # I would expect `pandas_df` to only include a,b as column. ``` **Describe alternatives you've considered** I could remove all columns that I don't want? But I don't know all of them in advance. **Additional context** I can probably make a PR with some pointers.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4476/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4476/timeline
null
completed
null
null
false
[ "Thanks for opening a discussion :)\r\n\r\nNote that you can use `.remove_columns(...)` to keep only the ones you're interested in before calling `.to_pandas()`", "Yes I can do that thank you!\r\n\r\nDo you think that conceptually my example should work? If not, I'm happy to close this issue. \r\n\r\nIf yes, I ca...
https://api.github.com/repos/huggingface/datasets/issues/3457
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3457/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3457/comments
https://api.github.com/repos/huggingface/datasets/issues/3457/events
https://github.com/huggingface/datasets/issues/3457
1,084,862,121
I_kwDODunzps5Aqa6p
3,457
Add CMU Graphics Lab Motion Capture dataset
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" }, { "color": "bfdadc",...
open
false
null
3
2021-12-20T14:34:39Z
2022-03-16T16:53:09Z
null
null
## Adding a Dataset - **Name:** CMU Graphics Lab Motion Capture database - **Description:** The database contains free motions which you can download and use. - **Data:** http://mocap.cs.cmu.edu/ - **Motivation:** Nice motion capture dataset Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3457/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3457/timeline
null
null
null
null
false
[ "This dataset has files in ASF/AMC format. [ The skeleton file is the ASF file (Acclaim Skeleton File). The motion file is the AMC file (Acclaim Motion Capture data). ] \r\n\r\nSome questions : \r\n1. How do we go about representing these features using datasets.Features and generate examples ?\r\n2. The dataset d...
https://api.github.com/repos/huggingface/datasets/issues/2313
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2313/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2313/comments
https://api.github.com/repos/huggingface/datasets/issues/2313/events
https://github.com/huggingface/datasets/pull/2313
875,475,367
MDExOlB1bGxSZXF1ZXN0NjI5ODEwNTc4
2,313
Remove unused head_hf_s3 function
[]
closed
false
null
0
2021-05-04T13:42:06Z
2021-05-07T09:31:42Z
2021-05-07T09:31:42Z
null
Currently, the function `head_hf_s3` is not used: - neither its returned result is used - nor it raises any exception, as exceptions are catched and returned (not raised) This PR removes it.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2313/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2313/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2313.diff", "html_url": "https://github.com/huggingface/datasets/pull/2313", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2313.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2313" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/1933
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1933/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1933/comments
https://api.github.com/repos/huggingface/datasets/issues/1933/events
https://github.com/huggingface/datasets/pull/1933
814,335,846
MDExOlB1bGxSZXF1ZXN0NTc4MzQwMzk3
1,933
Use arrow ipc file format
[]
open
false
null
0
2021-02-23T10:38:24Z
2022-07-06T15:19:48Z
null
null
According to the [documentation](https://arrow.apache.org/docs/format/Columnar.html?highlight=arrow1#ipc-file-format), it's identical to the streaming format except that it contains the memory offsets of each sample: > We define a “file format” supporting random access that is build with the stream format. The file starts and ends with a magic string ARROW1 (plus padding). What follows in the file is identical to the stream format. At the end of the file, we write a footer containing a redundant copy of the schema (which is a part of the streaming format) plus memory offsets and sizes for each of the data blocks in the file. This enables random access any record batch in the file. See File.fbs for the precise details of the file footer. Since it stores more metadata regarding the positions of the examples in the file, it should enable better example retrieval performances. However from the discussion in https://github.com/huggingface/datasets/issues/1803 it looks like it's not the case unfortunately. Maybe in the future this will allow speed gains. I think it's still a good idea to start using it anyway for these reasons: - in the future we may have speed gains - it contains the arrow streaming format data - it's compatible with the pyarrow Dataset implementation (it allows to load remote dataframes for example) if we want to use it in the future - it's also the format used by arrow feather if we want to use it in the future - it's roughly the same size as the streaming format - it's easy to have backward compatibility with the streaming format
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1933/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1933/timeline
null
null
true
{ "diff_url": "https://github.com/huggingface/datasets/pull/1933.diff", "html_url": "https://github.com/huggingface/datasets/pull/1933", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1933.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1933" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/1906
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1906/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1906/comments
https://api.github.com/repos/huggingface/datasets/issues/1906/events
https://github.com/huggingface/datasets/issues/1906
811,405,274
MDU6SXNzdWU4MTE0MDUyNzQ=
1,906
Feature Request: Support for Pandas `Categorical`
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "c5def5", "default": fals...
open
false
null
3
2021-02-18T19:46:05Z
2021-02-23T14:38:50Z
null
null
``` from datasets import Dataset import pandas as pd import pyarrow df = pd.DataFrame(pd.Series(["a", "b", "c", "a"], dtype="category")) pyarrow.Table.from_pandas(df) Dataset.from_pandas(df) # Throws NotImplementedError # TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_table ``` I'm curious if https://github.com/huggingface/datasets/blob/master/src/datasets/features.py#L796 could be built out in a way similar to `Sequence`? e.g. a `Map` class (or whatever name the maintainers might prefer) that can accept: ``` index_type = generate_from_arrow_type(pa_type.index_type) value_type = generate_from_arrow_type(pa_type.value_type) ``` and then additional code points to modify: - FeatureType: https://github.com/huggingface/datasets/blob/master/src/datasets/features.py#L694 - A branch to handle Map in get_nested_type: https://github.com/huggingface/datasets/blob/master/src/datasets/features.py#L719 - I don't quite understand what `encode_nested_example` does but perhaps a branch there? https://github.com/huggingface/datasets/blob/master/src/datasets/features.py#L755 - Similarly, I don't quite understand why `Sequence` is used this way in `generate_from_dict`, but perhaps a branch here? https://github.com/huggingface/datasets/blob/master/src/datasets/features.py#L775 I couldn't find other usages of `Sequence` outside of defining specific datasets, so I'm not sure if that's a comprehensive set of touchpoints.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1906/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1906/timeline
null
null
null
null
false
[ "We already have a ClassLabel type that does this kind of mapping between the label ids (integers) and actual label values (strings).\r\n\r\nI wonder if actually we should use the DictionaryType from Arrow and the Categorical type from pandas for the `datasets` ClassLabel feature type.\r\nCurrently ClassLabel corre...
https://api.github.com/repos/huggingface/datasets/issues/4679
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4679/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4679/comments
https://api.github.com/repos/huggingface/datasets/issues/4679/events
https://github.com/huggingface/datasets/pull/4679
1,303,980,648
PR_kwDODunzps47XX67
4,679
Added method to remove excess nesting in a DatasetDict
[]
closed
false
null
11
2022-07-13T21:49:37Z
2022-07-21T15:55:26Z
2022-07-21T10:55:02Z
null
Added the ability for a DatasetDict to remove additional nested layers within its features to avoid conflicts when collating. It is meant to accompany [this PR](https://github.com/huggingface/transformers/pull/18119) to resolve the same issue [#15505](https://github.com/huggingface/transformers/issues/15505). @stas00 @lhoestq
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4679/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4679/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4679.diff", "html_url": "https://github.com/huggingface/datasets/pull/4679", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4679.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4679" }
true
[ "Hi ! I think the issue you linked is closed and suggests to use `remove_columns`.\r\n\r\nMoreover if you end up with a dataset with an unnecessarily nested data, please modify your processing functions to not output nested data, or use `map(..., batched=True)` if you function take batches as input", "Hi @lhoestq...
https://api.github.com/repos/huggingface/datasets/issues/5458
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5458/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5458/comments
https://api.github.com/repos/huggingface/datasets/issues/5458/events
https://github.com/huggingface/datasets/issues/5458
1,555,054,737
I_kwDODunzps5csECR
5,458
slice split while streaming
[]
closed
false
null
2
2023-01-24T14:08:17Z
2023-01-24T15:11:47Z
2023-01-24T15:11:47Z
null
### Describe the bug When using the `load_dataset` function with streaming set to True, slicing splits is apparently not supported. Did I miss this in the documentation? ### Steps to reproduce the bug `load_dataset("lhoestq/demo1",revision=None, streaming=True, split="train[:3]")` causes ValueError: Bad split: train[:3]. Available splits: ['train', 'test'] in builder.py, line 1213, in as_streaming_dataset ### Expected behavior The first 3 entries of the dataset as a stream ### Environment info - `datasets` version: 2.8.0 - Platform: Windows-10-10.0.19045-SP0 - Python version: 3.10.9 - PyArrow version: 10.0.1 - Pandas version: 1.5.2
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5458/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5458/timeline
null
completed
null
null
false
[ "Hi! Yes, that's correct. When `streaming` is `True`, only split names can be specified as `split`, and for slicing, you have to use `.skip`/`.take` instead.\r\n\r\nE.g. \r\n`load_dataset(\"lhoestq/demo1\",revision=None, streaming=True, split=\"train[:3]\")`\r\n\r\nrewritten with `.skip`/`.take`:\r\n`load_dataset(\...
https://api.github.com/repos/huggingface/datasets/issues/4895
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4895/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4895/comments
https://api.github.com/repos/huggingface/datasets/issues/4895/events
https://github.com/huggingface/datasets/issues/4895
1,350,798,527
I_kwDODunzps5Qg4y_
4,895
load_dataset method returns Unknown split "validation" even if this dir exists
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
17
2022-08-25T12:11:00Z
2022-10-06T17:49:28Z
2022-09-29T08:07:50Z
null
## Describe the bug The `datasets.load_dataset` returns a `ValueError: Unknown split "validation". Should be one of ['train', 'test'].` when running `load_dataset(local_data_dir_path, split="validation")` even if the `validation` sub-directory exists in the local data path. The data directories are as follows and attached to this issue: ``` test_data1 |_ train |_ 1012.png |_ metadata.jsonl ... |_ test ... |_ validation |_ 234.png |_ metadata.jsonl ... test_data2 |_ train |_ train_1012.png |_ metadata.jsonl ... |_ test ... |_ validation |_ val_234.png |_ metadata.jsonl ... ``` They contain the same image files and `metadata.jsonl` but the images in `test_data2` have the split names prepended i.e. `train_1012.png, val_234.png` and the images in `test_data1` do not have the split names prepended to the image names i.e. `1012.png, 234.png` I actually saw in another issue `val` was not recognized as a split name but here I would expect the files to take the split from the parent directory name i.e. val should become part of the validation split? ## Steps to reproduce the bug ```python import datasets datasets.logging.set_verbosity_error() from datasets import load_dataset, get_dataset_split_names # the following only finds train, validation and test splits correctly path = "./test_data1" print("######################", get_dataset_split_names(path), "######################") dataset_list = [] for spt in ["train", "test", "validation"]: dataset = load_dataset(path, split=spt) dataset_list.append(dataset) # the following only finds train and test splits path = "./test_data2" print("######################", get_dataset_split_names(path), "######################") dataset_list = [] for spt in ["train", "test", "validation"]: dataset = load_dataset(path, split=spt) dataset_list.append(dataset) ``` ## Expected results ``` ###################### ['train', 'test', 'validation'] ###################### ###################### ['train', 'test', 'validation'] ###################### ``` ## Actual results ``` Traceback (most recent call last): File "test_data_loader.py", line 11, in <module> dataset = load_dataset(path, split=spt) File "/home/venv/lib/python3.8/site-packages/datasets/load.py", line 1758, in load_dataset ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory) File "/home/venv/lib/python3.8/site-packages/datasets/builder.py", line 893, in as_dataset datasets = map_nested( File "/home/venv/lib/python3.8/site-packages/datasets/utils/py_utils.py", line 385, in map_nested return function(data_struct) File "/home/venv/lib/python3.8/site-packages/datasets/builder.py", line 924, in _build_single_dataset ds = self._as_dataset( File "/home/venv/lib/python3.8/site-packages/datasets/builder.py", line 993, in _as_dataset dataset_kwargs = ArrowReader(self._cache_dir, self.info).read( File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 211, in read files = self.get_file_instructions(name, instructions, split_infos) File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 184, in get_file_instructions file_instructions = make_file_instructions( File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 107, in make_file_instructions absolute_instructions = instruction.to_absolute(name2len) File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 616, in to_absolute return [_rel_to_abs_instr(rel_instr, name2len) for rel_instr in self._relative_instructions] File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 616, in <listcomp> return [_rel_to_abs_instr(rel_instr, name2len) for rel_instr in self._relative_instructions] File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 433, in _rel_to_abs_instr raise ValueError(f'Unknown split "{split}". Should be one of {list(name2len)}.') ValueError: Unknown split "validation". Should be one of ['train', 'test']. ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - Platform: Linux Ubuntu 18.04 - Python version: 3.8.12 - PyArrow version: 9.0.0 Data files [test_data1.zip](https://github.com/huggingface/datasets/files/9424463/test_data1.zip) [test_data2.zip](https://github.com/huggingface/datasets/files/9424468/test_data2.zip)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4895/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4895/timeline
null
completed
null
null
false
[ "I don't know the main problem but it looks like, it is ignoring the last directory in your case. So, create a directory called 'zzz' in the same folder as train, validation and test. if it doesn't work, create a directory called \"aaa\". It worked for me.\r\n", "@SamSamhuns could you please try to load it with t...
https://api.github.com/repos/huggingface/datasets/issues/646
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/646/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/646/comments
https://api.github.com/repos/huggingface/datasets/issues/646/events
https://github.com/huggingface/datasets/pull/646
704,607,371
MDExOlB1bGxSZXF1ZXN0NDg5NTAyMTM3
646
Fix docs typos
[]
closed
false
null
0
2020-09-18T19:32:27Z
2020-09-21T16:30:54Z
2020-09-21T16:14:12Z
null
This PR fixes few typos in the docs and the error in the code snippet in the set_format section in docs/source/torch_tensorflow.rst. `torch.utils.data.Dataloader` expects padded batches so it throws an error due to not being able to stack the unpadded tensors. If we follow the Quick tour from the docs where they add the `truncation=True, padding='max_length'` arguments to the tokenizer before passing data to Dataloader, we can easily fix the issue.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/646/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/646/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/646.diff", "html_url": "https://github.com/huggingface/datasets/pull/646", "merged_at": "2020-09-21T16:14:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/646.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/646" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/1393
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1393/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1393/comments
https://api.github.com/repos/huggingface/datasets/issues/1393/events
https://github.com/huggingface/datasets/pull/1393
760,436,267
MDExOlB1bGxSZXF1ZXN0NTM1MjY4MjUx
1,393
Add script_version suggestion when dataset/metric not found
[]
closed
false
null
0
2020-12-09T15:37:38Z
2020-12-10T18:17:05Z
2020-12-10T18:17:05Z
null
Adds a helpful prompt to the error message when a dataset/metric is not found, suggesting the user might need to pass `script_version="master"` if the dataset was added recently. The whole error looks like: > Couldn't find file locally at blah/blah.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1/metrics/blah/blah.py or https://s3.amazonaws.com/datasets.huggingface.co/datasets/met rics/blah/blah.py. If the dataset was added recently, you may need to to pass script_version="master" to find the loading script on the master branch.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1393/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1393/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1393.diff", "html_url": "https://github.com/huggingface/datasets/pull/1393", "merged_at": "2020-12-10T18:17:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/1393.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1393" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/1701
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1701/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1701/comments
https://api.github.com/repos/huggingface/datasets/issues/1701/events
https://github.com/huggingface/datasets/issues/1701
781,345,717
MDU6SXNzdWU3ODEzNDU3MTc=
1,701
Some datasets miss dataset_infos.json or dummy_data.zip
[]
closed
false
null
2
2021-01-07T14:17:13Z
2022-11-04T15:11:16Z
2022-11-04T15:06:00Z
null
While working on dataset REAME generation script at https://github.com/madlag/datasets_readme_generator , I noticed that some datasets miss a dataset_infos.json : ``` c4 lm1b reclor wikihow ``` And some does not have a dummy_data.zip : ``` kor_nli math_dataset mlqa ms_marco newsgroup qa4mre qangaroo reddit_tifu super_glue trivia_qa web_of_science wmt14 wmt15 wmt16 wmt17 wmt18 wmt19 xtreme ``` But it seems that some of those last do have a "dummy" directory .
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/1701/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1701/timeline
null
completed
null
null
false
[ "Thanks for reporting.\r\nWe should indeed add all the missing dummy_data.zip and also the dataset_infos.json at least for lm1b, reclor and wikihow.\r\n\r\nFor c4 I haven't tested the script and I think we'll require some optimizations regarding beam datasets before processing it.\r\n", "Closing since the dummy d...
https://api.github.com/repos/huggingface/datasets/issues/2319
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2319/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2319/comments
https://api.github.com/repos/huggingface/datasets/issues/2319/events
https://github.com/huggingface/datasets/issues/2319
876,251,376
MDU6SXNzdWU4NzYyNTEzNzY=
2,319
UnicodeDecodeError for OSCAR (Afrikaans)
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
3
2021-05-05T09:22:52Z
2021-05-05T10:57:31Z
2021-05-05T10:50:55Z
null
## Describe the bug When loading the [OSCAR dataset](https://huggingface.co/datasets/oscar) (specifically `unshuffled_deduplicated_af`), I encounter a `UnicodeDecodeError`. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("oscar", "unshuffled_deduplicated_af") ``` ## Expected results Anything but an error, really. ## Actual results ```python >>> from datasets import load_dataset >>> dataset = load_dataset("oscar", "unshuffled_deduplicated_af") Downloading: 14.7kB [00:00, 4.91MB/s] Downloading: 3.07MB [00:00, 32.6MB/s] Downloading and preparing dataset oscar/unshuffled_deduplicated_af (download: 62.93 MiB, generated: 163.38 MiB, post-processed: Unknown size, total: 226.32 MiB) to C:\Users\sgraaf\.cache\huggingface\datasets\oscar\unshuffled_deduplicated_af\1.0.0\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464... Downloading: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 81.0/81.0 [00:00<00:00, 40.5kB/s] Downloading: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 66.0M/66.0M [00:18<00:00, 3.50MB/s] Traceback (most recent call last): File "<stdin>", line 1, in <module> File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\site-packages\datasets\load.py", line 745, in load_dataset builder_instance.download_and_prepare( File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\site-packages\datasets\builder.py", line 574, in download_and_prepare self._download_and_prepare( File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\site-packages\datasets\builder.py", line 652, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\site-packages\datasets\builder.py", line 979, in _prepare_split for key, record in utils.tqdm( File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\site-packages\tqdm\std.py", line 1133, in __iter__ for obj in iterable: File "C:\Users\sgraaf\.cache\huggingface\modules\datasets_modules\datasets\oscar\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464\oscar.py", line 359, in _generate_examples for line in f: File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\encodings\cp1252.py", line 23, in decode return codecs.charmap_decode(input,self.errors,decoding_table)[0] UnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 7454: character maps to <undefined> ``` ## Versions Paste the output of the following code: ```python import datasets import sys import platform print(f""" - Datasets: {datasets.__version__} - Python: {sys.version} - Platform: {platform.platform()} """) ``` - Datasets: 1.6.2 - Python: 3.9.4 (tags/v3.9.4:1f2e308, Apr 6 2021, 13:40:21) [MSC v.1928 64 bit (AMD64)] - Platform: Windows-10-10.0.19041-SP0
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2319/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2319/timeline
null
completed
null
null
false
[ "Thanks for reporting, @sgraaf.\r\n\r\nI am going to have a look at it. \r\n\r\nI guess the expected codec is \"UTF-8\". Normally, when no explicitly codec is passed, Python uses one which is platform-dependent. For Linux machines, the default codec is `utf_8`, which is OK. However for Windows machine, the default ...
https://api.github.com/repos/huggingface/datasets/issues/4240
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4240/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4240/comments
https://api.github.com/repos/huggingface/datasets/issues/4240/events
https://github.com/huggingface/datasets/pull/4240
1,217,287,594
PR_kwDODunzps423xRl
4,240
Fix yield for crd3
[]
closed
false
null
2
2022-04-27T12:31:36Z
2022-04-29T12:41:41Z
2022-04-29T12:41:41Z
null
Modified the `_generate_examples` function to consider all the turns for a chunk id as a single example Modified the features accordingly ``` "turns": [ { "names": datasets.features.Sequence(datasets.Value("string")), "utterances": datasets.features.Sequence(datasets.Value("string")), "number": datasets.Value("int32"), } ], } ``` I wasn't able to run `datasets-cli dummy_data datasets` command. Is there a workaround for this?
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4240/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4240/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4240.diff", "html_url": "https://github.com/huggingface/datasets/pull/4240", "merged_at": "2022-04-29T12:41:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/4240.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4240" }
true
[ "I don't think you need to generate new dummy data, since they're in the same format as the original data.\r\n\r\nThe CI is failing because of this error:\r\n```python\r\n> turn[\"names\"] = turn[\"NAMES\"]\r\nE TypeError: tuple indices must be integers or slices, not str...
https://api.github.com/repos/huggingface/datasets/issues/4309
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4309/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4309/comments
https://api.github.com/repos/huggingface/datasets/issues/4309/events
https://github.com/huggingface/datasets/pull/4309
1,231,232,935
PR_kwDODunzps43lKpm
4,309
[WIP] Add TEDLIUM dataset
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" }, { "color": "d93f0b",...
closed
false
null
11
2022-05-10T14:12:47Z
2022-06-17T12:54:40Z
2022-06-17T11:44:01Z
null
Adds the TED-LIUM dataset https://www.tensorflow.org/datasets/catalog/tedlium#tedliumrelease3 TODO: - [x] Port `tedium.py` from TF datasets using `convert_dataset.sh` script - [x] Make `load_dataset` work - [ ] ~~Run `datasets-cli` command to generate `dataset_infos.json`~~ - [ ] ~~Create dummy data for continuous testing~~ - [ ] ~~Dummy data tests~~ - [ ] ~~Real data tests~~ - [ ] Create the metadata JSON - [ ] Close PR and add directly to the Hub under LIUM org
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4309/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4309/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4309.diff", "html_url": "https://github.com/huggingface/datasets/pull/4309", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4309.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4309" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset('./datasets/tedlium', 'release1', cache_dir='/home/sanchitgandhi/cache')\r\n```\r\n\r\n```\r\nDownloading and preparing dataset tedlium/release1 to /home/sanchit...
https://api.github.com/repos/huggingface/datasets/issues/5243
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5243/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5243/comments
https://api.github.com/repos/huggingface/datasets/issues/5243/events
https://github.com/huggingface/datasets/issues/5243
1,449,523,962
I_kwDODunzps5WZfr6
5,243
Download only split data
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
5
2022-11-15T10:15:54Z
2023-05-02T09:27:51Z
null
null
### Feature request Is it possible to download only the data that I am requesting and not the entire dataset? I run out of disk spaceas it seems to download the entire dataset, instead of only the part needed. common_voice["test"] = load_dataset("mozilla-foundation/common_voice_11_0", "en", split="test", cache_dir="cache/path...", use_auth_token=True, download_config=DownloadConfig(delete_extracted='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') ) ### Motivation efficiency improvement ### Your contribution n/a
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5243/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5243/timeline
null
null
null
null
false
[ "Hi @capsabogdan! Unfortunately, it's hard to implement because quite often datasets data is being hosted in a single archive for all splits :( So we have to download the whole archive to split it into splits. This is the case for CommonVoice too. \r\n\r\nHowever, for cases when data is distributed in separate arch...
https://api.github.com/repos/huggingface/datasets/issues/1015
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1015/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1015/comments
https://api.github.com/repos/huggingface/datasets/issues/1015/events
https://github.com/huggingface/datasets/pull/1015
755,508,841
MDExOlB1bGxSZXF1ZXN0NTMxMjA2MTgy
1,015
add hard dataset
[]
closed
false
null
1
2020-12-02T18:27:36Z
2020-12-03T15:03:54Z
2020-12-03T15:03:54Z
null
Hotel Reviews in Arabic language.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1015/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1015/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1015.diff", "html_url": "https://github.com/huggingface/datasets/pull/1015", "merged_at": "2020-12-03T15:03:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/1015.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1015" }
true
[ "Thanks @sumanthd17 that fixed it. " ]
https://api.github.com/repos/huggingface/datasets/issues/1898
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1898/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1898/comments
https://api.github.com/repos/huggingface/datasets/issues/1898/events
https://github.com/huggingface/datasets/issues/1898
810,157,251
MDU6SXNzdWU4MTAxNTcyNTE=
1,898
ALT dataset has repeating instances in all splits
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
null
4
2021-02-17T12:51:42Z
2021-02-19T06:18:46Z
2021-02-19T06:18:46Z
null
The [ALT](https://huggingface.co/datasets/alt) dataset has all the same instances within each split :/ Seemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits. Would be great if this could be fixed :) Added a snapshot of the contents from `explore-datset` feature, for quick reference. ![image](https://user-images.githubusercontent.com/33179372/108206321-442a2d00-714c-11eb-882f-b4b6e708ef9c.png)
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1898/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1898/timeline
null
completed
null
null
false
[ "Thanks for reporting. This looks like a very bad issue. I'm looking into it", "I just merged a fix, we'll do a patch release soon. Thanks again for reporting, and sorry for the inconvenience.\r\nIn the meantime you can load `ALT` using `datasets` from the master branch", "Thanks!!! works perfectly in the blead...
https://api.github.com/repos/huggingface/datasets/issues/4787
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4787/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4787/comments
https://api.github.com/repos/huggingface/datasets/issues/4787/events
https://github.com/huggingface/datasets/issues/4787
1,328,243,911
I_kwDODunzps5PK2TH
4,787
NonMatchingChecksumError in mbpp dataset
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
0
2022-08-04T08:15:51Z
2022-08-04T17:21:01Z
2022-08-04T17:21:01Z
null
## Describe the bug As reported on the Hub [Fix Checksum Mismatch](https://huggingface.co/datasets/mbpp/discussions/1), there is a `NonMatchingChecksumError` when loading mbpp dataset ## Steps to reproduce the bug ```python ds = load_dataset("mbpp", "full") ``` ## Expected results Loading of the dataset without any exception raised. ## Actual results ``` NonMatchingChecksumError Traceback (most recent call last) <ipython-input-1-a3fbdd3ed82e> in <module> ----> 1 ds = load_dataset("mbpp", "full") .../huggingface/datasets/src/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, **config_kwargs) 1791 1792 # Download and prepare data -> 1793 builder_instance.download_and_prepare( 1794 download_config=download_config, 1795 download_mode=download_mode, .../huggingface/datasets/src/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs) 702 logger.warning("HF google storage unreachable. Downloading and preparing it from source") 703 if not downloaded_from_gcs: --> 704 self._download_and_prepare( 705 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs 706 ) .../huggingface/datasets/src/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos) 1225 1226 def _download_and_prepare(self, dl_manager, verify_infos): -> 1227 super()._download_and_prepare(dl_manager, verify_infos, check_duplicate_keys=verify_infos) 1228 1229 def _get_examples_iterable_for_split(self, split_generator: SplitGenerator) -> ExamplesIterable: .../huggingface/datasets/src/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 773 # Checksums verification 774 if verify_infos and dl_manager.record_checksums: --> 775 verify_checksums( 776 self.info.download_checksums, dl_manager.get_recorded_sizes_checksums(), "dataset source files" 777 ) .../huggingface/datasets/src/datasets/utils/info_utils.py in verify_checksums(expected_checksums, recorded_checksums, verification_name) 38 if len(bad_urls) > 0: 39 error_msg = "Checksums didn't match" + for_verification_name + ":\n" ---> 40 raise NonMatchingChecksumError(error_msg + str(bad_urls)) 41 logger.info("All the checksums matched successfully" + for_verification_name) 42 NonMatchingChecksumError: Checksums didn't match for dataset source files: ['https://raw.githubusercontent.com/google-research/google-research/master/mbpp/mbpp.jsonl'] ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4787/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4787/timeline
null
completed
null
null
false
[]
https://api.github.com/repos/huggingface/datasets/issues/4014
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4014/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4014/comments
https://api.github.com/repos/huggingface/datasets/issues/4014/events
https://github.com/huggingface/datasets/pull/4014
1,180,481,229
PR_kwDODunzps41AGBu
4,014
Support streaming id_clickbait dataset
[]
closed
false
null
1
2022-03-25T08:18:28Z
2022-03-25T08:58:31Z
2022-03-25T08:53:32Z
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4014/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4014/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4014.diff", "html_url": "https://github.com/huggingface/datasets/pull/4014", "merged_at": "2022-03-25T08:53:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/4014.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4014" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/2369
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2369/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2369/comments
https://api.github.com/repos/huggingface/datasets/issues/2369/events
https://github.com/huggingface/datasets/pull/2369
893,554,153
MDExOlB1bGxSZXF1ZXN0NjQ2MDQ5NDM1
2,369
correct labels of conll2003
[]
closed
false
null
0
2021-05-17T17:37:54Z
2021-05-18T08:27:42Z
2021-05-18T08:27:42Z
null
# What does this PR It fixes/extends the `ner_tags` for conll2003 to include all. Paper reference https://arxiv.org/pdf/cs/0306050v1.pdf Model reference https://huggingface.co/elastic/distilbert-base-cased-finetuned-conll03-english/blob/main/config.json
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2369/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2369/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2369.diff", "html_url": "https://github.com/huggingface/datasets/pull/2369", "merged_at": "2021-05-18T08:27:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/2369.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2369" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/2924
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2924/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2924/comments
https://api.github.com/repos/huggingface/datasets/issues/2924/events
https://github.com/huggingface/datasets/issues/2924
997,378,113
I_kwDODunzps47cshB
2,924
"File name too long" error for file locks
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
10
2021-09-15T18:16:50Z
2023-03-28T06:50:18Z
2021-10-29T09:42:24Z
null
## Describe the bug Getting the following error when calling `load_dataset("gar1t/test")`: ``` OSError: [Errno 36] File name too long: '<user>/.cache/huggingface/datasets/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock' ``` ## Steps to reproduce the bug Where the user cache dir (e.g. `~/.cache`) is on a file system that limits filenames to 255 chars (e.g. ext4): ```python from datasets import load_dataset load_dataset("gar1t/test") ``` ## Expected results Expect the function to return without an error. ## Actual results ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "<python_venv>/lib/python3.9/site-packages/datasets/load.py", line 1112, in load_dataset builder_instance.download_and_prepare( File "<python_venv>/lib/python3.9/site-packages/datasets/builder.py", line 644, in download_and_prepare self._save_info() File "<python_venv>/lib/python3.9/site-packages/datasets/builder.py", line 765, in _save_info with FileLock(lock_path): File "<python_venv>/lib/python3.9/site-packages/datasets/utils/filelock.py", line 323, in __enter__ self.acquire() File "<python_venv>/lib/python3.9/site-packages/datasets/utils/filelock.py", line 272, in acquire self._acquire() File "<python_venv>/lib/python3.9/site-packages/datasets/utils/filelock.py", line 403, in _acquire fd = os.open(self._lock_file, open_mode) OSError: [Errno 36] File name too long: '<user>/.cache/huggingface/datasets/_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock' ``` ## Environment info - `datasets` version: 1.12.1 - Platform: Linux-5.11.0-27-generic-x86_64-with-glibc2.31 - Python version: 3.9.7 - PyArrow version: 5.0.0
{ "+1": 0, "-1": 0, "confused": 1, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2924/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2924/timeline
null
completed
null
null
false
[ "Hi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_datasets_csv_test-7c856aea083a7043_0.0.0_9144e0a4e8435090117cea53e6c7537173ef2304525df4a077c435d8ee7828ff.incomplete.lock\")\r\n154\r\n```\r\nso not sure why it's considered too long for your filesystem.\r\n(also note...
https://api.github.com/repos/huggingface/datasets/issues/4254
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4254/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4254/comments
https://api.github.com/repos/huggingface/datasets/issues/4254/events
https://github.com/huggingface/datasets/pull/4254
1,220,204,395
PR_kwDODunzps43Bwnj
4,254
Replace data URL in SAMSum dataset and support streaming
[]
closed
false
null
1
2022-04-29T08:21:43Z
2022-05-06T08:38:16Z
2022-04-29T16:26:09Z
null
This PR replaces data URL in SAMSum dataset: - original host (arxiv.org) does not allow HTTP Range requests - we have hosted the data on the Hub (license: CC BY-NC-ND 4.0) Moreover, it implements support for streaming. Fix #4146. Related to: #4236. CC: @severo
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4254/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4254/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4254.diff", "html_url": "https://github.com/huggingface/datasets/pull/4254", "merged_at": "2022-04-29T16:26:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/4254.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4254" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/4935
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4935/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4935/comments
https://api.github.com/repos/huggingface/datasets/issues/4935/events
https://github.com/huggingface/datasets/issues/4935
1,363,226,736
I_kwDODunzps5RQTBw
4,935
Dataset Viewer issue for ubuntu_dialogs_corpus
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
closed
false
null
1
2022-09-06T12:41:50Z
2022-09-06T12:51:25Z
2022-09-06T12:51:25Z
null
### Link _No response_ ### Description _No response_ ### Owner _No response_
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4935/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4935/timeline
null
completed
null
null
false
[ "The dataset maintainers (https://huggingface.co/datasets/ubuntu_dialogs_corpus) decided to forbid the dataset from being downloaded automatically (https://huggingface.co/docs/datasets/v2.4.0/en/loading#manual-download), and the dataset viewer respects this.\r\nWe will try to improve the error display though. Thank...
https://api.github.com/repos/huggingface/datasets/issues/3785
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3785/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3785/comments
https://api.github.com/repos/huggingface/datasets/issues/3785/events
https://github.com/huggingface/datasets/pull/3785
1,150,069,801
PR_kwDODunzps4zciES
3,785
Fix: Bypass Virus Checks in Google Drive Links (CNN-DM dataset)
[]
closed
false
null
8
2022-02-25T05:48:57Z
2022-03-03T16:43:47Z
2022-03-03T14:03:37Z
null
This commit fixes the issue described in #3784. By adding an extra parameter to the end of Google Drive links, we are able to bypass the virus check and download the datasets. So, if the original link looked like https://drive.google.com/uc?export=download&id=0BwmD_VLjROrfTHk4NFg2SndKcjQ The new link now looks like https://drive.google.com/uc?export=download&id=0BwmD_VLjROrfTHk4NFg2SndKcjQ&confirm=t Fixes #3784
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3785/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3785/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/3785.diff", "html_url": "https://github.com/huggingface/datasets/pull/3785", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/3785.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3785" }
true
[ "Thank you, @albertvillanova!", "Got it. Thanks for explaining this, @albertvillanova!\r\n\r\n> On the other hand, the tests are not passing because the dummy data should also be fixed. Once done, this PR will be able to be merged into master.\r\n\r\nWill do this 👍", "Hi ! I think we need to fix the issue for ...
https://api.github.com/repos/huggingface/datasets/issues/2630
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2630/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2630/comments
https://api.github.com/repos/huggingface/datasets/issues/2630/events
https://github.com/huggingface/datasets/issues/2630
942,102,956
MDU6SXNzdWU5NDIxMDI5NTY=
2,630
Progress bars are not properly rendered in Jupyter notebook
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
2
2021-07-12T14:07:13Z
2022-02-03T15:55:33Z
2022-02-03T15:55:33Z
null
## Describe the bug The progress bars are not Jupyter widgets; regular progress bars appear (like in a terminal). ## Steps to reproduce the bug ```python ds.map(tokenize, num_proc=10) ``` ## Expected results Jupyter widgets displaying the progress bars. ## Actual results Simple plane progress bars. cc: Reported by @thomwolf
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2630/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2630/timeline
null
completed
null
null
false
[ "To add my experience when trying to debug this issue:\r\n\r\nSeems like previously the workaround given [here](https://github.com/tqdm/tqdm/issues/485#issuecomment-473338308) worked around this issue. But with the latest version of jupyter/tqdm I still get terminal warnings that IPython tried to send a message fro...
https://api.github.com/repos/huggingface/datasets/issues/4083
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4083/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4083/comments
https://api.github.com/repos/huggingface/datasets/issues/4083/events
https://github.com/huggingface/datasets/pull/4083
1,190,025,878
PR_kwDODunzps41gEbu
4,083
Add SacreBLEU Metric Card
[]
closed
false
null
1
2022-04-01T16:24:56Z
2022-04-12T20:45:00Z
2022-04-12T20:38:40Z
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4083/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4083/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4083.diff", "html_url": "https://github.com/huggingface/datasets/pull/4083", "merged_at": "2022-04-12T20:38:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/4083.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4083" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/4853
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4853/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4853/comments
https://api.github.com/repos/huggingface/datasets/issues/4853/events
https://github.com/huggingface/datasets/pull/4853
1,339,456,490
PR_kwDODunzps49NFNL
4,853
Fix bug and checksums in exams dataset
[]
closed
false
null
1
2022-08-15T20:17:57Z
2022-08-16T06:43:57Z
2022-08-16T06:29:06Z
null
Fix #4852.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4853/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4853/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4853.diff", "html_url": "https://github.com/huggingface/datasets/pull/4853", "merged_at": "2022-08-16T06:29:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/4853.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4853" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
https://api.github.com/repos/huggingface/datasets/issues/2075
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2075/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2075/comments
https://api.github.com/repos/huggingface/datasets/issues/2075/events
https://github.com/huggingface/datasets/issues/2075
834,301,246
MDU6SXNzdWU4MzQzMDEyNDY=
2,075
ConnectionError: Couldn't reach common_voice.py
[]
closed
false
null
2
2021-03-18T01:19:06Z
2021-03-20T10:29:41Z
2021-03-20T10:29:41Z
null
When I run: from datasets import load_dataset, load_metric common_voice_train = load_dataset("common_voice", "zh-CN", split="train+validation") common_voice_test = load_dataset("common_voice", "zh-CN", split="test") Got: ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/master/datasets/common_voice/common_voice.py Version: 1.4.1 Thanks! @lhoestq @LysandreJik @thomwolf
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2075/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2075/timeline
null
completed
null
null
false
[ "Hi @LifaSun, thanks for reporting this issue.\r\n\r\nSometimes, GitHub has some connectivity problems. Could you confirm that the problem persists?", "@albertvillanova Thanks! It works well now. " ]
https://api.github.com/repos/huggingface/datasets/issues/1421
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1421/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1421/comments
https://api.github.com/repos/huggingface/datasets/issues/1421/events
https://github.com/huggingface/datasets/pull/1421
760,706,851
MDExOlB1bGxSZXF1ZXN0NTM1NDkzMzU4
1,421
adding fake-news-english-2
[]
closed
false
null
0
2020-12-09T22:05:13Z
2020-12-13T00:48:49Z
2020-12-13T00:48:49Z
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1421/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1421/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1421.diff", "html_url": "https://github.com/huggingface/datasets/pull/1421", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1421.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1421" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/4092
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4092/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4092/comments
https://api.github.com/repos/huggingface/datasets/issues/4092/events
https://github.com/huggingface/datasets/pull/4092
1,192,499,903
PR_kwDODunzps41n40R
4,092
Fix dataset `amazon_us_reviews` metadata - 4/4/2022
[]
closed
false
null
2
2022-04-05T01:39:45Z
2022-04-08T12:35:41Z
2022-04-08T12:29:31Z
null
Fixes #4048 by running `dataset-cli test` to reprocess data and regenerate metadata. Additionally I've updated the README to include up-to-date counts for the subsets.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4092/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4092/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4092.diff", "html_url": "https://github.com/huggingface/datasets/pull/4092", "merged_at": "2022-04-08T12:29:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/4092.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4092" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "cc: @albertvillanova just FYI" ]
https://api.github.com/repos/huggingface/datasets/issues/1411
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1411/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1411/comments
https://api.github.com/repos/huggingface/datasets/issues/1411/events
https://github.com/huggingface/datasets/pull/1411
760,606,290
MDExOlB1bGxSZXF1ZXN0NTM1NDEwNjU3
1,411
2 typos
[]
closed
false
null
0
2020-12-09T19:24:34Z
2020-12-11T10:39:05Z
2020-12-11T10:39:05Z
null
Corrected 2 typos
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1411/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1411/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1411.diff", "html_url": "https://github.com/huggingface/datasets/pull/1411", "merged_at": "2020-12-11T10:39:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/1411.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1411" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/4453
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4453/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4453/comments
https://api.github.com/repos/huggingface/datasets/issues/4453/events
https://github.com/huggingface/datasets/issues/4453
1,262,674,105
I_kwDODunzps5LQuC5
4,453
Dataset Viewer issue for Yaxin/SemEval2015
[]
closed
false
null
3
2022-06-07T03:30:08Z
2022-06-09T08:34:16Z
2022-06-09T08:34:16Z
null
### Link _No response_ ### Description _No response_ ### Owner _No response_
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4453/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4453/timeline
null
completed
null
null
false
[ "I understand that the issue is that a remote file (URL) is being loaded as a local file. Right @albertvillanova @lhoestq?\r\n\r\n```\r\nMessage: [Errno 2] No such file or directory: 'https://raw.githubusercontent.com/YaxinCui/ABSADataset/main/SemEval2015Task12Corrected/train/restaurants_train.xml'\r\n```", ...
https://api.github.com/repos/huggingface/datasets/issues/4460
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/4460/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/4460/comments
https://api.github.com/repos/huggingface/datasets/issues/4460/events
https://github.com/huggingface/datasets/pull/4460
1,264,644,205
PR_kwDODunzps45UHIs
4,460
Drop Python 3.6 support
[]
closed
false
null
5
2022-06-08T12:10:18Z
2022-07-26T19:16:39Z
2022-07-26T19:04:21Z
null
Remove the fallback imports/checks in the code needed for Python 3.6 and update the requirements/CI files. Also, use Python types for the NumPy dtype wherever possible to avoid deprecation warnings in newer NumPy versions.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4460/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/4460/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/4460.diff", "html_url": "https://github.com/huggingface/datasets/pull/4460", "merged_at": "2022-07-26T19:04:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/4460.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4460" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "I've disabled the `test_dummy_dataset_serialize_s3` tests in the Linux CI to avoid the failures (these tests only fail on Windows in 3.6). These failures are unrelated to this PR's changes, and I would like to address this in a new P...
https://api.github.com/repos/huggingface/datasets/issues/1808
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1808/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1808/comments
https://api.github.com/repos/huggingface/datasets/issues/1808/events
https://github.com/huggingface/datasets/issues/1808
798,879,180
MDU6SXNzdWU3OTg4NzkxODA=
1,808
writing Datasets in a human readable format
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "d876e3", "default": true...
closed
false
null
3
2021-02-02T02:55:40Z
2022-06-01T15:38:13Z
2022-06-01T15:38:13Z
null
Hi I see there is a save_to_disk function to save data, but this is not human readable format, is there a way I could save a Dataset object in a human readable format to a file like json? thanks @lhoestq
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1808/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1808/timeline
null
completed
null
null
false
[ "AFAIK, there is currently no built-in method on the `Dataset` object to do this.\r\nHowever, a workaround is to directly use the Arrow table backing the dataset, **but it implies loading the whole dataset in memory** (correct me if I'm mistaken @lhoestq).\r\n\r\nYou can convert the Arrow table to a pandas datafram...
https://api.github.com/repos/huggingface/datasets/issues/99
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/99/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/99/comments
https://api.github.com/repos/huggingface/datasets/issues/99/events
https://github.com/huggingface/datasets/pull/99
618,026,700
MDExOlB1bGxSZXF1ZXN0NDE3ODMxNjky
99
[Cmrc 2018] fix cmrc2018
[]
closed
false
null
0
2020-05-14T08:22:03Z
2020-05-14T08:49:42Z
2020-05-14T08:49:41Z
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/99/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/99/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/99.diff", "html_url": "https://github.com/huggingface/datasets/pull/99", "merged_at": "2020-05-14T08:49:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/99.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/99" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/3427
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3427/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3427/comments
https://api.github.com/repos/huggingface/datasets/issues/3427/events
https://github.com/huggingface/datasets/pull/3427
1,078,782,159
PR_kwDODunzps4vxb_y
3,427
Add The Pile Enron Emails subset
[]
closed
false
null
0
2021-12-13T17:14:16Z
2021-12-14T17:30:59Z
2021-12-14T17:30:57Z
null
Add: - Enron Emails subset of The Pile: "enron_emails" config Close bigscience-workshop/data_tooling#310. CC: @StellaAthena
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3427/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3427/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/3427.diff", "html_url": "https://github.com/huggingface/datasets/pull/3427", "merged_at": "2021-12-14T17:30:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/3427.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3427" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/5714
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5714/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5714/comments
https://api.github.com/repos/huggingface/datasets/issues/5714/events
https://github.com/huggingface/datasets/pull/5714
1,657,388,033
PR_kwDODunzps5NxIOc
5,714
Fix xnumpy_load for .npz files
[]
closed
false
null
2
2023-04-06T13:01:45Z
2023-04-07T09:23:54Z
2023-04-07T09:16:57Z
null
PR: - #5626 implemented support for streaming `.npy` files by using `numpy.load`. However, it introduced a bug when used with `.npz` files, within a context manager: ``` ValueError: seek of closed file ``` or in streaming mode: ``` ValueError: I/O operation on closed file. ``` This PR fixes the bug and tests for both `.npy` and `.npz` files. Fix #5711.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5714/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5714/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/5714.diff", "html_url": "https://github.com/huggingface/datasets/pull/5714", "merged_at": "2023-04-07T09:16:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/5714.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5714" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
https://api.github.com/repos/huggingface/datasets/issues/147
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/147/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/147/comments
https://api.github.com/repos/huggingface/datasets/issues/147/events
https://github.com/huggingface/datasets/issues/147
619,581,907
MDU6SXNzdWU2MTk1ODE5MDc=
147
Error with sklearn train_test_split
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
2
2020-05-17T00:28:24Z
2020-06-18T16:23:23Z
2020-06-18T16:23:23Z
null
It would be nice if we could use sklearn `train_test_split` to quickly generate subsets from the dataset objects returned by `nlp.load_dataset`. At the moment the code: ```python data = nlp.load_dataset('imdb', cache_dir=data_cache) f_half, s_half = train_test_split(data['train'], test_size=0.5, random_state=seed) ``` throws: ``` ValueError: Can only get row(s) (int or slice) or columns (string). ``` It's not a big deal, since there are other ways to split the data, but it would be a cool thing to have.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/147/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/147/timeline
null
completed
null
null
false
[ "Indeed. Probably we will want to have a similar method directly in the library", "Related: #166 " ]
https://api.github.com/repos/huggingface/datasets/issues/2122
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2122/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2122/comments
https://api.github.com/repos/huggingface/datasets/issues/2122/events
https://github.com/huggingface/datasets/pull/2122
842,194,588
MDExOlB1bGxSZXF1ZXN0NjAxODE3MjI0
2,122
Fast table queries with interpolation search
[]
closed
false
null
0
2021-03-26T18:09:20Z
2021-08-04T18:11:59Z
2021-04-06T14:33:01Z
null
## Intro This should fix issue #1803 Currently querying examples in a dataset is O(n) because of the underlying pyarrow ChunkedArrays implementation. To fix this I implemented interpolation search that is pretty effective since datasets usually verifies the condition of evenly distributed chunks (the default chunk size is fixed). ## Benchmark Here is a [benchmark](https://pastebin.com/utEXUqsR) I did on bookcorpus (74M rows): for the current implementation ```python >>> python speed.py Loaded dataset 'bookcorpus', len=74004228, nbytes=4835358766 ========================= Querying unshuffled bookcorpus ========================= Avg access time key=1 : 0.018ms Avg access time key=74004227 : 0.215ms Avg access time key=range(74003204, 74004228) : 1.416ms Avg access time key=RandIter(low=0, high=74004228, size=1024, seed=42): 92.532ms ========================== Querying shuffled bookcorpus ========================== Avg access time key=1 : 0.187ms Avg access time key=74004227 : 6.642ms Avg access time key=range(74003204, 74004228) : 90.941ms Avg access time key=RandIter(low=0, high=74004228, size=1024, seed=42): 3448.456ms ``` for the new one using interpolation search: ```python >>> python speed.py Loaded dataset 'bookcorpus', len=74004228, nbytes=4835358766 ========================= Querying unshuffled bookcorpus ========================= Avg access time key=1 : 0.076ms Avg access time key=74004227 : 0.056ms Avg access time key=range(74003204, 74004228) : 1.807ms Avg access time key=RandIter(low=0, high=74004228, size=1024, seed=42): 24.028ms ========================== Querying shuffled bookcorpus ========================== Avg access time key=1 : 0.061ms Avg access time key=74004227 : 0.058ms Avg access time key=range(74003204, 74004228) : 22.166ms Avg access time key=RandIter(low=0, high=74004228, size=1024, seed=42): 42.757ms ``` The RandIter class is just an iterable of 1024 random indices from 0 to 74004228. Here is also a plot showing the speed improvement depending on the dataset size: ![image](https://user-images.githubusercontent.com/42851186/112673587-32335c80-8e65-11eb-9a0c-58ad774abaec.png) ## Implementation details: - `datasets.table.Table` objects implement interpolation search for the `slice` method - The interpolation search requires to store the offsets of all the chunks of a table. The offsets are stored when the `Table` is initialized. - `datasets.table.Table.slice` returns a `datasets.table.Table` using interpolation search - `datasets.table.Table.fast_slice` returns a `pyarrow.Table` object using interpolation search. This is useful to get a part of a dataset if we don't need the indexing structure for future computations. For example it's used when querying an example as a dictionary. - Now a `Dataset` object is always backed by a `datasets.table.Table` object. If one passes a `pyarrow.Table` to initialize a `Dataset`, then it's converted to a `datasets.table.Table` ## Checklist: - [x] implement interpolation search - [x] use `datasets.table.Table` in `Dataset` objects - [x] update current tests - [x] add tests for interpolation search - [x] comments and docstring - [x] add the benchmark to the CI Fix #1803.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 5, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 5, "url": "https://api.github.com/repos/huggingface/datasets/issues/2122/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2122/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2122.diff", "html_url": "https://github.com/huggingface/datasets/pull/2122", "merged_at": "2021-04-06T14:33:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/2122.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2122" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/2939
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2939/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2939/comments
https://api.github.com/repos/huggingface/datasets/issues/2939/events
https://github.com/huggingface/datasets/pull/2939
999,639,630
PR_kwDODunzps4r58Gu
2,939
MENYO-20k repo has moved, updating URL
[]
closed
false
null
0
2021-09-17T19:01:54Z
2021-09-21T15:31:37Z
2021-09-21T15:31:36Z
null
Dataset repo moved to https://github.com/uds-lsv/menyo-20k_MT, now editing URL to match. https://github.com/uds-lsv/menyo-20k_MT/blob/master/data/train.tsv is the file we're looking for
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2939/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2939/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/2939.diff", "html_url": "https://github.com/huggingface/datasets/pull/2939", "merged_at": "2021-09-21T15:31:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/2939.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2939" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/344
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/344/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/344/comments
https://api.github.com/repos/huggingface/datasets/issues/344/events
https://github.com/huggingface/datasets/pull/344
651,495,246
MDExOlB1bGxSZXF1ZXN0NDQ0NzQwMTIw
344
Search qa
[]
closed
false
null
1
2020-07-06T12:23:16Z
2020-07-16T08:58:16Z
2020-07-16T08:58:16Z
null
This PR adds the Search QA dataset used in **SearchQA: A New Q&A Dataset Augmented with Context from a Search Engine**. The dataset has the following config name: - raw_jeopardy: raw data - train_test_val: which is the splitted version #336
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/344/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/344/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/344.diff", "html_url": "https://github.com/huggingface/datasets/pull/344", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/344.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/344" }
true
[ "Could you rebase from master just to make sure we won't break anything for `fever` pls @mariamabarham ?" ]
https://api.github.com/repos/huggingface/datasets/issues/5075
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5075/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5075/comments
https://api.github.com/repos/huggingface/datasets/issues/5075/events
https://github.com/huggingface/datasets/issues/5075
1,397,865,501
I_kwDODunzps5TUbwd
5,075
Throw EnvironmentError when token is not present
[ { "color": "7057ff", "default": true, "description": "Good for newcomers", "id": 1935892877, "name": "good first issue", "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue" }, { "color": "DF8D62", "defa...
closed
false
null
1
2022-10-05T14:14:18Z
2022-10-07T14:33:28Z
2022-10-07T14:33:28Z
null
Throw EnvironmentError instead of OSError ([link](https://github.com/huggingface/datasets/blob/6ad430ba0cdeeb601170f732d4bd977f5c04594d/src/datasets/arrow_dataset.py#L4306) to the line) in `push_to_hub` when the Hub token is not present.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5075/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5075/timeline
null
completed
null
null
false
[ "@mariosasko I've raised a PR #5076 against this issue. Please help to review. Thanks." ]
https://api.github.com/repos/huggingface/datasets/issues/6076
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/6076/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/6076/comments
https://api.github.com/repos/huggingface/datasets/issues/6076/events
https://github.com/huggingface/datasets/pull/6076
1,822,345,597
PR_kwDODunzps5WcGVR
6,076
No gzip encoding from github
[]
closed
false
null
3
2023-07-26T12:46:07Z
2023-07-27T16:15:11Z
2023-07-27T16:14:40Z
null
Don't accept gzip encoding from github, otherwise some files are not streamable + seekable. fix https://huggingface.co/datasets/code_x_glue_cc_code_to_code_trans/discussions/2#64c0e0c1a04a514ba6303e84 and making sure https://github.com/huggingface/datasets/issues/2918 works as well
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/6076/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/6076/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/6076.diff", "html_url": "https://github.com/huggingface/datasets/pull/6076", "merged_at": "2023-07-27T16:14:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/6076.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/6076" }
true
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
https://api.github.com/repos/huggingface/datasets/issues/740
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/740/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/740/comments
https://api.github.com/repos/huggingface/datasets/issues/740/events
https://github.com/huggingface/datasets/pull/740
723,047,958
MDExOlB1bGxSZXF1ZXN0NTA0NzAyNTc0
740
Fix TREC urls
[]
closed
false
null
0
2020-10-16T09:11:28Z
2020-10-19T08:54:37Z
2020-10-19T08:54:36Z
null
The old TREC urls are now redirections. I updated the urls to the new ones, since we don't support redirections for downloads. Fix #737
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/740/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/740/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/740.diff", "html_url": "https://github.com/huggingface/datasets/pull/740", "merged_at": "2020-10-19T08:54:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/740.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/740" }
true
[]
https://api.github.com/repos/huggingface/datasets/issues/3071
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3071/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3071/comments
https://api.github.com/repos/huggingface/datasets/issues/3071/events
https://github.com/huggingface/datasets/issues/3071
1,024,893,493
I_kwDODunzps49FqI1
3,071
Custom plain text dataset, plain json dataset and plain csv dataset are remove from datasets template folder
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
null
1
2021-10-13T07:32:10Z
2021-10-13T08:27:04Z
2021-10-13T08:27:03Z
null
## Adding a Dataset - **Name:** text, json, csv - **Description:** I am developing a customized dataset loading script. The problem is mainly about my custom dataset is seperate into many files and I only find a dataset loading template in [https://github.com/huggingface/datasets/blob/1.2.1/datasets/json/json.py](https://github.com/huggingface/datasets/blob/1.2.1/datasets/json/json.py) that can handle my circumstance. I'm afraid these templates are too old to use. Could you re-add these three templates to current master branch?
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3071/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3071/timeline
null
completed
null
null
false
[ "Hi @zixiliuUSC, \r\n\r\nAs explained in the documentation (https://huggingface.co/docs/datasets/loading.html#json), we support loading any dataset in JSON (as well as CSV, text, Parquet) format:\r\n```python\r\nds = load_dataset('json', data_files='my_file.json')\r\n```" ]
https://api.github.com/repos/huggingface/datasets/issues/5158
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/5158/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/5158/comments
https://api.github.com/repos/huggingface/datasets/issues/5158/events
https://github.com/huggingface/datasets/issues/5158
1,422,059,287
I_kwDODunzps5UwucX
5,158
Fix language and license tag names in all Hub datasets
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
closed
false
null
6
2022-10-25T08:19:29Z
2022-10-25T11:27:26Z
2022-10-25T10:42:19Z
null
While working on this: - #5137 we realized there are still many datasets with deprecated "languages" and "licenses" tag names (instead of "language" and "license"). This is a blocking issue: no subsequent PR can be opened to modify their metadata: a ValueError will be thrown. We should fix the "language" and "license" tag names in all Hub datasets. TODO: - [x] Fix language and license tag names in 402 Hub datasets CC: @julien-c
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5158/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/5158/timeline
null
completed
null
null
false
[ "There are currently 402 datasets with deprecated \"languages\" or \"licenses\".", "hey @albertvillanova ,i would love to work on this issue if you like.", "Hi @ayushthe1, thanks for your offer.\r\n\r\nBut as you can see, I self-assigned this issue.\r\n\r\nI have already fixed 200 out of the 402 datasets. My sc...
https://api.github.com/repos/huggingface/datasets/issues/2175
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2175/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2175/comments
https://api.github.com/repos/huggingface/datasets/issues/2175/events
https://github.com/huggingface/datasets/issues/2175
851,836,096
MDU6SXNzdWU4NTE4MzYwOTY=
2,175
dataset.search_batch() function outputs all -1 indices sometime.
[]
closed
false
null
6
2021-04-06T21:50:49Z
2021-04-16T12:21:16Z
2021-04-16T12:21:15Z
null
I am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, "IVF65536_HNSW32,Flat")**. During the retrieval phase exactly in [this line of retrieval_rag.py](https://github.com/huggingface/transformers/blob/master/src/transformers/models/rag/retrieval_rag.py#L231) an error issue when all retrieved indices are -1. Please refer to the screenshot of a PID worker. ![image](https://user-images.githubusercontent.com/16892570/113782387-37a67600-9786-11eb-9c29-acad661a9648.png) Here, my retrieve batch size is 2 and n_docs is 5. I can solve this by working around np. stack, but I want to ask, why we get an output index of -1. Do you have any idea :) ? Is this a problem of the index, where the faiss can't find any similar vector? Is there documentation on the output index being -1? @lhoestq
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2175/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/2175/timeline
null
completed
null
null
false
[ "Actually, I found the answer [here](https://github.com/facebookresearch/faiss/wiki/FAQ#what-does-it-mean-when-a-search-returns--1-ids). \r\n\r\nSo we have to do some modifications to the code for instances where the index doesn't retrieve any IDs.", "@lhoestq @patrickvonplaten \r\n\r\nI also found another short...
https://api.github.com/repos/huggingface/datasets/issues/835
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/835/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/835/comments
https://api.github.com/repos/huggingface/datasets/issues/835/events
https://github.com/huggingface/datasets/issues/835
740,102,210
MDU6SXNzdWU3NDAxMDIyMTA=
835
Wikipedia postprocessing
[]
closed
false
null
3
2020-11-10T17:26:38Z
2020-11-10T18:23:20Z
2020-11-10T17:49:21Z
null
Hi, thanks for this library! Running this code: ```py import datasets wikipedia = datasets.load_dataset("wikipedia", "20200501.de") print(wikipedia['train']['text'][0]) ``` I get: ``` mini|Ricardo Flores Magón mini|Mexikanische Revolutionäre, Magón in der Mitte anführend, gegen die Diktatur von Porfirio Diaz, Ausschnitt des Gemälde „Tierra y Libertad“ von Idelfonso Carrara (?) von 1930. Ricardo Flores Magón (* 16. September 1874 in San Antonio Eloxochitlán im mexikanischen Bundesstaat Oaxaca; † 22. November 1922 im Bundesgefängnis Leavenworth im US-amerikanischen Bundesstaat Kansas) war als Journalist, Gewerkschafter und Literat ein führender anarchistischer Theoretiker und Aktivist, der die revolutionäre mexikanische Bewegung radikal beeinflusste. Magón war Gründer der Partido Liberal Mexicano und Mitglied der Industrial Workers of the World. Politische Biografie Journalistisch und politisch kämpfte er und sein Bruder sehr kompromisslos gegen die Diktatur Porfirio Diaz. Philosophisch und politisch orientiert an radikal anarchistischen Idealen und den Erfahrungen seiner indigenen Vorfahren bei der gemeinschaftlichen Bewirtschaftung des Gemeindelandes, machte er die Forderung „Land und Freiheit“ (Tierra y Libertad) populär. Besonders Francisco Villa und Emiliano Zapata griffen die Forderung Land und Freiheit auf. Seine Philosophie hatte großen Einfluss auf die Landarbeiter. 1904 floh er in die USA und gründete 1906 die Partido Liberal Mexicano. Im Exil lernte er u. a. Emma Goldman kennen. Er verbrachte die meiste Zeit seines Lebens in Gefängnissen und im Exil und wurde 1918 in den USA wegen „Behinderung der Kriegsanstrengungen“ zu zwanzig Jahren Gefängnis verurteilt. Zu seinem Tod gibt es drei verschiedene Theorien. Offiziell starb er an Herzversagen. Librado Rivera, der die Leiche mit eigenen Augen gesehen hat, geht davon aus, dass Magón von einem Mitgefangenen erdrosselt wurde. Die staatstreue Gewerkschaftszeitung CROM veröffentlichte 1923 einen Beitrag, nachdem Magón von einem Gefängniswärter erschlagen wurde. mini|Die Brüder Ricardo (links) und Enrique Flores Magón (rechts) vor dem Los Angeles County Jail, 1917 [...] ``` so some Markup like `mini|` is still left. Should I run another parser on this text before feeding it to an ML model or is this a known imperfection of parsing Wiki markup? Apologies if this has been asked before.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/835/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/835/timeline
null
completed
null
null
false
[ "Hi @bminixhofer ! Parsing WikiMedia is notoriously difficult: this processing used [mwparserfromhell](https://github.com/earwig/mwparserfromhell) which is pretty good but not perfect.\r\n\r\nAs an alternative, you can also use the Wiki40b dataset which was pre-processed using an un-released Google internal tool", ...
https://api.github.com/repos/huggingface/datasets/issues/3545
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3545/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3545/comments
https://api.github.com/repos/huggingface/datasets/issues/3545/events
https://github.com/huggingface/datasets/pull/3545
1,096,189,889
PR_kwDODunzps4wpziv
3,545
fix: 🐛 pass token when retrieving the split names
[]
closed
false
null
3
2022-01-07T10:29:22Z
2022-01-10T10:51:47Z
2022-01-10T10:51:46Z
null
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3545/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3545/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/3545.diff", "html_url": "https://github.com/huggingface/datasets/pull/3545", "merged_at": "2022-01-10T10:51:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/3545.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/3545" }
true
[ "Currently, it does not work with https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0/blob/main/common_voice_7_0.py#L146 (which was the goal), because `dl_manager.download_config.use_auth_token` is ignored, and the authentication is required to be use `huggingface-cli login`.\r\nIn my use case (data...
https://api.github.com/repos/huggingface/datasets/issues/3459
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/3459/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/3459/comments
https://api.github.com/repos/huggingface/datasets/issues/3459/events
https://github.com/huggingface/datasets/issues/3459
1,084,969,672
I_kwDODunzps5Aq1LI
3,459
dataset.filter overwriting previously set dataset._indices values, resulting in the wrong elements being selected.
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
closed
false
null
2
2021-12-20T16:16:49Z
2021-12-20T16:34:57Z
2021-12-20T16:34:57Z
null
## Describe the bug When using dataset.select to select a subset of a dataset, dataset._indices are set to indicate which elements are now considered in the dataset. The same thing happens when you shuffle the dataset; dataset._indices are set to indicate what the new order of the data is. However, if you then use a dataset.filter, that filter interacts with those dataset._indices values in a non-intuitive manner. https://huggingface.co/docs/datasets/_modules/datasets/arrow_dataset.html#Dataset.filter Effectively, it looks like the original set of _indices were discared and overwritten by the set created during the filter operation. I think this is actually an issue with how the map function handles dataset._indices. Ideally it should use the _indices it gets passed, and then return an updated _indices which reflect the map transformation applied to the starting _indices. ## Steps to reproduce the bug ```python dataset = load_dataset('imdb', split='train', keep_in_memory=True) dataset = dataset.shuffle(keep_in_memory=True) dataset = dataset.select(range(0, 10), keep_in_memory=True) print("initial 10 elements") print(dataset['label']) # -> [1, 1, 0, 1, 0, 0, 0, 1, 0, 0] dataset = dataset.filter(lambda x: x['label'] == 0, keep_in_memory=True) print("filtered 10 elements looking for label 0") print(dataset['label']) # -> [1, 1, 1, 1, 1, 1] ``` ## Actual results ``` $ python indices_bug.py initial 10 elements [1, 1, 0, 1, 0, 0, 0, 1, 0, 0] filtered 10 elements looking for label 0 [1, 1, 1, 1, 1, 1] ``` This code block first shuffles the dataset (to get a mix of label 0 and label 1). Then it selects just the first 10 elements (the number of elements does not matter, 10 is just easy to visualize). The important part is that you select some subset of the dataset. Finally, a filter is applied to pull out just the elements with `label == 0`. The bug is that you cannot combine any dataset operation which sets the dataset._indices with filter. In this case I have 2, shuffle and subset. If you just use a single dataset._indices operation (in this case shuffle) the bug still shows up. The shuffle sets the dataset._indices and then filter uses those indices in the map, then overwrites dataset._indices with the filter results. ```python dataset = load_dataset('imdb', split='train', keep_in_memory=True) dataset = dataset.shuffle(keep_in_memory=True) dataset = dataset.filter(lambda x: x['label'] == 0, keep_in_memory=True) dataset = dataset.select(range(0, 10), keep_in_memory=True) print(dataset['label']) # -> [1, 1, 1, 1, 1, 1, 1, 1, 1, 1] ``` ## Expected results In an ideal world, the dataset filter would respect any dataset._indices values which had previously been set. If you use dataset.filter with the base dataset (where dataset._indices has not been set) then the filter command works as expected. ## Environment info Here are the commands required to rebuild the conda environment from scratch. ``` # create a virtual environment conda create -n dataset_indices python=3.8 -y # activate the virtual environment conda activate dataset_indices # install huggingface datasets conda install datasets ``` <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 1.12.1 - Platform: Linux-5.11.0-41-generic-x86_64-with-glibc2.17 - Python version: 3.8.12 - PyArrow version: 3.0.0 ### Full Conda Environment ``` $ conda env export name: dasaset_indices channels: - defaults dependencies: - _libgcc_mutex=0.1=main - _openmp_mutex=4.5=1_gnu - abseil-cpp=20210324.2=h2531618_0 - aiohttp=3.8.1=py38h7f8727e_0 - aiosignal=1.2.0=pyhd3eb1b0_0 - arrow-cpp=3.0.0=py38h6b21186_4 - attrs=21.2.0=pyhd3eb1b0_0 - aws-c-common=0.4.57=he6710b0_1 - aws-c-event-stream=0.1.6=h2531618_5 - aws-checksums=0.1.9=he6710b0_0 - aws-sdk-cpp=1.8.185=hce553d0_0 - bcj-cffi=0.5.1=py38h295c915_0 - blas=1.0=mkl - boost-cpp=1.73.0=h27cfd23_11 - bottleneck=1.3.2=py38heb32a55_1 - brotli=1.0.9=he6710b0_2 - brotli-python=1.0.9=py38heb0550a_2 - brotlicffi=1.0.9.2=py38h295c915_0 - brotlipy=0.7.0=py38h27cfd23_1003 - bzip2=1.0.8=h7b6447c_0 - c-ares=1.17.1=h27cfd23_0 - ca-certificates=2021.10.26=h06a4308_2 - certifi=2021.10.8=py38h06a4308_0 - cffi=1.14.6=py38h400218f_0 - conllu=4.4.1=pyhd3eb1b0_0 - cryptography=36.0.0=py38h9ce1e76_0 - dataclasses=0.8=pyh6d0b6a4_7 - dill=0.3.4=pyhd3eb1b0_0 - double-conversion=3.1.5=he6710b0_1 - et_xmlfile=1.1.0=py38h06a4308_0 - filelock=3.4.0=pyhd3eb1b0_0 - frozenlist=1.2.0=py38h7f8727e_0 - gflags=2.2.2=he6710b0_0 - glog=0.5.0=h2531618_0 - gmp=6.2.1=h2531618_2 - grpc-cpp=1.39.0=hae934f6_5 - huggingface_hub=0.0.17=pyhd3eb1b0_0 - icu=58.2=he6710b0_3 - idna=3.3=pyhd3eb1b0_0 - importlib-metadata=4.8.2=py38h06a4308_0 - importlib_metadata=4.8.2=hd3eb1b0_0 - intel-openmp=2021.4.0=h06a4308_3561 - krb5=1.19.2=hac12032_0 - ld_impl_linux-64=2.35.1=h7274673_9 - libboost=1.73.0=h3ff78a5_11 - libcurl=7.80.0=h0b77cf5_0 - libedit=3.1.20210910=h7f8727e_0 - libev=4.33=h7f8727e_1 - libevent=2.1.8=h1ba5d50_1 - libffi=3.3=he6710b0_2 - libgcc-ng=9.3.0=h5101ec6_17 - libgomp=9.3.0=h5101ec6_17 - libnghttp2=1.46.0=hce63b2e_0 - libprotobuf=3.17.2=h4ff587b_1 - libssh2=1.9.0=h1ba5d50_1 - libstdcxx-ng=9.3.0=hd4cf53a_17 - libthrift=0.14.2=hcc01f38_0 - libxml2=2.9.12=h03d6c58_0 - libxslt=1.1.34=hc22bd24_0 - lxml=4.6.3=py38h9120a33_0 - lz4-c=1.9.3=h295c915_1 - mkl=2021.4.0=h06a4308_640 - mkl-service=2.4.0=py38h7f8727e_0 - mkl_fft=1.3.1=py38hd3c417c_0 - mkl_random=1.2.2=py38h51133e4_0 - multiprocess=0.70.12.2=py38h7f8727e_0 - multivolumefile=0.2.3=pyhd3eb1b0_0 - ncurses=6.3=h7f8727e_2 - numexpr=2.7.3=py38h22e1b3c_1 - numpy=1.21.2=py38h20f2e39_0 - numpy-base=1.21.2=py38h79a1101_0 - openpyxl=3.0.9=pyhd3eb1b0_0 - openssl=1.1.1l=h7f8727e_0 - orc=1.6.9=ha97a36c_3 - packaging=21.3=pyhd3eb1b0_0 - pip=21.2.4=py38h06a4308_0 - py7zr=0.16.1=pyhd3eb1b0_1 - pycparser=2.21=pyhd3eb1b0_0 - pycryptodomex=3.10.1=py38h27cfd23_1 - pyopenssl=21.0.0=pyhd3eb1b0_1 - pyparsing=3.0.4=pyhd3eb1b0_0 - pyppmd=0.16.1=py38h295c915_0 - pysocks=1.7.1=py38h06a4308_0 - python=3.8.12=h12debd9_0 - python-dateutil=2.8.2=pyhd3eb1b0_0 - python-xxhash=2.0.2=py38h7f8727e_0 - pyzstd=0.14.4=py38h7f8727e_3 - re2=2020.11.01=h2531618_1 - readline=8.1=h27cfd23_0 - requests=2.26.0=pyhd3eb1b0_0 - setuptools=58.0.4=py38h06a4308_0 - six=1.16.0=pyhd3eb1b0_0 - snappy=1.1.8=he6710b0_0 - sqlite=3.36.0=hc218d9a_0 - texttable=1.6.4=pyhd3eb1b0_0 - tk=8.6.11=h1ccaba5_0 - typing_extensions=3.10.0.2=pyh06a4308_0 - uriparser=0.9.3=he6710b0_1 - utf8proc=2.6.1=h27cfd23_0 - wheel=0.37.0=pyhd3eb1b0_1 - xxhash=0.8.0=h7f8727e_3 - xz=5.2.5=h7b6447c_0 - zipp=3.6.0=pyhd3eb1b0_0 - zlib=1.2.11=h7f8727e_4 - zstd=1.4.9=haebb681_0 - pip: - async-timeout==4.0.2 - charset-normalizer==2.0.9 - datasets==1.16.1 - fsspec==2021.11.1 - huggingface-hub==0.2.1 - multidict==5.2.0 - pandas==1.3.5 - pyarrow==6.0.1 - pytz==2021.3 - pyyaml==6.0 - tqdm==4.62.3 - typing-extensions==4.0.1 - urllib3==1.26.7 - yarl==1.7.2 ```
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/3459/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/3459/timeline
null
completed
null
null
false
[ "I think this is a duplicate of [#3190](https://github.com/huggingface/datasets/issues/3190)?", "Upgrading the datasets version as per #3190 fixes this bug. \r\nI'm Marking as closed." ]
https://api.github.com/repos/huggingface/datasets/issues/1752
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1752/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1752/comments
https://api.github.com/repos/huggingface/datasets/issues/1752/events
https://github.com/huggingface/datasets/pull/1752
789,822,459
MDExOlB1bGxSZXF1ZXN0NTU4MTA5NTA5
1,752
COMET metric citation
[]
closed
false
null
1
2021-01-20T09:54:43Z
2021-01-20T10:27:07Z
2021-01-20T10:25:02Z
null
In my last pull request to add COMET metric, the citations where not following the usual "format". Because of that they where not correctly displayed on the website: <img width="814" alt="Screenshot 2021-01-20 at 09 48 44" src="https://user-images.githubusercontent.com/17256847/105158000-686efb80-5b05-11eb-8bb0-9c85fdac2938.png"> This pull request is only intended to fix that.
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/1752/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/1752/timeline
null
null
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/1752.diff", "html_url": "https://github.com/huggingface/datasets/pull/1752", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/1752.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/1752" }
true
[ "I think its better to create a new branch with this fix. I forgot I was still using the old branch." ]