id
int64
599M
3.29B
url
stringlengths
58
61
html_url
stringlengths
46
51
number
int64
1
7.72k
title
stringlengths
1
290
state
stringclasses
2 values
comments
int64
0
70
created_at
timestamp[s]date
2020-04-14 10:18:02
2025-08-05 09:28:51
updated_at
timestamp[s]date
2020-04-27 16:04:17
2025-08-05 11:39:56
closed_at
timestamp[s]date
2020-04-14 12:01:40
2025-08-01 05:15:45
user_login
stringlengths
3
26
labels
listlengths
0
4
body
stringlengths
0
228k
is_pull_request
bool
2 classes
1,388,201,146
https://api.github.com/repos/huggingface/datasets/issues/5031
https://github.com/huggingface/datasets/pull/5031
5,031
Support hfh 0.10 implicit auth
closed
4
2022-09-27T18:37:49
2022-09-30T09:18:24
2022-09-30T09:15:59
lhoestq
[]
In huggingface-hub 0.10 the `token` parameter is deprecated for dataset_info and list_repo_files in favor of use_auth_token. Moreover if use_auth_token=None then the user's token is used implicitly. I took those two changes into account Close https://github.com/huggingface/datasets/issues/4990 TODO: - [x] fi...
true
1,388,061,340
https://api.github.com/repos/huggingface/datasets/issues/5030
https://github.com/huggingface/datasets/pull/5030
5,030
Fast dataset iter
closed
2
2022-09-27T16:44:51
2022-09-29T15:50:44
2022-09-29T15:48:17
mariosasko
[]
Use `pa.Table.to_reader` to make iteration over examples/batches faster in `Dataset.{__iter__, map}` TODO: * [x] benchmarking (the only benchmark for now - iterating over (single) examples of `bookcorpus` (75 mil examples) in Colab is approx. 2.3x faster) * [x] check if iterating over bigger chunks + slicing to fe...
true
1,387,600,960
https://api.github.com/repos/huggingface/datasets/issues/5029
https://github.com/huggingface/datasets/pull/5029
5,029
Fix import in `ClassLabel` docstring example
closed
1
2022-09-27T11:35:29
2022-09-27T14:03:24
2022-09-27T12:27:50
alvarobartt
[]
This PR addresses a super-simple fix: adding a missing `import` to the `ClassLabel` docstring example, as it was formatted as `from datasets Features`, so it's been fixed to `from datasets import Features`.
true
1,386,272,533
https://api.github.com/repos/huggingface/datasets/issues/5028
https://github.com/huggingface/datasets/issues/5028
5,028
passing parameters to the method passed to Dataset.from_generator()
closed
1
2022-09-26T15:20:06
2022-10-03T13:00:00
2022-10-03T13:00:00
Basir-mahmood
[ "enhancement" ]
Big thanks for providing dataset creation via a generator. I want to ask whether there is any way that parameters can be passed to the method Dataset.from_generator() method, like as follows. ``` from datasets import Dataset def gen(param1): for idx in len(custom_dataset): yield custom_dataset[id...
false
1,386,153,072
https://api.github.com/repos/huggingface/datasets/issues/5027
https://github.com/huggingface/datasets/pull/5027
5,027
Fix typo in error message
closed
1
2022-09-26T14:10:09
2022-09-27T12:28:03
2022-09-27T12:26:02
severo
[]
null
true
1,386,071,154
https://api.github.com/repos/huggingface/datasets/issues/5026
https://github.com/huggingface/datasets/pull/5026
5,026
patch CI_HUB_TOKEN_PATH with Path instead of str
closed
1
2022-09-26T13:19:01
2022-09-26T14:30:55
2022-09-26T14:28:45
Wauplin
[]
Should fix the tests for `huggingface_hub==0.10.0rc0` prerelease (see [failed CI](https://github.com/huggingface/datasets/actions/runs/3127805250/jobs/5074879144)). Related to [this thread](https://huggingface.slack.com/archives/C02V5EA0A95/p1664195165294559) (internal link). Note: this should be a backward compat...
true
1,386,011,239
https://api.github.com/repos/huggingface/datasets/issues/5025
https://github.com/huggingface/datasets/issues/5025
5,025
Custom Json Dataset Throwing Error when batch is False
closed
2
2022-09-26T12:38:39
2022-09-27T19:50:00
2022-09-27T19:50:00
jmandivarapu1
[ "bug" ]
## Describe the bug A clear and concise description of what the bug is. I tried to create my custom dataset using below code ``` from datasets import Features, Sequence, ClassLabel, Value, Array2D, Array3D from torchvision import transforms from transformers import AutoProcessor # we'll use the Auto API here -...
false
1,385,947,624
https://api.github.com/repos/huggingface/datasets/issues/5024
https://github.com/huggingface/datasets/pull/5024
5,024
Fix string features of xcsr dataset
closed
1
2022-09-26T11:55:36
2022-09-28T07:56:18
2022-09-28T07:54:19
albertvillanova
[ "dataset contribution" ]
This PR fixes string features of `xcsr` dataset to avoid character splitting. Fix #5023. CC: @yangxqiao, @yuchenlin
true
1,385,881,112
https://api.github.com/repos/huggingface/datasets/issues/5023
https://github.com/huggingface/datasets/issues/5023
5,023
Text strings are split into lists of characters in xcsr dataset
closed
0
2022-09-26T11:11:50
2022-09-28T07:54:20
2022-09-28T07:54:20
albertvillanova
[ "dataset bug" ]
## Describe the bug Text strings are split into lists of characters. Example for "X-CSQA-en": ``` {'id': 'd3845adc08414fda', 'lang': 'en', 'question': {'stem': ['T', 'h', 'e', ' ', 'd', 'e', 'n', 't', 'a', 'l', ' ', 'o', 'f', 'f', 'i', 'c', 'e', ...
false
1,385,432,859
https://api.github.com/repos/huggingface/datasets/issues/5022
https://github.com/huggingface/datasets/pull/5022
5,022
Fix languages of X-CSQA configs in xcsr dataset
closed
4
2022-09-26T05:13:39
2022-09-26T12:27:20
2022-09-26T10:57:30
albertvillanova
[ "dataset contribution" ]
Fix #5017. CC: @yangxqiao, @yuchenlin
true
1,385,351,250
https://api.github.com/repos/huggingface/datasets/issues/5021
https://github.com/huggingface/datasets/issues/5021
5,021
Split is inferred from filename and overrides metadata.jsonl
closed
3
2022-09-26T03:22:14
2022-09-29T08:07:50
2022-09-29T08:07:50
float-trip
[ "bug", "duplicate" ]
## Describe the bug Including the strings "test" or "train" anywhere in a filename causes `datasets` to infer the split and silently ignore all other files. This behavior is documented for directory names but not filenames: https://huggingface.co/docs/datasets/image_dataset#imagefolder ## Steps to reproduce th...
false
1,384,684,078
https://api.github.com/repos/huggingface/datasets/issues/5020
https://github.com/huggingface/datasets/pull/5020
5,020
Fix URLs of sbu_captions dataset
closed
1
2022-09-24T14:00:33
2022-09-28T07:20:20
2022-09-28T07:18:23
donglixp
[ "dataset contribution" ]
Forbidden You don't have permission to access /~vicente/sbucaptions/sbu-captions-all.tar.gz on this server. Additionally, a 403 Forbidden error was encountered while trying to use an ErrorDocument to handle the request. Apache/2.4.6 (Red Hat Enterprise Linux) OpenSSL/1.0.2k-fips PHP/5.4.16 mod_fcgid/2.3.9 mod_ws...
true
1,384,673,718
https://api.github.com/repos/huggingface/datasets/issues/5019
https://github.com/huggingface/datasets/pull/5019
5,019
Update swiss judgment prediction
closed
4
2022-09-24T13:28:57
2022-09-28T07:13:39
2022-09-28T05:48:50
JoelNiklaus
[ "dataset contribution" ]
Hi, I updated the dataset to include additional data made available recently. When I test it locally, it seems to work. However, I get the following error with the dummy data creation: `Dummy data generation done but dummy data test failed since splits ['train', 'validation', 'test'] have 0 examples for config 'fr...
true
1,384,146,585
https://api.github.com/repos/huggingface/datasets/issues/5018
https://github.com/huggingface/datasets/pull/5018
5,018
Create all YAML dataset_info
closed
2
2022-09-23T18:08:15
2023-09-24T09:33:21
2022-10-03T17:08:05
lhoestq
[ "dataset contribution" ]
Following https://github.com/huggingface/datasets/pull/4926 Creates all the `dataset_info` YAML fields in the dataset cards The JSON are also updated using the simplified backward compatible format added in https://github.com/huggingface/datasets/pull/4926 Needs https://github.com/huggingface/datasets/pull/4926 ...
true
1,384,022,463
https://api.github.com/repos/huggingface/datasets/issues/5017
https://github.com/huggingface/datasets/issues/5017
5,017
xcsr: X-CSQA simply uses english for all alleged non-english data
closed
1
2022-09-23T16:11:54
2022-09-26T10:57:31
2022-09-26T10:57:31
thesofakillers
[ "dataset bug" ]
## Describe the bug All the alleged non-english subcollections for the X-CSQA task in the [xcsr benchmark dataset ](https://huggingface.co/datasets/xcsr) seem to be copies of the english subcollection, rather than translations. This is in contrast to the data description: > we automatically translate the original C...
false
1,383,883,058
https://api.github.com/repos/huggingface/datasets/issues/5016
https://github.com/huggingface/datasets/pull/5016
5,016
Fix tar extraction vuln
closed
1
2022-09-23T14:22:21
2022-09-29T12:42:26
2022-09-29T12:40:28
lhoestq
[]
Fix for CVE-2007-4559 Description: Directory traversal vulnerability in the (1) extract and (2) extractall functions in the tarfile module in Python allows user-assisted remote attackers to overwrite arbitrary files via a .. (dot dot) sequence in filenames in a TAR archive, a related issue to CVE-2001-1267. I ...
true
1,383,485,558
https://api.github.com/repos/huggingface/datasets/issues/5015
https://github.com/huggingface/datasets/issues/5015
5,015
Transfer dataset scripts to Hub
closed
1
2022-09-23T08:48:10
2022-10-05T07:15:57
2022-10-05T07:15:57
albertvillanova
[]
Before merging: - #4974 TODO: - [x] Create label: ["dataset contribution"](https://github.com/huggingface/datasets/pulls?q=label%3A%22dataset+contribution%22) - [x] Create project: [Datasets: Transfer datasets to Hub](https://github.com/orgs/huggingface/projects/22/) - [x] PRs: - [x] Add dataset: we should r...
false
1,383,422,639
https://api.github.com/repos/huggingface/datasets/issues/5014
https://github.com/huggingface/datasets/issues/5014
5,014
I need to read the custom dataset in conll format
open
3
2022-09-23T07:49:42
2022-11-02T11:57:15
null
shell-nlp
[ "enhancement" ]
I need to read the custom dataset in conll format
false
1,383,415,971
https://api.github.com/repos/huggingface/datasets/issues/5013
https://github.com/huggingface/datasets/issues/5013
5,013
would huggingface like publish cpp binding for datasets package ?
closed
5
2022-09-23T07:42:49
2023-02-24T16:20:57
2023-02-24T16:20:57
mullerhai
[ "wontfix" ]
HI: I use cpp env libtorch, I like use hugggingface ,but huggingface not cpp binding, would you like publish cpp binding for it. thanks
false
1,382,851,096
https://api.github.com/repos/huggingface/datasets/issues/5012
https://github.com/huggingface/datasets/issues/5012
5,012
Force JSON format regardless of file naming on S3
closed
4
2022-09-22T18:28:15
2023-08-16T09:58:36
2023-08-16T09:58:36
junwang-wish
[ "enhancement" ]
I have a file on S3 created by Data Version Control, it looks like `s3://dvc/ac/badff5b134382a0f25248f1b45d7b2` but contains a json file. If I run ```python dataset = load_dataset( "json", data_files='s3://dvc/ac/badff5b134382a0f25248f1b45d7b2' ) ``` It gives me ``` InvalidSchema: No connection adap...
false
1,382,609,587
https://api.github.com/repos/huggingface/datasets/issues/5011
https://github.com/huggingface/datasets/issues/5011
5,011
Audio: `encode_example` fails with IndexError
closed
1
2022-09-22T15:07:27
2022-09-23T09:05:18
2022-09-23T09:05:18
sanchit-gandhi
[ "bug" ]
## Describe the bug Loading the dataset [earnings-22](https://huggingface.co/datasets/sanchit-gandhi/earnings22_split) from the Hub yields an Index Error. I created this dataset locally and then pushed to hub at the specified URL. Thus, I expect the dataset should work out-of-the-box! Indeed, the dataset viewer functi...
false
1,382,308,799
https://api.github.com/repos/huggingface/datasets/issues/5010
https://github.com/huggingface/datasets/pull/5010
5,010
Add deprecation warning to multilingual_librispeech dataset card
closed
1
2022-09-22T11:41:59
2022-09-23T12:04:37
2022-09-23T12:02:45
albertvillanova
[ "dataset contribution" ]
Besides the current deprecation warning in the script of `multilingual_librispeech`, this PR adds a deprecation warning to its dataset card as well. The format of the deprecation warning is aligned with the one in the library documentation when docstrings contain the `<Deprecated/>` tag. Related to: - #4060
true
1,381,194,067
https://api.github.com/repos/huggingface/datasets/issues/5009
https://github.com/huggingface/datasets/issues/5009
5,009
Error loading StonyBrookNLP/tellmewhy dataset from hub even though local copy loads correctly
closed
8
2022-09-21T16:23:06
2022-09-29T13:07:29
2022-09-29T13:07:29
ykl7
[ "bug" ]
## Describe the bug I have added a new dataset with the identifier `StonyBrookNLP/tellmewhy` to the hub. When I load the individual files using my local copy using `dataset = datasets.load_dataset("json", data_files="data/train.jsonl")`, it loads the dataset correctly. However, when I try to load it from the hub, I ge...
false
1,381,090,903
https://api.github.com/repos/huggingface/datasets/issues/5008
https://github.com/huggingface/datasets/pull/5008
5,008
Re-apply input columns change
closed
1
2022-09-21T15:09:01
2022-09-22T13:57:36
2022-09-22T13:55:23
mariosasko
[]
Fixes the `filter` + `input_columns` combination, which is used in the `transformers` examples for instance. Revert #5006 (which in turn reverts #4971) Fix https://github.com/huggingface/datasets/issues/4858
true
1,381,007,607
https://api.github.com/repos/huggingface/datasets/issues/5007
https://github.com/huggingface/datasets/pull/5007
5,007
Add some note about running the transformers ci before a release
closed
1
2022-09-21T14:14:25
2022-09-22T10:16:14
2022-09-22T10:14:06
lhoestq
[]
null
true
1,380,968,395
https://api.github.com/repos/huggingface/datasets/issues/5006
https://github.com/huggingface/datasets/pull/5006
5,006
Revert input_columns change
closed
2
2022-09-21T13:49:20
2022-09-21T14:14:33
2022-09-21T14:11:57
lhoestq
[]
Revert https://github.com/huggingface/datasets/pull/4971 Fix https://github.com/huggingface/datasets/issues/5005
true
1,380,952,960
https://api.github.com/repos/huggingface/datasets/issues/5005
https://github.com/huggingface/datasets/issues/5005
5,005
Release 2.5.0 breaks transformers CI
closed
1
2022-09-21T13:39:19
2022-09-21T14:11:57
2022-09-21T14:11:57
albertvillanova
[ "bug" ]
## Describe the bug As reported by @lhoestq: > see https://app.circleci.com/pipelines/github/huggingface/transformers/47634/workflows/b491886b-e66e-4edb-af96-8b459e72aa25/jobs/564563 this is used here: [https://github.com/huggingface/transformers/blob/3b19c0317b6909e2d7f11b5053895ac55[…]torch/speech-pretraining/ru...
false
1,380,860,606
https://api.github.com/repos/huggingface/datasets/issues/5004
https://github.com/huggingface/datasets/pull/5004
5,004
Remove license tag file and validation
closed
1
2022-09-21T12:35:14
2022-09-22T11:47:41
2022-09-22T11:45:46
albertvillanova
[]
As requested, we are removing the validation of the licenses from `datasets` because this is done on the Hub. Fix #4994. Related to: - #4926, which is removing all the validation from `datasets`
true
1,380,617,353
https://api.github.com/repos/huggingface/datasets/issues/5003
https://github.com/huggingface/datasets/pull/5003
5,003
Fix missing use_auth_token in streaming docstrings
closed
1
2022-09-21T09:27:03
2022-09-21T16:24:01
2022-09-21T16:20:59
albertvillanova
[]
This PRs fixes docstrings: - adds the missing `use_auth_token` param - updates syntax of param types - adds params to docstrings without them - fixes return/yield types - fixes syntax
true
1,380,589,402
https://api.github.com/repos/huggingface/datasets/issues/5002
https://github.com/huggingface/datasets/issues/5002
5,002
Dataset Viewer issue for loubnabnl/humaneval-x
closed
2
2022-09-21T09:06:17
2022-09-21T11:49:49
2022-09-21T11:49:49
loubnabnl
[ "dataset-viewer" ]
### Link https://huggingface.co/datasets/loubnabnl/humaneval-x/viewer/ ### Description The dataset has subsets but the viewer gets stuck in the default subset even when I select another one (the data loading of the subsets works fine) ### Owner Yes
false
1,379,844,820
https://api.github.com/repos/huggingface/datasets/issues/5001
https://github.com/huggingface/datasets/pull/5001
5,001
Support loading XML datasets
open
3
2022-09-20T18:42:58
2024-05-22T22:13:25
null
albertvillanova
[]
CC: @davanstrien
true
1,379,709,398
https://api.github.com/repos/huggingface/datasets/issues/5000
https://github.com/huggingface/datasets/issues/5000
5,000
Dataset Viewer issue for asapp/slue
closed
9
2022-09-20T16:45:45
2022-09-27T07:04:03
2022-09-21T07:24:07
fwu-asapp
[]
### Link https://huggingface.co/datasets/asapp/slue/viewer/ ### Description Hi, I wonder how to get the dataset viewer of our slue dataset to work. Best, Felix ### Owner Yes
false
1,379,610,030
https://api.github.com/repos/huggingface/datasets/issues/4999
https://github.com/huggingface/datasets/pull/4999
4,999
Add EmptyDatasetError
closed
1
2022-09-20T15:28:05
2022-09-21T12:23:43
2022-09-21T12:21:24
lhoestq
[]
examples: from the hub: ```python Traceback (most recent call last): File "playground/ttest.py", line 3, in <module> print(load_dataset("lhoestq/empty")) File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/load.py", line 1686, in load_dataset **config_kwargs, File "/Users/quentinlhoest/Deskto...
true
1,379,466,717
https://api.github.com/repos/huggingface/datasets/issues/4998
https://github.com/huggingface/datasets/pull/4998
4,998
Don't add a tag on the Hub on release
closed
1
2022-09-20T13:54:57
2022-09-20T14:11:46
2022-09-20T14:08:54
lhoestq
[]
Datasets with no namespace on the Hub have tags to redirect to the version of datasets where they come from. I’m about to remove them all because I think it looks bad/unexpected in the UI and it’s not actually useful Therefore I'm also disabling tagging. Note that the CI job will be completely removed in https:/...
true
1,379,430,711
https://api.github.com/repos/huggingface/datasets/issues/4997
https://github.com/huggingface/datasets/pull/4997
4,997
Add support for parsing JSON files in array form
closed
1
2022-09-20T13:31:26
2022-09-20T15:42:40
2022-09-20T15:40:06
mariosasko
[]
Support parsing JSON files in the array form (top-level object is an array). For simplicity, `json.load` is used for decoding. This means the entire file is loaded into memory. If requested, we can optimize this by introducing a param similar to `lines` in [`pandas.read_json`](https://pandas.pydata.org/docs/reference/a...
true
1,379,345,161
https://api.github.com/repos/huggingface/datasets/issues/4996
https://github.com/huggingface/datasets/issues/4996
4,996
Dataset Viewer issue for Jean-Baptiste/wikiner_fr
closed
2
2022-09-20T12:32:07
2022-09-27T12:35:44
2022-09-27T12:35:44
severo
[]
### Link https://huggingface.co/datasets/Jean-Baptiste/wikiner_fr ### Description ``` Error code: StreamingRowsError Exception: FileNotFoundError Message: [Errno 2] No such file or directory: 'zip:/data/train::https:/huggingface.co/datasets/Jean-Baptiste/wikiner_fr/resolve/main/data.zip/state.json' Tra...
false
1,379,108,482
https://api.github.com/repos/huggingface/datasets/issues/4995
https://github.com/huggingface/datasets/issues/4995
4,995
Get a specific Exception when the dataset has no data
closed
0
2022-09-20T09:31:59
2022-09-21T12:21:25
2022-09-21T12:21:25
severo
[ "enhancement", "dataset-viewer" ]
In the dataset viewer on the Hub (https://huggingface.co/datasets/glue/viewer), we would like (https://github.com/huggingface/moon-landing/issues/3882) to show a specific message when the repository lacks any data files. In that case, instead of showing a complex traceback, we want to show a call to action to help t...
false
1,379,084,015
https://api.github.com/repos/huggingface/datasets/issues/4994
https://github.com/huggingface/datasets/issues/4994
4,994
delete the hardcoded license list in `datasets`
closed
0
2022-09-20T09:14:41
2022-09-22T11:45:47
2022-09-22T11:45:47
julien-c
[]
> Feel free to delete the license list in `datasets` [...] > > Also FYI in #4926 I also removed all the validation steps anyway (language, license, types etc.) _Originally posted by @lhoestq in https://github.com/huggingface/datasets/issues/4930#issuecomment-1238401662_ > [...], in my opinion we can just delete...
false
1,379,044,435
https://api.github.com/repos/huggingface/datasets/issues/4993
https://github.com/huggingface/datasets/pull/4993
4,993
fix: avoid casting tuples after Dataset.map
closed
1
2022-09-20T08:45:16
2022-09-20T16:11:27
2022-09-20T13:08:29
szmoro
[]
This PR updates features.py to avoid casting tuples to lists when reading the results of Dataset.map as suggested by @lhoestq [here](https://github.com/huggingface/datasets/issues/4676#issuecomment-1187371367) in https://github.com/huggingface/datasets/issues/4676.
true
1,379,031,842
https://api.github.com/repos/huggingface/datasets/issues/4992
https://github.com/huggingface/datasets/pull/4992
4,992
Support streaming iwslt2017 dataset
closed
1
2022-09-20T08:35:41
2022-09-20T09:27:55
2022-09-20T09:15:24
albertvillanova
[]
Support streaming iwslt2017 dataset. Once this PR is merged: - [x] Remove old ".tgz" data files from the Hub.
true
1,378,898,752
https://api.github.com/repos/huggingface/datasets/issues/4991
https://github.com/huggingface/datasets/pull/4991
4,991
Fix missing tags in dataset cards
closed
1
2022-09-20T06:42:07
2022-09-22T12:25:32
2022-09-20T07:37:30
albertvillanova
[]
Fix missing tags in dataset cards: - aeslc - empathetic_dialogues - event2Mind - gap - iwslt2017 - newsgroup - qa4mre - scicite This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833 - #4891 - #4896 - #4908 - #4921 - #4931 - ...
true
1,378,120,806
https://api.github.com/repos/huggingface/datasets/issues/4990
https://github.com/huggingface/datasets/issues/4990
4,990
"no-token" is passed to `huggingface_hub` when token is `None`
closed
6
2022-09-19T15:14:40
2022-09-30T09:16:00
2022-09-30T09:16:00
Wauplin
[ "bug" ]
## Describe the bug In the 2 lines listed below, a token is passed to `huggingface_hub` to get information from a dataset. If no token is provided, a "no-token" string is passed. What is the purpose of it ? If no real, I would prefer if the `None` value could be sent directly to be handle by `huggingface_hub`. I fee...
false
1,376,832,233
https://api.github.com/repos/huggingface/datasets/issues/4989
https://github.com/huggingface/datasets/issues/4989
4,989
Running add_column() seems to corrupt existing sequence-type column info
closed
1
2022-09-17T17:42:05
2022-09-19T12:54:54
2022-09-19T12:54:54
derek-rocheleau
[ "bug" ]
I have a dataset that contains a column ("foo") that is a sequence type of length 4. So when I run .to_pandas() on it, the resulting dataframe correctly contains 4 columns - foo_0, foo_1, foo_2, foo_3. So the 1st row of the dataframe might look like: ds = load_dataset(...) df = ds.to_pandas() df: foo_0 | foo_1 ...
false
1,376,096,584
https://api.github.com/repos/huggingface/datasets/issues/4988
https://github.com/huggingface/datasets/issues/4988
4,988
Add `IterableDataset.from_generator` to the API
closed
3
2022-09-16T15:19:41
2022-10-05T12:10:49
2022-10-05T12:10:49
mariosasko
[ "enhancement", "good first issue" ]
We've just added `Dataset.from_generator` to the API. It would also be cool to add `IterableDataset.from_generator` to support creating an iterable dataset from a generator. cc @lhoestq
false
1,376,006,477
https://api.github.com/repos/huggingface/datasets/issues/4987
https://github.com/huggingface/datasets/pull/4987
4,987
Embed image/audio data in dl_and_prepare parquet
closed
1
2022-09-16T14:09:27
2022-09-16T16:24:47
2022-09-16T16:22:35
lhoestq
[]
Embed the bytes of the image or audio files in the Parquet files directly, instead of having a "path" that points to a local file. Indeed Parquet files are often used to share data or to be used by workers that may not have access to the local files.
true
1,375,895,035
https://api.github.com/repos/huggingface/datasets/issues/4986
https://github.com/huggingface/datasets/pull/4986
4,986
[doc] Fix broken snippet that had too many quotes
closed
2
2022-09-16T12:41:07
2022-09-16T22:12:21
2022-09-16T17:32:14
tomaarsen
[]
Hello! ### Pull request overview * Fix broken snippet in https://huggingface.co/docs/datasets/main/en/process that has too many quotes ### Details The snippet in question can be found here: https://huggingface.co/docs/datasets/main/en/process#map This screenshot shows the issue, there is a quote too many, caus...
true
1,375,807,768
https://api.github.com/repos/huggingface/datasets/issues/4985
https://github.com/huggingface/datasets/pull/4985
4,985
Prefer split patterns from directories over split patterns from filenames
closed
4
2022-09-16T11:20:40
2022-11-02T11:54:28
2022-09-29T08:07:49
polinaeterna
[]
related to https://github.com/huggingface/datasets/issues/4895
true
1,375,690,330
https://api.github.com/repos/huggingface/datasets/issues/4984
https://github.com/huggingface/datasets/pull/4984
4,984
docs: ✏️ add links to the Datasets API
closed
2
2022-09-16T09:34:12
2022-09-16T13:10:14
2022-09-16T13:07:33
severo
[]
I added some links to the Datasets API in the docs. See https://github.com/huggingface/datasets-server/pull/566 for a companion PR in the datasets-server. The idea is to improve the discovery of the API through the docs. I'm a bit shy about pasting a lot of links to the API in the docs, so it's minimal for now. I'm ...
true
1,375,667,654
https://api.github.com/repos/huggingface/datasets/issues/4983
https://github.com/huggingface/datasets/issues/4983
4,983
How to convert torch.utils.data.Dataset to huggingface dataset?
closed
15
2022-09-16T09:15:10
2023-12-14T20:54:15
2022-09-20T11:23:43
DEROOCE
[ "enhancement" ]
I look through the huggingface dataset docs, and it seems that there is no offical support function to convert `torch.utils.data.Dataset` to huggingface dataset. However, there is a way to convert huggingface dataset to `torch.utils.data.Dataset`, like below: ```python from datasets import Dataset data = [[1, 2]...
false
1,375,604,693
https://api.github.com/repos/huggingface/datasets/issues/4982
https://github.com/huggingface/datasets/issues/4982
4,982
Create dataset_infos.json with VALIDATION and TEST splits
closed
3
2022-09-16T08:21:19
2022-09-28T07:59:39
2022-09-28T07:59:39
skalinin
[ "bug" ]
The problem is described in that [issue](https://github.com/huggingface/datasets/issues/4895#issuecomment-1247975569). > When I try to create data_infos.json using datasets-cli test Peter.py --save_infos --all_configs I get an error: > ValueError: Unknown split "test". Should be one of ['train']. > > The data_i...
false
1,375,086,773
https://api.github.com/repos/huggingface/datasets/issues/4981
https://github.com/huggingface/datasets/issues/4981
4,981
Can't create a dataset with `float16` features
open
8
2022-09-15T21:03:24
2025-06-12T11:47:42
null
dconathan
[ "bug" ]
## Describe the bug I can't create a dataset with `float16` features. I understand from the traceback that this is a `pyarrow` error, but I don't see anywhere in the `datasets` documentation about how to successfully do this. Is it actually supported? I've tried older versions of `pyarrow` as well with the same e...
false
1,374,868,083
https://api.github.com/repos/huggingface/datasets/issues/4980
https://github.com/huggingface/datasets/issues/4980
4,980
Make `pyarrow` optional
closed
3
2022-09-15T17:38:03
2022-09-16T17:23:47
2022-09-16T17:23:47
KOLANICH
[ "enhancement" ]
**Is your feature request related to a problem? Please describe.** Is `pyarrow` really needed for every dataset? **Describe the solution you'd like** It is made optional. **Describe alternatives you've considered** Likely, no.
false
1,374,820,758
https://api.github.com/repos/huggingface/datasets/issues/4979
https://github.com/huggingface/datasets/pull/4979
4,979
Fix missing tags in dataset cards
closed
1
2022-09-15T16:51:03
2022-09-22T12:37:55
2022-09-15T17:12:09
albertvillanova
[]
Fix missing tags in dataset cards: - amazon_us_reviews - art - discofuse - indic_glue - ubuntu_dialogs_corpus This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833 - #4891 - #4896 - #4908 - #4921 - #4931
true
1,374,271,504
https://api.github.com/repos/huggingface/datasets/issues/4978
https://github.com/huggingface/datasets/pull/4978
4,978
Update IndicGLUE download links
closed
1
2022-09-15T10:05:57
2022-09-15T22:00:20
2022-09-15T21:57:34
sumanthd17
[]
null
true
1,372,962,157
https://api.github.com/repos/huggingface/datasets/issues/4977
https://github.com/huggingface/datasets/issues/4977
4,977
Providing dataset size
open
3
2022-09-14T13:09:27
2022-09-15T16:03:58
null
sashavor
[ "enhancement" ]
**Is your feature request related to a problem? Please describe.** Especially for big datasets like [LAION](https://huggingface.co/datasets/laion/laion2B-en/), it's hard to know exactly the downloaded size (because there are many files and you don't have their exact size when downloaded). **Describe the solution yo...
false
1,372,322,382
https://api.github.com/repos/huggingface/datasets/issues/4976
https://github.com/huggingface/datasets/issues/4976
4,976
Hope to adapt Python3.9 as soon as possible
open
3
2022-09-14T04:42:22
2022-09-26T16:32:35
null
RedHeartSecretMan
[ "enhancement" ]
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. **Describe the solution you'd like** A clear and concise description of what you want to happen. **Describe alternatives you've considered** A clear and concise description of any alternat...
false
1,371,703,691
https://api.github.com/repos/huggingface/datasets/issues/4975
https://github.com/huggingface/datasets/pull/4975
4,975
Add `fn_kwargs` param to `IterableDataset.map`
closed
4
2022-09-13T16:19:05
2023-05-05T16:53:43
2022-09-13T16:45:34
mariosasko
[]
Add the `fn_kwargs` parameter to `IterableDataset.map`. ("Resolves" https://discuss.huggingface.co/t/how-to-use-large-image-text-datasets-in-hugging-face-hub-without-downloading-for-free/22780/3)
true
1,371,682,020
https://api.github.com/repos/huggingface/datasets/issues/4974
https://github.com/huggingface/datasets/pull/4974
4,974
[GH->HF] Part 2: Remove all dataset scripts from github
closed
6
2022-09-13T16:01:12
2022-10-03T17:09:39
2022-10-03T17:07:32
lhoestq
[]
Now that all the datasets live on the Hub we can remove the /datasets directory that contains all the dataset scripts of this repository - [x] Needs https://github.com/huggingface/datasets/pull/4973 to be merged first - [x] and PR to be enabled on the Hub for non-namespaced datasets
true
1,371,600,074
https://api.github.com/repos/huggingface/datasets/issues/4973
https://github.com/huggingface/datasets/pull/4973
4,973
[GH->HF] Load datasets from the Hub
closed
2
2022-09-13T15:01:41
2023-09-24T10:06:02
2022-09-15T15:24:26
lhoestq
[]
Currently datasets with no namespace (e.g. squad, glue) are loaded from github. In this PR I changed this logic to use the Hugging Face Hub instead. This is the first step in removing all the dataset scripts in this repository related to discussions in https://github.com/huggingface/datasets/pull/4059 (I shoul...
true
1,371,443,306
https://api.github.com/repos/huggingface/datasets/issues/4972
https://github.com/huggingface/datasets/pull/4972
4,972
Fix map batched with torch output
closed
1
2022-09-13T13:16:34
2022-09-20T09:42:02
2022-09-20T09:39:33
lhoestq
[]
Reported in https://discuss.huggingface.co/t/typeerror-when-applying-map-after-set-format-type-torch/23067/2 Currently it fails if one uses batched `map` and the map function returns a torch tensor. I fixed it for torch, tf, jax and pandas series.
true
1,370,319,516
https://api.github.com/repos/huggingface/datasets/issues/4971
https://github.com/huggingface/datasets/pull/4971
4,971
Preserve non-`input_colums` in `Dataset.map` if `input_columns` are specified
closed
1
2022-09-12T18:08:24
2022-09-13T13:51:08
2022-09-13T13:48:45
mariosasko
[]
Currently, if the `input_columns` list in `Dataset.map` is specified, the columns not in that list are dropped after the `map` transform. This makes the behavior inconsistent with `IterableDataset.map`. (It seems this issue was introduced by mistake in https://github.com/huggingface/datasets/pull/2246) Fix h...
true
1,369,433,074
https://api.github.com/repos/huggingface/datasets/issues/4970
https://github.com/huggingface/datasets/pull/4970
4,970
Support streaming nli_tr dataset
closed
1
2022-09-12T07:48:45
2022-09-12T08:45:04
2022-09-12T08:43:08
albertvillanova
[]
Support streaming nli_tr dataset. This PR removes legacy `codecs.open` and replaces it with `open` that supports passing encoding. Fix #3186.
true
1,369,334,740
https://api.github.com/repos/huggingface/datasets/issues/4969
https://github.com/huggingface/datasets/pull/4969
4,969
Fix data URL and metadata of vivos dataset
closed
1
2022-09-12T06:12:34
2022-09-12T07:16:15
2022-09-12T07:14:19
albertvillanova
[]
After contacting the authors of the VIVOS dataset to report that their data server is down, we have received a reply from Hieu-Thi Luong that their data is now hosted on Zenodo: https://doi.org/10.5281/zenodo.7068130 This PR updates their data URL and some metadata (homepage, citation and license). Fix #4936.
true
1,369,312,877
https://api.github.com/repos/huggingface/datasets/issues/4968
https://github.com/huggingface/datasets/pull/4968
4,968
Support streaming compguesswhat dataset
closed
1
2022-09-12T05:42:24
2022-09-12T08:00:06
2022-09-12T07:58:06
albertvillanova
[]
Support streaming `compguesswhat` dataset. Fix #3191.
true
1,369,092,452
https://api.github.com/repos/huggingface/datasets/issues/4967
https://github.com/huggingface/datasets/pull/4967
4,967
Strip "/" in local dataset path to avoid empty dataset name error
closed
2
2022-09-11T23:09:16
2022-09-29T10:46:21
2022-09-12T15:30:38
apohllo
[]
null
true
1,368,661,002
https://api.github.com/repos/huggingface/datasets/issues/4965
https://github.com/huggingface/datasets/issues/4965
4,965
[Apple M1] MemoryError: Cannot allocate write+execute memory for ffi.callback()
closed
6
2022-09-10T15:55:49
2024-03-21T17:25:53
2023-07-21T14:45:50
hoangtnm
[ "bug" ]
## Describe the bug I'm trying to run `cast_column("audio", Audio())` on Apple M1 Pro, but it seems that it doesn't work. ## Steps to reproduce the bug ```python import datasets dataset = load_dataset("csv", data_files="./train.csv")["train"] dataset = dataset.map(lambda x: {"audio": str(DATA_DIR / "audio" / ...
false
1,368,617,322
https://api.github.com/repos/huggingface/datasets/issues/4964
https://github.com/huggingface/datasets/issues/4964
4,964
Column of arrays (2D+) are using unreasonably high memory
open
10
2022-09-10T13:07:22
2022-09-22T18:29:22
null
vigsterkr
[ "bug" ]
## Describe the bug When trying to store `Array2D, Array3D, etc` as column values in a dataset, accessing that column (or creating depending on how you create it, see code below) will cause more than 10 fold of memory usage. ## Steps to reproduce the bug ```python from datasets import Dataset, Features, Array2D, ...
false
1,368,201,188
https://api.github.com/repos/huggingface/datasets/issues/4963
https://github.com/huggingface/datasets/issues/4963
4,963
Dataset without script does not support regular JSON data file
closed
1
2022-09-09T18:45:33
2022-09-20T15:40:07
2022-09-20T15:40:07
julien-c
[]
### Link https://huggingface.co/datasets/julien-c/label-studio-my-dogs ### Description <img width="1115" alt="image" src="https://user-images.githubusercontent.com/326577/189422048-7e9c390f-bea7-4521-a232-43f049ccbd1f.png"> ### Owner Yes
false
1,368,155,365
https://api.github.com/repos/huggingface/datasets/issues/4962
https://github.com/huggingface/datasets/pull/4962
4,962
Update setup.py
closed
2
2022-09-09T17:57:56
2022-09-12T14:33:04
2022-09-12T14:33:04
DCNemesis
[]
exclude broken version of fsspec. See the [related issue](https://github.com/huggingface/datasets/issues/4961)
true
1,368,124,033
https://api.github.com/repos/huggingface/datasets/issues/4961
https://github.com/huggingface/datasets/issues/4961
4,961
fsspec 2022.8.2 breaks xopen in streaming mode
closed
6
2022-09-09T17:26:55
2022-09-12T17:45:50
2022-09-12T14:32:05
DCNemesis
[ "bug" ]
## Describe the bug When fsspec 2022.8.2 is installed in your environment, xopen will prematurely close files, making streaming mode inoperable. ## Steps to reproduce the bug ```python import datasets data = datasets.load_dataset('MLCommons/ml_spoken_words', 'id_wav', split='train', streaming=True) ``` ...
false
1,368,035,159
https://api.github.com/repos/huggingface/datasets/issues/4960
https://github.com/huggingface/datasets/issues/4960
4,960
BioASQ AttributeError: 'BuilderConfig' object has no attribute 'schema'
open
2
2022-09-09T16:06:43
2022-09-13T08:51:03
null
DSLituiev
[ "dataset bug" ]
## Describe the bug I am trying to load a dataset from drive and running into an error. ## Steps to reproduce the bug ```python data_dir = "/Users/dlituiev/repos/datasets/bioasq/BioASQ-training9b" bioasq_task_b = load_dataset("aps/bioasq_task_b", data_dir=data_dir) ``` ## Actual results `AttributeError: ...
false
1,367,924,429
https://api.github.com/repos/huggingface/datasets/issues/4959
https://github.com/huggingface/datasets/pull/4959
4,959
Fix data URLs of compguesswhat dataset
closed
1
2022-09-09T14:36:10
2022-09-09T16:01:34
2022-09-09T15:59:04
albertvillanova
[]
After we informed the `compguesswhat` dataset authors about an error with their data URLs, they have updated them: - https://github.com/CompGuessWhat/compguesswhat.github.io/issues/1 This PR updates their data URLs in our loading script. Related to: - #3191
true
1,367,695,376
https://api.github.com/repos/huggingface/datasets/issues/4958
https://github.com/huggingface/datasets/issues/4958
4,958
ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/2.4.0/datasets/jsonl/jsonl.py
closed
1
2022-09-09T11:29:55
2022-09-09T11:38:44
2022-09-09T11:38:44
hasakikiki
[ "bug" ]
Hi, When I use load_dataset from local jsonl files, below error happens, and I type the link into the browser prompting me `404: Not Found`. I download the other `.py` files using the same method and it works. It seems that the server is missing the appropriate file, or it is a problem with the code version. ``` C...
false
1,366,532,849
https://api.github.com/repos/huggingface/datasets/issues/4957
https://github.com/huggingface/datasets/pull/4957
4,957
Add `Dataset.from_generator`
closed
3
2022-09-08T15:08:25
2022-09-16T14:46:35
2022-09-16T14:44:18
mariosasko
[]
Add `Dataset.from_generator` to the API to allow creating datasets from data larger than RAM. The implementation relies on a packaged module not exposed in `load_dataset` to tie this method with `datasets`' caching mechanism. Closes https://github.com/huggingface/datasets/issues/4417
true
1,366,475,160
https://api.github.com/repos/huggingface/datasets/issues/4956
https://github.com/huggingface/datasets/pull/4956
4,956
Fix TF tests for 2.10
closed
1
2022-09-08T14:39:10
2022-09-08T15:16:51
2022-09-08T15:14:44
Rocketknight1
[]
Fixes #4953
true
1,366,382,314
https://api.github.com/repos/huggingface/datasets/issues/4955
https://github.com/huggingface/datasets/issues/4955
4,955
Raise a more precise error when the URL is unreachable in streaming mode
open
0
2022-09-08T13:52:37
2022-09-08T13:53:36
null
severo
[ "enhancement" ]
See for example: - https://github.com/huggingface/datasets/issues/3191 - https://github.com/huggingface/datasets/issues/3186 It would help provide clearer information on the Hub and help the dataset maintainer solve the issue by themselves quicker. Currently: - https://huggingface.co/datasets/compguesswhat ...
false
1,366,369,682
https://api.github.com/repos/huggingface/datasets/issues/4954
https://github.com/huggingface/datasets/pull/4954
4,954
Pin TensorFlow temporarily
closed
1
2022-09-08T13:46:15
2022-09-08T14:12:33
2022-09-08T14:10:03
albertvillanova
[]
Temporarily fix TensorFlow until a permanent solution is found. Related to: - #4953
true
1,366,356,514
https://api.github.com/repos/huggingface/datasets/issues/4953
https://github.com/huggingface/datasets/issues/4953
4,953
CI test of TensorFlow is failing
closed
0
2022-09-08T13:39:29
2022-09-08T15:14:45
2022-09-08T15:14:45
albertvillanova
[ "bug" ]
## Describe the bug The following CI test fails: https://github.com/huggingface/datasets/runs/8246722693?check_suite_focus=true ``` FAILED tests/test_py_utils.py::TempSeedTest::test_tensorflow - AssertionError: ``` Details: ``` _________________________ TempSeedTest.test_tensorflow _________________________ [...
false
1,366,354,604
https://api.github.com/repos/huggingface/datasets/issues/4952
https://github.com/huggingface/datasets/pull/4952
4,952
Add test-datasets CI job
closed
2
2022-09-08T13:38:30
2023-09-24T10:05:57
2022-09-16T13:25:48
lhoestq
[]
To avoid having too many conflicts in the datasets and metrics dependencies I split the CI into test and test-catalog test does the test of the core of the `datasets` lib, while test-catalog tests the datasets scripts and metrics scripts This also makes `pip install -e .[dev]` much smaller for developers WDYT ...
true
1,365,954,814
https://api.github.com/repos/huggingface/datasets/issues/4951
https://github.com/huggingface/datasets/pull/4951
4,951
Fix license information in qasc dataset card
closed
1
2022-09-08T10:04:39
2022-09-08T14:54:47
2022-09-08T14:52:05
albertvillanova
[]
This PR adds the license information to `qasc` dataset, once reported via GitHub by Tushar Khot, the dataset is licensed under CC BY 4.0: - https://github.com/allenai/qasc/issues/5
true
1,365,458,633
https://api.github.com/repos/huggingface/datasets/issues/4950
https://github.com/huggingface/datasets/pull/4950
4,950
Update Enwik8 broken link and information
closed
1
2022-09-08T03:15:00
2022-09-24T22:14:35
2022-09-08T14:51:00
mtanghu
[]
The current enwik8 dataset link give a 502 bad gateway error which can be view on https://huggingface.co/datasets/enwik8 (click the dropdown to see the dataset preview, it will show the error). This corrects the links, and json metadata as well as adds a little bit more information about enwik8.
true
1,365,251,916
https://api.github.com/repos/huggingface/datasets/issues/4949
https://github.com/huggingface/datasets/pull/4949
4,949
Update enwik8 fixing the broken link
closed
1
2022-09-07T22:17:14
2022-09-08T03:14:04
2022-09-08T03:14:04
mtanghu
[]
The current enwik8 dataset link give a 502 bad gateway error which can be view on https://huggingface.co/datasets/enwik8 (click the dropdown to see the dataset preview, it will show the error). This corrects the links, and json metadata as well as adds a little bit more information about enwik8.
true
1,364,973,778
https://api.github.com/repos/huggingface/datasets/issues/4948
https://github.com/huggingface/datasets/pull/4948
4,948
Fix minor typo in error message for missing imports
closed
1
2022-09-07T17:20:51
2022-09-08T14:59:31
2022-09-08T14:57:15
mariosasko
[]
null
true
1,364,967,957
https://api.github.com/repos/huggingface/datasets/issues/4947
https://github.com/huggingface/datasets/pull/4947
4,947
Try to fix the Windows CI after TF update 2.10
closed
1
2022-09-07T17:14:49
2023-09-24T10:05:38
2022-09-08T09:13:10
lhoestq
[]
null
true
1,364,692,069
https://api.github.com/repos/huggingface/datasets/issues/4946
https://github.com/huggingface/datasets/pull/4946
4,946
Introduce regex check when pushing as well
closed
2
2022-09-07T13:45:58
2022-09-13T10:19:01
2022-09-13T10:16:34
LysandreJik
[]
Closes https://github.com/huggingface/datasets/issues/4945 by adding a regex check when pushing to hub. Let me know if this is helpful and if it's the fix you would have in mind for the issue and I'm happy to contribute tests.
true
1,364,691,096
https://api.github.com/repos/huggingface/datasets/issues/4945
https://github.com/huggingface/datasets/issues/4945
4,945
Push to hub can push splits that do not respect the regex
closed
0
2022-09-07T13:45:17
2022-09-13T10:16:35
2022-09-13T10:16:35
LysandreJik
[ "bug" ]
## Describe the bug The `push_to_hub` method can push splits that do not respect the regex check that is used for downloads. Therefore, splits may be pushed but never re-used, which can be painful if the split was done after runtime preprocessing. ## Steps to reproduce the bug ```python >>> from datasets import...
false
1,364,313,569
https://api.github.com/repos/huggingface/datasets/issues/4944
https://github.com/huggingface/datasets/issues/4944
4,944
larger dataset, larger GPU memory in the training phase? Is that correct?
closed
2
2022-09-07T08:46:30
2022-09-07T12:34:58
2022-09-07T12:34:58
debby1103
[ "bug" ]
from datasets import set_caching_enabled set_caching_enabled(False) for ds_name in ["squad","newsqa","nqopen","narrativeqa"]: train_ds = load_from_disk("../../../dall/downstream/processedproqa/{}-train.hf".format(ds_name)) break train_ds = concatenate_datasets([train_ds,train_...
false
1,363,967,650
https://api.github.com/repos/huggingface/datasets/issues/4943
https://github.com/huggingface/datasets/pull/4943
4,943
Add splits to MBPP dataset
closed
4
2022-09-07T01:18:31
2022-09-13T12:29:19
2022-09-13T12:27:21
cwarny
[]
This PR addresses https://github.com/huggingface/datasets/issues/4795
true
1,363,869,421
https://api.github.com/repos/huggingface/datasets/issues/4942
https://github.com/huggingface/datasets/issues/4942
4,942
Trec Dataset has incorrect labels
closed
1
2022-09-06T22:13:40
2022-09-08T11:12:03
2022-09-08T11:12:03
wmpauli
[ "bug" ]
## Describe the bug Both coarse and fine labels seem to be out of line. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = "trec" raw_datasets = load_dataset(dataset) df = pd.DataFrame(raw_datasets["test"]) df.head() ``` ## Expected results text (string) | coarse_labe...
false
1,363,622,861
https://api.github.com/repos/huggingface/datasets/issues/4941
https://github.com/huggingface/datasets/pull/4941
4,941
Add Papers with Code ID to scifact dataset
closed
1
2022-09-06T17:46:37
2022-09-06T18:28:17
2022-09-06T18:26:01
albertvillanova
[]
This PR: - adds Papers with Code ID - forces sync between GitHub and Hub, which previously failed due to Hub validation error of the license tag: https://github.com/huggingface/datasets/runs/8200223631?check_suite_focus=true
true
1,363,513,058
https://api.github.com/repos/huggingface/datasets/issues/4940
https://github.com/huggingface/datasets/pull/4940
4,940
Fix multilinguality tag and missing sections in xquad_r dataset card
closed
1
2022-09-06T16:05:35
2022-09-12T10:11:07
2022-09-12T10:08:48
albertvillanova
[]
This PR fixes issue reported on the Hub: - Label as multilingual: https://huggingface.co/datasets/xquad_r/discussions/1
true
1,363,468,679
https://api.github.com/repos/huggingface/datasets/issues/4939
https://github.com/huggingface/datasets/pull/4939
4,939
Fix NonMatchingChecksumError in adv_glue dataset
closed
1
2022-09-06T15:31:16
2022-09-06T17:42:10
2022-09-06T17:39:16
albertvillanova
[]
Fix issue reported on the Hub: https://huggingface.co/datasets/adv_glue/discussions/1
true
1,363,429,228
https://api.github.com/repos/huggingface/datasets/issues/4938
https://github.com/huggingface/datasets/pull/4938
4,938
Remove main branch rename notice
closed
1
2022-09-06T15:03:05
2022-09-06T16:46:11
2022-09-06T16:43:53
lhoestq
[]
We added a notice in README.md to show that we renamed the master branch to main, but we can remove it now (it's been 2 months) I also unpinned the github issue about the branch renaming
true
1,363,426,946
https://api.github.com/repos/huggingface/datasets/issues/4937
https://github.com/huggingface/datasets/pull/4937
4,937
Remove deprecated identical_ok
closed
1
2022-09-06T15:01:24
2022-09-06T22:24:09
2022-09-06T22:21:57
lhoestq
[]
`huggingface-hub` says that the `identical_ok` argument of `HfApi.upload_file` is now deprecated, and will be removed soon. It even has no effect at the moment when it's passed: ```python Args: ... identical_ok (`bool`, *optional*, defaults to `True`): Deprecated: will be removed in 0.11.0. ...
true
1,363,274,907
https://api.github.com/repos/huggingface/datasets/issues/4936
https://github.com/huggingface/datasets/issues/4936
4,936
vivos (Vietnamese speech corpus) dataset not accessible
closed
3
2022-09-06T13:17:55
2022-09-21T06:06:02
2022-09-12T07:14:20
polinaeterna
[ "dataset bug" ]
## Describe the bug VIVOS data is not accessible anymore, neither of these links work (at least from France): * https://ailab.hcmus.edu.vn/assets/vivos.tar.gz (data) * https://ailab.hcmus.edu.vn/vivos (dataset page) Therefore `load_dataset` doesn't work. ## Steps to reproduce the bug ```python ds = load_dat...
false
1,363,226,736
https://api.github.com/repos/huggingface/datasets/issues/4935
https://github.com/huggingface/datasets/issues/4935
4,935
Dataset Viewer issue for ubuntu_dialogs_corpus
closed
1
2022-09-06T12:41:50
2022-09-06T12:51:25
2022-09-06T12:51:25
CibinQuadance
[ "dataset-viewer" ]
### Link _No response_ ### Description _No response_ ### Owner _No response_
false
1,363,034,253
https://api.github.com/repos/huggingface/datasets/issues/4934
https://github.com/huggingface/datasets/issues/4934
4,934
Dataset Viewer issue for indonesian-nlp/librivox-indonesia
closed
6
2022-09-06T10:03:23
2022-09-06T12:46:40
2022-09-06T12:46:40
cahya-wirawan
[]
### Link https://huggingface.co/datasets/indonesian-nlp/librivox-indonesia ### Description I created a new speech dataset https://huggingface.co/datasets/indonesian-nlp/librivox-indonesia, but the dataset preview doesn't work with following error message: ``` Server error Status code: 400 Exception: TypeEr...
false
1,363,013,023
https://api.github.com/repos/huggingface/datasets/issues/4933
https://github.com/huggingface/datasets/issues/4933
4,933
Dataset/DatasetDict.filter() cannot have `batched=True` due to `mask` (numpy array?) being non-iterable.
closed
2
2022-09-06T09:47:48
2022-09-06T11:44:27
2022-09-06T11:44:27
tianjianjiang
[ "bug" ]
## Describe the bug `Dataset/DatasetDict.filter()` cannot have `batched=True` due to `mask` (numpy array?) being non-iterable. ## Steps to reproduce the bug (In a python 3.7.12 env, I've tried 2.4.0 and 2.3.2 with both `pyarraw==9.0.0` and `pyarrow==8.0.0`.) ```python from datasets import load_dataset ds_...
false
1,362,522,423
https://api.github.com/repos/huggingface/datasets/issues/4932
https://github.com/huggingface/datasets/issues/4932
4,932
Dataset Viewer issue for bigscience-biomedical/biosses
closed
4
2022-09-05T22:40:32
2022-09-06T14:24:56
2022-09-06T14:24:56
galtay
[]
### Link https://huggingface.co/datasets/bigscience-biomedical/biosses ### Description I've just been working on adding the dataset loader script to this dataset and working with the relative imports. I'm not sure how to interpret the error below (show where the dataset preview used to be) . ``` Status code: 40...
false
1,362,298,764
https://api.github.com/repos/huggingface/datasets/issues/4931
https://github.com/huggingface/datasets/pull/4931
4,931
Fix missing tags in dataset cards
closed
1
2022-09-05T17:03:04
2022-09-22T12:40:15
2022-09-06T05:39:29
albertvillanova
[]
Fix missing tags in dataset cards: - coqa - hyperpartisan_news_detection - opinosis - scientific_papers - scifact - search_qa - wiki_qa - wiki_split - wikisql This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833 - #4891 - #489...
true