url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
48
51
id
int64
600M
1.08B
node_id
stringlengths
18
24
number
int64
2
3.45k
title
stringlengths
1
276
user
dict
labels
list
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
list
milestone
dict
comments
list
created_at
int64
1,587B
1,640B
updated_at
int64
1,588B
1,640B
closed_at
int64
1,588B
1,640B
author_association
stringclasses
3 values
active_lock_reason
null
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
draft
null
pull_request
null
is_pull_request
bool
1 class
https://api.github.com/repos/huggingface/datasets/issues/1643
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1643/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1643/comments
https://api.github.com/repos/huggingface/datasets/issues/1643/events
https://github.com/huggingface/datasets/issues/1643
775,280,046
MDU6SXNzdWU3NzUyODAwNDY=
1,643
Dataset social_bias_frames 404
{ "login": "atemate", "id": 7501517, "node_id": "MDQ6VXNlcjc1MDE1MTc=", "avatar_url": "https://avatars.githubusercontent.com/u/7501517?v=4", "gravatar_id": "", "url": "https://api.github.com/users/atemate", "html_url": "https://github.com/atemate", "followers_url": "https://api.github.com/users/atemate/...
[]
closed
false
null
[]
null
[ "I see, master is already fixed in https://github.com/huggingface/datasets/commit/9e058f098a0919efd03a136b9b9c3dec5076f626" ]
1,609,144,534,000
1,609,144,687,000
1,609,144,687,000
NONE
null
``` >>> from datasets import load_dataset >>> dataset = load_dataset("social_bias_frames") ... Downloading and preparing dataset social_bias_frames/default ... ~/.pyenv/versions/3.7.6/lib/python3.7/site-packages/datasets/utils/file_utils.py in get_from_cache(url, cache_dir, force_download, proxies, etag_timeout, ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1643/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1643/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1641
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1641/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1641/comments
https://api.github.com/repos/huggingface/datasets/issues/1641/events
https://github.com/huggingface/datasets/issues/1641
775,110,872
MDU6SXNzdWU3NzUxMTA4NzI=
1,641
muchocine dataset cannot be dowloaded
{ "login": "mrm8488", "id": 3653789, "node_id": "MDQ6VXNlcjM2NTM3ODk=", "avatar_url": "https://avatars.githubusercontent.com/u/3653789?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mrm8488", "html_url": "https://github.com/mrm8488", "followers_url": "https://api.github.com/users/mrm8488/...
[ { "id": 1935892913, "node_id": "MDU6TGFiZWwxOTM1ODkyOTEz", "url": "https://api.github.com/repos/huggingface/datasets/labels/wontfix", "name": "wontfix", "color": "ffffff", "default": true, "description": "This will not be worked on" }, { "id": 2067388877, "node_id": "MDU6TGFi...
closed
false
null
[]
null
[ "I have encountered the same error with `v1.0.1` and `v1.0.2` on both Windows and Linux environments. However, cloning the repo and using the path to the dataset's root directory worked for me. Even after having the dataset cached - passing the path is the only way (for now) to load the dataset.\r\n\r\n```python\r\...
1,609,104,388,000
1,627,967,249,000
1,627,967,249,000
NONE
null
```python --------------------------------------------------------------------------- FileNotFoundError Traceback (most recent call last) /usr/local/lib/python3.6/dist-packages/datasets/load.py in prepare_module(path, script_version, download_config, download_mode, dataset, force_local_path, ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1641/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1641/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1639
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1639/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1639/comments
https://api.github.com/repos/huggingface/datasets/issues/1639/events
https://github.com/huggingface/datasets/issues/1639
774,903,472
MDU6SXNzdWU3NzQ5MDM0NzI=
1,639
bug with sst2 in glue
{ "login": "ghost", "id": 10137, "node_id": "MDQ6VXNlcjEwMTM3", "avatar_url": "https://avatars.githubusercontent.com/u/10137?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghost", "html_url": "https://github.com/ghost", "followers_url": "https://api.github.com/users/ghost/followers", "f...
[]
open
false
null
[]
null
[ "Maybe you can use nltk's treebank detokenizer ?\r\n```python\r\nfrom nltk.tokenize.treebank import TreebankWordDetokenizer\r\n\r\nTreebankWordDetokenizer().detokenize(\"it 's a charming and often affecting journey . \".split())\r\n# \"it's a charming and often affecting journey.\"\r\n```", "I am looking for alte...
1,609,001,843,000
1,630,076,603,000
null
NONE
null
Hi I am getting very low accuracy on SST2 I investigate this and observe that for this dataset sentences are tokenized, while this is correct for the other datasets in GLUE, please see below. Is there any alternatives I could get untokenized sentences? I am unfortunately under time pressure to report some results on ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1639/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1639/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1636
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1636/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1636/comments
https://api.github.com/repos/huggingface/datasets/issues/1636/events
https://github.com/huggingface/datasets/issues/1636
774,574,378
MDU6SXNzdWU3NzQ1NzQzNzg=
1,636
winogrande cannot be dowloaded
{ "login": "ghost", "id": 10137, "node_id": "MDQ6VXNlcjEwMTM3", "avatar_url": "https://avatars.githubusercontent.com/u/10137?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghost", "html_url": "https://github.com/ghost", "followers_url": "https://api.github.com/users/ghost/followers", "f...
[]
open
false
null
[]
null
[ "I have same issue for other datasets (`myanmar_news` in my case).\r\n\r\nA version of `datasets` runs correctly on my local machine (**without GPU**) which looking for the dataset at \r\n```\r\nhttps://raw.githubusercontent.com/huggingface/datasets/master/datasets/myanmar_news/myanmar_news.py\r\n```\r\n\r\nMeanwhi...
1,608,848,902,000
1,609,163,629,000
null
NONE
null
Hi, I am getting this error when trying to run the codes on the cloud. Thank you for any suggestion and help on this @lhoestq ``` File "./finetune_trainer.py", line 318, in <module> main() File "./finetune_trainer.py", line 148, in main for task in data_args.tasks] File "./finetune_trainer.py", ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1636/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1636/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1635
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1635/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1635/comments
https://api.github.com/repos/huggingface/datasets/issues/1635/events
https://github.com/huggingface/datasets/issues/1635
774,524,492
MDU6SXNzdWU3NzQ1MjQ0OTI=
1,635
Persian Abstractive/Extractive Text Summarization
{ "login": "m3hrdadfi", "id": 2601833, "node_id": "MDQ6VXNlcjI2MDE4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/2601833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/m3hrdadfi", "html_url": "https://github.com/m3hrdadfi", "followers_url": "https://api.github.com/users/m3...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[]
1,608,832,032,000
1,609,773,064,000
1,609,773,064,000
CONTRIBUTOR
null
Assembling datasets tailored to different tasks and languages is a precious target. This would be great to have this dataset included. ## Adding a Dataset - **Name:** *pn-summary* - **Description:** *A well-structured summarization dataset for the Persian language consists of 93,207 records. It is prepared for Abs...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1635/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1635/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1634
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1634/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1634/comments
https://api.github.com/repos/huggingface/datasets/issues/1634/events
https://github.com/huggingface/datasets/issues/1634
774,487,934
MDU6SXNzdWU3NzQ0ODc5MzQ=
1,634
Inspecting datasets per category
{ "login": "ghost", "id": 10137, "node_id": "MDQ6VXNlcjEwMTM3", "avatar_url": "https://avatars.githubusercontent.com/u/10137?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghost", "html_url": "https://github.com/ghost", "followers_url": "https://api.github.com/users/ghost/followers", "f...
[]
open
false
null
[]
null
[ "That's interesting, can you tell me what you think would be useful to access to inspect a dataset?\r\n\r\nYou can filter them in the hub with the search by the way: https://huggingface.co/datasets have you seen it?", "Hi @thomwolf \r\nthank you, I was not aware of this, I was looking into the data viewer linked ...
1,608,823,594,000
1,610,098,084,000
null
NONE
null
Hi Is there a way I could get all NLI datasets/all QA datasets to get some understanding of available datasets per category? this is hard for me to inspect the datasets one by one in the webpage, thanks for the suggestions @lhoestq
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1634/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1634/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1633
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1633/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1633/comments
https://api.github.com/repos/huggingface/datasets/issues/1633/events
https://github.com/huggingface/datasets/issues/1633
774,422,603
MDU6SXNzdWU3NzQ0MjI2MDM=
1,633
social_i_qa wrong format of labels
{ "login": "ghost", "id": 10137, "node_id": "MDQ6VXNlcjEwMTM3", "avatar_url": "https://avatars.githubusercontent.com/u/10137?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghost", "html_url": "https://github.com/ghost", "followers_url": "https://api.github.com/users/ghost/followers", "f...
[]
closed
false
null
[]
null
[ "@lhoestq, should I raise a PR for this? Just a minor change while reading labels text file", "Sure feel free to open a PR thanks !" ]
1,608,815,514,000
1,609,348,729,000
1,609,348,729,000
NONE
null
Hi, there is extra "\n" in labels of social_i_qa datasets, no big deal, but I was wondering if you could remove it to make it consistent. so label is 'label': '1\n', not '1' thanks ``` >>> import datasets >>> from datasets import load_dataset >>> dataset = load_dataset( ... 'social_i_qa') cahce dir /jul...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1633/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1633/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1632
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1632/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1632/comments
https://api.github.com/repos/huggingface/datasets/issues/1632/events
https://github.com/huggingface/datasets/issues/1632
774,388,625
MDU6SXNzdWU3NzQzODg2MjU=
1,632
SICK dataset
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[]
1,608,813,614,000
1,612,540,165,000
1,612,540,165,000
CONTRIBUTOR
null
Hi, this would be great to have this dataset included. I might be missing something, but I could not find it in the list of already included datasets. Thank you. ## Adding a Dataset - **Name:** SICK - **Description:** SICK consists of about 10,000 English sentence pairs that include many examples of the lexical,...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1632/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1632/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1630
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1630/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1630/comments
https://api.github.com/repos/huggingface/datasets/issues/1630/events
https://github.com/huggingface/datasets/issues/1630
774,332,129
MDU6SXNzdWU3NzQzMzIxMjk=
1,630
Adding UKP Argument Aspect Similarity Corpus
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[ "Adding a link to the guide on adding a dataset if someone want to give it a try: https://github.com/huggingface/datasets#add-a-new-dataset-to-the-hub\r\n\r\nwe should add this guide to the issue template @lhoestq ", "thanks @thomwolf , this is added now. The template is correct, sorry my mistake not to include i...
1,608,807,691,000
1,608,809,418,000
null
CONTRIBUTOR
null
Hi, this would be great to have this dataset included. ## Adding a Dataset - **Name:** UKP Argument Aspect Similarity Corpus - **Description:** The UKP Argument Aspect Similarity Corpus (UKP ASPECT) includes 3,595 sentence pairs over 28 controversial topics. Each sentence pair was annotated via crowdsourcing as ei...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1630/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1630/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1627
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1627/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1627/comments
https://api.github.com/repos/huggingface/datasets/issues/1627/events
https://github.com/huggingface/datasets/issues/1627
773,960,255
MDU6SXNzdWU3NzM5NjAyNTU=
1,627
`Dataset.map` disable progress bar
{ "login": "Nickil21", "id": 8767964, "node_id": "MDQ6VXNlcjg3Njc5NjQ=", "avatar_url": "https://avatars.githubusercontent.com/u/8767964?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Nickil21", "html_url": "https://github.com/Nickil21", "followers_url": "https://api.github.com/users/Nicki...
[]
closed
false
null
[]
null
[ "Progress bar can be disabled like this:\r\n```python\r\nfrom datasets.utils.logging import set_verbosity_error\r\nset_verbosity_error()\r\n```\r\n\r\nThere is this line in `Dataset.map`:\r\n```python\r\nnot_verbose = bool(logger.getEffectiveLevel() > WARNING)\r\n```\r\n\r\nSo any logging level higher than `WARNING...
1,608,746,022,000
1,609,012,656,000
1,609,012,637,000
NONE
null
I can't find anything to turn off the `tqdm` progress bars while running a preprocessing function using `Dataset.map`. I want to do akin to `disable_tqdm=True` in the case of `transformers`. Is there something like that?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1627/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1627/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1624
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1624/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1624/comments
https://api.github.com/repos/huggingface/datasets/issues/1624/events
https://github.com/huggingface/datasets/issues/1624
773,669,700
MDU6SXNzdWU3NzM2Njk3MDA=
1,624
Cannot download ade_corpus_v2
{ "login": "him1411", "id": 20259310, "node_id": "MDQ6VXNlcjIwMjU5MzEw", "avatar_url": "https://avatars.githubusercontent.com/u/20259310?v=4", "gravatar_id": "", "url": "https://api.github.com/users/him1411", "html_url": "https://github.com/him1411", "followers_url": "https://api.github.com/users/him141...
[]
closed
false
null
[]
null
[ "Hi @him1411, the dataset you are trying to load has been added during the community sprint and has not been released yet. It will be available with the v2 of `datasets`.\r\nFor now, you should be able to load the datasets after installing the latest (master) version of `datasets` using pip:\r\n`pip install git+htt...
1,608,721,094,000
1,627,967,334,000
1,627,967,334,000
NONE
null
I tried this to get the dataset following this url : https://huggingface.co/datasets/ade_corpus_v2 but received this error : `Traceback (most recent call last): File "/opt/anaconda3/lib/python3.7/site-packages/datasets/load.py", line 267, in prepare_module local_path = cached_path(file_path, download_con...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1624/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1624/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1622
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1622/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1622/comments
https://api.github.com/repos/huggingface/datasets/issues/1622/events
https://github.com/huggingface/datasets/issues/1622
772,940,768
MDU6SXNzdWU3NzI5NDA3Njg=
1,622
Can't call shape on the output of select()
{ "login": "noaonoszko", "id": 47183162, "node_id": "MDQ6VXNlcjQ3MTgzMTYy", "avatar_url": "https://avatars.githubusercontent.com/u/47183162?v=4", "gravatar_id": "", "url": "https://api.github.com/users/noaonoszko", "html_url": "https://github.com/noaonoszko", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "Indeed that's a typo, do you want to open a PR to fix it?", "Yes, created a PR" ]
1,608,643,120,000
1,608,730,633,000
1,608,730,632,000
CONTRIBUTOR
null
I get the error `TypeError: tuple expected at most 1 argument, got 2` when calling `shape` on the output of `select()`. It's line 531 in shape in arrow_dataset.py that causes the problem: ``return tuple(self._indices.num_rows, self._data.num_columns)`` This makes sense, since `tuple(num1, num2)` is not a valid call....
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1622/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1622/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1618
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1618/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1618/comments
https://api.github.com/repos/huggingface/datasets/issues/1618/events
https://github.com/huggingface/datasets/issues/1618
772,248,730
MDU6SXNzdWU3NzIyNDg3MzA=
1,618
Can't filter language:EN on https://huggingface.co/datasets
{ "login": "davidefiocco", "id": 4547987, "node_id": "MDQ6VXNlcjQ1NDc5ODc=", "avatar_url": "https://avatars.githubusercontent.com/u/4547987?v=4", "gravatar_id": "", "url": "https://api.github.com/users/davidefiocco", "html_url": "https://github.com/davidefiocco", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "cc'ing @mapmeld ", "Full language list is now deployed to https://huggingface.co/datasets ! Recommend close", "Cool @mapmeld ! My 2 cents (for a next iteration), it would be cool to have a small search widget in the filter dropdown as you have a ton of languages now here! Closing this in the meantime." ]
1,608,564,203,000
1,608,657,420,000
1,608,657,369,000
NONE
null
When visiting https://huggingface.co/datasets, I don't see an obvious way to filter only English datasets. This is unexpected for me, am I missing something? I'd expect English to be selectable in the language widget. This problem reproduced on Mozilla Firefox and MS Edge: ![screenshot](https://user-images.githubuse...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1618/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1618/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1615
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1615/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1615/comments
https://api.github.com/repos/huggingface/datasets/issues/1615/events
https://github.com/huggingface/datasets/issues/1615
771,641,088
MDU6SXNzdWU3NzE2NDEwODg=
1,615
Bug: Can't download TriviaQA with `load_dataset` - custom `cache_dir`
{ "login": "SapirWeissbuch", "id": 44585792, "node_id": "MDQ6VXNlcjQ0NTg1Nzky", "avatar_url": "https://avatars.githubusercontent.com/u/44585792?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SapirWeissbuch", "html_url": "https://github.com/SapirWeissbuch", "followers_url": "https://api.gi...
[]
open
false
null
[]
null
[ "Hi @SapirWeissbuch,\r\nWhen you are saying it freezes, at that time it is unzipping the file from the zip file it downloaded. Since it's a very heavy file it'll take some time. It was taking ~11GB after unzipping when it started reading examples for me. Hope that helps!\r\n![Screenshot 2020-12-21 at 23 40 52](http...
1,608,485,258,000
1,624,626,693,000
null
NONE
null
Hello, I'm having issue downloading TriviaQA dataset with `load_dataset`. ## Environment info - `datasets` version: 1.1.3 - Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1 - Python version: 3.7.3 ## The code I'm running: ```python import datasets dataset = datasets.load_dataset("trivia_qa", "rc", c...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1615/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1615/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1611
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1611/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1611/comments
https://api.github.com/repos/huggingface/datasets/issues/1611/events
https://github.com/huggingface/datasets/issues/1611
771,486,456
MDU6SXNzdWU3NzE0ODY0NTY=
1,611
shuffle with torch generator
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Is there a way one can convert the two generator? not sure overall what alternatives I could have to shuffle the datasets with a torch generator, thanks ", "@lhoestq let me please expalin in more details, maybe you could help me suggesting an alternative to solve the issue for now, I have multiple large dataset...
1,608,425,834,000
1,608,574,339,000
null
NONE
null
Hi I need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1611/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1611/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1610
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1610/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1610/comments
https://api.github.com/repos/huggingface/datasets/issues/1610/events
https://github.com/huggingface/datasets/issues/1610
771,453,599
MDU6SXNzdWU3NzE0NTM1OTk=
1,610
shuffle does not accept seed
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi, did you check the doc on `shuffle`?\r\nhttps://huggingface.co/docs/datasets/package_reference/main_classes.html?datasets.Dataset.shuffle#datasets.Dataset.shuffle", "Hi Thomas\r\nthanks for reponse, yes, I did checked it, but this does not work for me please see \r\n\r\n```\r\n(internship) rkarimi@italix17:/i...
1,608,411,579,000
1,609,754,403,000
1,609,754,403,000
CONTRIBUTOR
null
Hi I need to shuffle the dataset, but this needs to be based on epoch+seed to be consistent across the cores, when I pass seed to shuffle, this does not accept seed, could you assist me with this? thanks @lhoestq
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1610/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1610/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1609
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1609/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1609/comments
https://api.github.com/repos/huggingface/datasets/issues/1609/events
https://github.com/huggingface/datasets/issues/1609
771,421,881
MDU6SXNzdWU3NzE0MjE4ODE=
1,609
Not able to use 'jigsaw_toxicity_pred' dataset
{ "login": "jassimran", "id": 7424133, "node_id": "MDQ6VXNlcjc0MjQxMzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7424133?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jassimran", "html_url": "https://github.com/jassimran", "followers_url": "https://api.github.com/users/ja...
[]
closed
false
null
[]
null
[ "Hi @jassimran,\r\nThe `jigsaw_toxicity_pred` dataset has not been released yet, it will be available with version 2 of `datasets`, coming soon.\r\nYou can still access it by installing the master (unreleased) version of datasets directly :\r\n`pip install git+https://github.com/huggingface/datasets.git@master`\r\n...
1,608,399,348,000
1,608,655,344,000
1,608,655,343,000
NONE
null
When trying to use jigsaw_toxicity_pred dataset, like this in a [colab](https://colab.research.google.com/drive/1LwO2A5M2X5dvhkAFYE4D2CUT3WUdWnkn?usp=sharing): ``` from datasets import list_datasets, list_metrics, load_dataset, load_metric ds = load_dataset("jigsaw_toxicity_pred") ``` I see below error: >...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1609/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1609/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1605
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1605/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1605/comments
https://api.github.com/repos/huggingface/datasets/issues/1605/events
https://github.com/huggingface/datasets/issues/1605
770,979,620
MDU6SXNzdWU3NzA5Nzk2MjA=
1,605
Navigation version breaking
{ "login": "mttk", "id": 3007947, "node_id": "MDQ6VXNlcjMwMDc5NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3007947?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mttk", "html_url": "https://github.com/mttk", "followers_url": "https://api.github.com/users/mttk/followers", ...
[]
open
false
null
[]
null
[]
1,608,305,784,000
1,608,306,112,000
null
NONE
null
Hi, when navigating docs (Chrome, Ubuntu) (e.g. on this page: https://huggingface.co/docs/datasets/loading_metrics.html#using-a-custom-metric-script) the version control dropdown has the wrong string displayed as the current version: ![image](https://user-images.githubusercontent.com/3007947/102632187-02cad080-...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1605/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1605/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1604
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1604/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1604/comments
https://api.github.com/repos/huggingface/datasets/issues/1604/events
https://github.com/huggingface/datasets/issues/1604
770,862,112
MDU6SXNzdWU3NzA4NjIxMTI=
1,604
Add tests for the download functions ?
{ "login": "SBrandeis", "id": 33657802, "node_id": "MDQ6VXNlcjMzNjU3ODAy", "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SBrandeis", "html_url": "https://github.com/SBrandeis", "followers_url": "https://api.github.com/users/...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[]
1,608,295,765,000
1,608,295,765,000
null
CONTRIBUTOR
null
AFAIK the download functions in `DownloadManager` are not tested yet. It could be good to add some to ensure behavior is as expected.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1604/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1604/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1600
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1600/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1600/comments
https://api.github.com/repos/huggingface/datasets/issues/1600/events
https://github.com/huggingface/datasets/issues/1600
770,582,960
MDU6SXNzdWU3NzA1ODI5NjA=
1,600
AttributeError: 'DatasetDict' object has no attribute 'train_test_split'
{ "login": "david-waterworth", "id": 5028974, "node_id": "MDQ6VXNlcjUwMjg5NzQ=", "avatar_url": "https://avatars.githubusercontent.com/u/5028974?v=4", "gravatar_id": "", "url": "https://api.github.com/users/david-waterworth", "html_url": "https://github.com/david-waterworth", "followers_url": "https://ap...
[ { "id": 1935892912, "node_id": "MDU6TGFiZWwxOTM1ODkyOTEy", "url": "https://api.github.com/repos/huggingface/datasets/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "Further information is requested" } ]
closed
false
null
[]
null
[ "Hi @david-waterworth!\r\n\r\nAs indicated in the error message, `load_dataset(\"csv\")` returns a `DatasetDict` object, which is mapping of `str` to `Dataset` objects. I believe in this case the behavior is to return a `train` split with all the data.\r\n`train_test_split` is a method of the `Dataset` object, so y...
1,608,269,830,000
1,623,756,346,000
1,608,536,338,000
NONE
null
The following code fails with "'DatasetDict' object has no attribute 'train_test_split'" - am I doing something wrong? ``` from datasets import load_dataset dataset = load_dataset('csv', data_files='data.txt') dataset = dataset.train_test_split(test_size=0.1) ``` > AttributeError: 'DatasetDict' object has no at...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1600/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1600/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1594
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1594/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1594/comments
https://api.github.com/repos/huggingface/datasets/issues/1594/events
https://github.com/huggingface/datasets/issues/1594
769,747,767
MDU6SXNzdWU3Njk3NDc3Njc=
1,594
connection error
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url...
[]
open
false
null
[]
null
[ "This happen quite often when they are too many concurrent requests to github.\r\n\r\ni can understand it’s a bit cumbersome to handle on the user side. Maybe we should try a few times in the lib (eg with timeout) before failing, what do you think @lhoestq ?", "Yes currently there's no retry afaik. We should add ...
1,608,196,714,000
1,608,850,653,000
null
NONE
null
Hi I am hitting to this error, thanks ``` > Traceback (most recent call last): File "finetune_t5_trainer.py", line 379, in <module> main() File "finetune_t5_trainer.py", line 208, in main if training_args.do_eval or training_args.evaluation_strategy != EvaluationStrategy.NO File "finetune_t5_tr...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1594/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1594/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1593
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1593/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1593/comments
https://api.github.com/repos/huggingface/datasets/issues/1593/events
https://github.com/huggingface/datasets/issues/1593
769,611,386
MDU6SXNzdWU3Njk2MTEzODY=
1,593
Access to key in DatasetDict map
{ "login": "ZhaofengWu", "id": 11954789, "node_id": "MDQ6VXNlcjExOTU0Nzg5", "avatar_url": "https://avatars.githubusercontent.com/u/11954789?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ZhaofengWu", "html_url": "https://github.com/ZhaofengWu", "followers_url": "https://api.github.com/use...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Indeed that would be cool\r\n\r\nAlso FYI right now the easiest way to do this is\r\n```python\r\ndataset_dict[\"train\"] = dataset_dict[\"train\"].map(my_transform_for_the_train_set)\r\ndataset_dict[\"test\"] = dataset_dict[\"test\"].map(my_transform_for_the_test_set)\r\n```" ]
1,608,188,540,000
1,610,534,283,000
null
NONE
null
It is possible that we want to do different things in the `map` function (and possibly other functions too) of a `DatasetDict`, depending on the key. I understand that `DatasetDict.map` is a really thin wrapper of `Dataset.map`, so it is easy to directly implement this functionality in the client code. Still, it'd be n...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1593/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1593/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1592
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1592/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1592/comments
https://api.github.com/repos/huggingface/datasets/issues/1592/events
https://github.com/huggingface/datasets/issues/1592
769,529,421
MDU6SXNzdWU3Njk1Mjk0MjE=
1,592
Using datasets.Metric with Trainer()
{ "login": "YipingNUS", "id": 5652584, "node_id": "MDQ6VXNlcjU2NTI1ODQ=", "avatar_url": "https://avatars.githubusercontent.com/u/5652584?v=4", "gravatar_id": "", "url": "https://api.github.com/users/YipingNUS", "html_url": "https://github.com/YipingNUS", "followers_url": "https://api.github.com/users/Yi...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[ "We are indeed working on the integration with `Trainer` :)" ]
1,608,182,224,000
1,608,205,744,000
null
NONE
null
## Using datasets.Metric with Trainer() Hi team, I was quite surprised in the [Metric documentation](https://huggingface.co/docs/datasets/using_metrics.html) I don't see how it can be used with `Trainer()`. That would be the most intuitive use case instead of having to iterate the batches and add predictions and ref...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1592/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1592/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1591
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1591/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1591/comments
https://api.github.com/repos/huggingface/datasets/issues/1591/events
https://github.com/huggingface/datasets/issues/1591
769,383,714
MDU6SXNzdWU3NjkzODM3MTQ=
1,591
IWSLT-17 Link Broken
{ "login": "ZhaofengWu", "id": 11954789, "node_id": "MDQ6VXNlcjExOTU0Nzg5", "avatar_url": "https://avatars.githubusercontent.com/u/11954789?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ZhaofengWu", "html_url": "https://github.com/ZhaofengWu", "followers_url": "https://api.github.com/use...
[ { "id": 1935892865, "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate", "name": "duplicate", "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists" }, { "id": 2067388877, "...
closed
false
null
[]
null
[ "Sorry, this is a duplicate of #1287. Not sure why it didn't come up when I searched `iwslt` in the issues list.", "Closing this since its a duplicate" ]
1,608,166,002,000
1,608,278,796,000
1,608,278,728,000
NONE
null
``` FileNotFoundError: Couldn't find file at https://wit3.fbk.eu/archive/2017-01-trnmted//texts/DeEnItNlRo/DeEnItNlRo/DeEnItNlRo-DeEnItNlRo.tgz ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1591/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1591/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1590
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1590/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1590/comments
https://api.github.com/repos/huggingface/datasets/issues/1590/events
https://github.com/huggingface/datasets/issues/1590
769,242,858
MDU6SXNzdWU3NjkyNDI4NTg=
1,590
Add helper to resolve namespace collision
{ "login": "jramapuram", "id": 8204807, "node_id": "MDQ6VXNlcjgyMDQ4MDc=", "avatar_url": "https://avatars.githubusercontent.com/u/8204807?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jramapuram", "html_url": "https://github.com/jramapuram", "followers_url": "https://api.github.com/users...
[]
open
false
null
[]
null
[ "Do you have an example?", "I was thinking about using something like [importlib](https://docs.python.org/3/library/importlib.html#importing-a-source-file-directly) to over-ride the collision. \r\n\r\n**Reason requested**: I use the [following template](https://github.com/jramapuram/ml_base/) repo where I house a...
1,608,149,844,000
1,608,349,238,000
null
NONE
null
Many projects use a module called `datasets`, however this is incompatible with huggingface datasets. It would be great if there if there was some helper or similar function to resolve such a common conflict.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1590/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1590/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1585
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1585/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1585/comments
https://api.github.com/repos/huggingface/datasets/issues/1585/events
https://github.com/huggingface/datasets/issues/1585
768,831,171
MDU6SXNzdWU3Njg4MzExNzE=
1,585
FileNotFoundError for `amazon_polarity`
{ "login": "phtephanx", "id": 24647404, "node_id": "MDQ6VXNlcjI0NjQ3NDA0", "avatar_url": "https://avatars.githubusercontent.com/u/24647404?v=4", "gravatar_id": "", "url": "https://api.github.com/users/phtephanx", "html_url": "https://github.com/phtephanx", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi @phtephanx , the `amazon_polarity` dataset has not been released yet. It will be available in the coming soon v2of `datasets` :) \r\n\r\nYou can still access it now if you want, but you will need to install datasets via the master branch:\r\n`pip install git+https://github.com/huggingface/datasets.git@master`" ...
1,608,123,065,000
1,608,134,576,000
1,608,134,576,000
NONE
null
Version: `datasets==v1.1.3` ### Reproduction ```python from datasets import load_dataset data = load_dataset("amazon_polarity") ``` crashes with ```bash FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/amazon_polarity/amazon_polarity.py ``` and ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1585/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1585/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1581
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1581/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1581/comments
https://api.github.com/repos/huggingface/datasets/issues/1581/events
https://github.com/huggingface/datasets/issues/1581
768,320,594
MDU6SXNzdWU3NjgzMjA1OTQ=
1,581
Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers'
{ "login": "eduardofv", "id": 702586, "node_id": "MDQ6VXNlcjcwMjU4Ng==", "avatar_url": "https://avatars.githubusercontent.com/u/702586?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eduardofv", "html_url": "https://github.com/eduardofv", "followers_url": "https://api.github.com/users/edua...
[]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nYou can override the directory in which cache file are stored using for example\r\n```\r\nENV HF_HOME=\"/root/cache/hf_cache_home\"\r\n```\r\n\r\nThis way both `transformers` and `datasets` will use this directory instead of the default `.cache`", "Great, thanks. I didn't see documentat...
1,608,076,941,000
1,623,944,445,000
1,623,944,445,000
NONE
null
I am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `/.cache`: ``` $ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)/data:/root/data ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1581/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1581/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1541
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1541/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1541/comments
https://api.github.com/repos/huggingface/datasets/issues/1541/events
https://github.com/huggingface/datasets/issues/1541
765,430,586
MDU6SXNzdWU3NjU0MzA1ODY=
1,541
connection issue while downloading data
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url...
[]
open
false
null
[]
null
[ "could you tell me how I can avoid download, by pre-downloading the data first, put them in a folder so the code does not try to redownload? could you tell me the path to put the downloaded data, and how to do it? thanks\r\n@lhoestq ", "Does your instance have an internet connection ?\r\n\r\nIf you don't have an ...
1,607,869,620,000
1,608,028,471,000
null
NONE
null
Hi I am running my codes on google cloud, and I am getting this error resulting in the failure of the codes when trying to download the data, could you assist me to solve this? also as a temporary solution, could you tell me how I can increase the number of retries and timeout to at least let the models run for now. t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1541/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1541/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1514
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1514/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1514/comments
https://api.github.com/repos/huggingface/datasets/issues/1514/events
https://github.com/huggingface/datasets/issues/1514
764,017,148
MDU6SXNzdWU3NjQwMTcxNDg=
1,514
how to get all the options of a property in datasets
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[ { "id": 1935892912, "node_id": "MDU6TGFiZWwxOTM1ODkyOTEy", "url": "https://api.github.com/repos/huggingface/datasets/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "Further information is requested" } ]
open
false
null
[]
null
[ "In a dataset, labels correspond to the `ClassLabel` feature that has the `names` property that returns string represenation of the integer classes (or `num_classes` to get the number of different classes).", "I think the `features` attribute of the dataset object is what you are looking for:\r\n```\r\n>>> datase...
1,607,790,248,000
1,608,278,937,000
null
CONTRIBUTOR
null
Hi could you tell me how I can get all unique options of a property of dataset? for instance in case of boolq, if the user wants to know which unique labels it has, is there a way to access unique labels without getting all training data lables and then forming a set i mean? thanks
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1514/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1514/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1478
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1478/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1478/comments
https://api.github.com/repos/huggingface/datasets/issues/1478/events
https://github.com/huggingface/datasets/issues/1478
762,293,076
MDU6SXNzdWU3NjIyOTMwNzY=
1,478
Inconsistent argument names.
{ "login": "Fraser-Greenlee", "id": 8402500, "node_id": "MDQ6VXNlcjg0MDI1MDA=", "avatar_url": "https://avatars.githubusercontent.com/u/8402500?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Fraser-Greenlee", "html_url": "https://github.com/Fraser-Greenlee", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "Also for the `Accuracy` metric the `accuracy_score` method should have its args in the opposite order so `accuracy_score(predictions, references,,,)`.", "Thanks for pointing this out ! 🕵🏻 \r\nPredictions and references should indeed be swapped in the docstring.\r\nHowever, the call to `accuracy_score` should n...
1,607,689,178,000
1,608,390,219,000
1,608,390,219,000
CONTRIBUTOR
null
Just find it a wee bit odd that in the transformers library `predictions` are those made by the model: https://github.com/huggingface/transformers/blob/master/src/transformers/trainer_utils.py#L51-L61 While in many datasets metrics they are the ground truth labels: https://github.com/huggingface/datasets/blob/c3f5...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1478/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1478/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1452
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1452/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1452/comments
https://api.github.com/repos/huggingface/datasets/issues/1452/events
https://github.com/huggingface/datasets/issues/1452
761,104,924
MDU6SXNzdWU3NjExMDQ5MjQ=
1,452
SNLI dataset contains labels with value -1
{ "login": "aarnetalman", "id": 11405654, "node_id": "MDQ6VXNlcjExNDA1NjU0", "avatar_url": "https://avatars.githubusercontent.com/u/11405654?v=4", "gravatar_id": "", "url": "https://api.github.com/users/aarnetalman", "html_url": "https://github.com/aarnetalman", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[ "I believe the `-1` label is used for missing/NULL data as per HuggingFace Dataset conventions. If I recall correctly SNLI has some entries with no (gold) labels in the dataset.", "Ah, you're right. The dataset has some pairs with missing labels. Thanks for reminding me." ]
1,607,595,415,000
1,607,622,595,000
1,607,622,595,000
NONE
null
``` import datasets nli_data = datasets.load_dataset("snli") train_data = nli_data['train'] train_labels = train_data['label'] label_set = set(train_labels) print(label_set) ``` **Output:** `{0, 1, 2, -1}`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1452/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1452/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1444
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1444/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1444/comments
https://api.github.com/repos/huggingface/datasets/issues/1444/events
https://github.com/huggingface/datasets/issues/1444
761,055,651
MDU6SXNzdWU3NjEwNTU2NTE=
1,444
FileNotFound remotly, can't load a dataset
{ "login": "sadakmed", "id": 18331629, "node_id": "MDQ6VXNlcjE4MzMxNjI5", "avatar_url": "https://avatars.githubusercontent.com/u/18331629?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sadakmed", "html_url": "https://github.com/sadakmed", "followers_url": "https://api.github.com/users/sad...
[]
closed
false
null
[]
null
[ "This dataset will be available in version-2 of the library. If you want to use this dataset now, install datasets from `master` branch rather.\r\n\r\nCommand to install datasets from `master` branch:\r\n`!pip install git+https://github.com/huggingface/datasets.git@master`", "Closing this, thanks @VasudevGupta7 "...
1,607,591,687,000
1,608,054,074,000
1,608,054,074,000
NONE
null
```py !pip install datasets import datasets as ds corpus = ds.load_dataset('large_spanish_corpus') ``` gives the error > FileNotFoundError: Couldn't find file locally at large_spanish_corpus/large_spanish_corpus.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/large_spa...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1444/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1444/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1422
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1422/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1422/comments
https://api.github.com/repos/huggingface/datasets/issues/1422/events
https://github.com/huggingface/datasets/issues/1422
760,707,113
MDU6SXNzdWU3NjA3MDcxMTM=
1,422
Can't map dataset (loaded from csv)
{ "login": "SolomidHero", "id": 28161779, "node_id": "MDQ6VXNlcjI4MTYxNzc5", "avatar_url": "https://avatars.githubusercontent.com/u/28161779?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SolomidHero", "html_url": "https://github.com/SolomidHero", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[ "Please could you post the whole script? I can't reproduce your issue. After updating the feature names/labels to match with the data, everything works fine for me. Try to update datasets/transformers to the newest version.", "Actually, the problem was how `tokenize` function was defined. This was completely my s...
1,607,551,542,000
1,608,228,820,000
1,608,228,820,000
NONE
null
Hello! I am trying to load single csv file with two columns: ('label': str, 'text' str), where is label is str of two possible classes. Below steps are similar with [this notebook](https://colab.research.google.com/drive/1-JIJlao4dI-Ilww_NnTc0rxtp-ymgDgM?usp=sharing), where bert model and tokenizer are used to class...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1422/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1422/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1324
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1324/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1324/comments
https://api.github.com/repos/huggingface/datasets/issues/1324/events
https://github.com/huggingface/datasets/issues/1324
759,587,864
MDU6SXNzdWU3NTk1ODc4NjQ=
1,324
❓ Sharing ElasticSearch indexed dataset
{ "login": "pietrolesci", "id": 61748653, "node_id": "MDQ6VXNlcjYxNzQ4NjUz", "avatar_url": "https://avatars.githubusercontent.com/u/61748653?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pietrolesci", "html_url": "https://github.com/pietrolesci", "followers_url": "https://api.github.com/...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[ "Hello @pietrolesci , I am not sure to understand what you are trying to do here.\r\n\r\nIf you're looking for ways to save a dataset on disk, you can you the `save_to_disk` method:\r\n```python\r\n>>> import datasets\r\n>>> loaded_dataset = datasets.load(\"dataset_name\")\r\n>>> loaded_dataset.save_to_disk(\"/path...
1,607,444,758,000
1,608,623,456,000
null
NONE
null
Hi there, First of all, thank you very much for this amazing library. Datasets have become my preferred data structure for basically everything I am currently doing. **Question:** I'm working with a dataset and I have an elasticsearch container running at localhost:9200. I added an elasticsearch index and I was w...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1324/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1324/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1299
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1299/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1299/comments
https://api.github.com/repos/huggingface/datasets/issues/1299/events
https://github.com/huggingface/datasets/issues/1299
759,414,566
MDU6SXNzdWU3NTk0MTQ1NjY=
1,299
can't load "german_legal_entity_recognition" dataset
{ "login": "nataly-obr", "id": 59837137, "node_id": "MDQ6VXNlcjU5ODM3MTM3", "avatar_url": "https://avatars.githubusercontent.com/u/59837137?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nataly-obr", "html_url": "https://github.com/nataly-obr", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "Please if you could tell me more about the error? \r\n\r\n1. Please check the directory you've been working on\r\n2. Check for any typos", "> Please if you could tell me more about the error?\r\n> \r\n> 1. Please check the directory you've been working on\r\n> 2. Check for any typos\r\n\r\nError happens during t...
1,607,431,321,000
1,608,134,593,000
1,608,134,593,000
NONE
null
FileNotFoundError: Couldn't find file locally at german_legal_entity_recognition/german_legal_entity_recognition.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.3/datasets/german_legal_entity_recognition/german_legal_entity_recognition.py or https://s3.amazonaws.com/datasets.huggingface.co...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1299/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1299/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1290
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1290/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1290/comments
https://api.github.com/repos/huggingface/datasets/issues/1290/events
https://github.com/huggingface/datasets/issues/1290
759,339,989
MDU6SXNzdWU3NTkzMzk5ODk=
1,290
imdb dataset cannot be downloaded
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
closed
false
null
[]
null
[ "Hi @rabeehk , I am unable to reproduce your problem locally.\r\nCan you try emptying the cache (removing the content of `/idiap/temp/rkarimi/cache_home_1/datasets`) and retry ?", "Hi,\r\nthanks, I did remove the cache and still the same error here\r\n\r\n```\r\n>>> a = datasets.load_dataset(\"imdb\", split=\"tra...
1,607,424,456,000
1,608,831,489,000
1,608,831,489,000
CONTRIBUTOR
null
hi please find error below getting imdb train spli: thanks ` datasets.load_dataset>>> datasets.load_dataset("imdb", split="train")` errors ``` cahce dir /idiap/temp/rkarimi/cache_home_1/datasets cahce dir /idiap/temp/rkarimi/cache_home_1/datasets Downloading and preparing dataset imdb/plain_text (d...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1290/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1290/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1287
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1287/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1287/comments
https://api.github.com/repos/huggingface/datasets/issues/1287/events
https://github.com/huggingface/datasets/issues/1287
759,300,992
MDU6SXNzdWU3NTkzMDA5OTI=
1,287
'iwslt2017-ro-nl', cannot be downloaded
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
null
[]
null
[ "the same issue with datasets.load_dataset(\"iwslt2017\", 'iwslt2017-en-nl', split=split), ..... ", "even with setting master like the following command, still remains \r\n\r\ndatasets.load_dataset(\"iwslt2017\", 'iwslt2017-en-nl', split=\"train\", script_version=\"master\")\r\n", "Looks like the data has been ...
1,607,421,415,000
1,608,056,694,000
null
CONTRIBUTOR
null
Hi I am trying `>>> datasets.load_dataset("iwslt2017", 'iwslt2017-ro-nl', split="train")` getting this error thank you for your help ``` cahce dir /idiap/temp/rkarimi/cache_home_1/datasets cahce dir /idiap/temp/rkarimi/cache_home_1/datasets Downloading and preparing dataset iwsl_t217/iwslt2017-ro-nl (downlo...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1287/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1287/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1286
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1286/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1286/comments
https://api.github.com/repos/huggingface/datasets/issues/1286/events
https://github.com/huggingface/datasets/issues/1286
759,291,509
MDU6SXNzdWU3NTkyOTE1MDk=
1,286
[libprotobuf FATAL /sentencepiece/src/../third_party/protobuf-lite/google/protobuf/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
closed
false
null
[]
null
[ "I remember also getting the same issue for several other translation datasets like all the iwslt2017 group, this is blokcing me and I really need to fix it and I was wondering if you have an idea on this. @lhoestq thanks,. ", "maybe there is an empty line or something inside these datasets? could you tell me wh...
1,607,420,655,000
1,607,801,782,000
1,607,790,156,000
CONTRIBUTOR
null
Hi I am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help {'epoch': 20.0} 100%|████████████████████████████...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1286/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1286/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1285
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1285/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1285/comments
https://api.github.com/repos/huggingface/datasets/issues/1285/events
https://github.com/huggingface/datasets/issues/1285
759,278,758
MDU6SXNzdWU3NTkyNzg3NTg=
1,285
boolq does not work
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
closed
false
null
[]
null
[ "here is the minimal code to reproduce\r\n\r\n`datasets>>> datasets.load_dataset(\"boolq\", \"train\")\r\n\r\nthe errors\r\n\r\n```\r\n`cahce dir /idiap/temp/rkarimi/cache_home_1/datasets\r\ncahce dir /idiap/temp/rkarimi/cache_home_1/datasets\r\nUsing custom data configuration train\r\nDownloading and preparing d...
1,607,419,727,000
1,607,420,830,000
1,607,420,830,000
CONTRIBUTOR
null
Hi I am getting this error when trying to load boolq, thanks for your help ts_boolq_default_0.1.0_2987db1f15deaa19500ae24de560eabeaf1f8ef51df88c0470beeec72943bf11.lock Traceback (most recent call last): File "finetune_t5_trainer.py", line 274, in <module> main() File "finetune_t5_trainer.py", line 147, ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1285/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1285/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1167
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1167/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1167/comments
https://api.github.com/repos/huggingface/datasets/issues/1167/events
https://github.com/huggingface/datasets/issues/1167
757,722,921
MDU6SXNzdWU3NTc3MjI5MjE=
1,167
❓ On-the-fly tokenization with datasets, tokenizers, and torch Datasets and Dataloaders
{ "login": "pietrolesci", "id": 61748653, "node_id": "MDQ6VXNlcjYxNzQ4NjUz", "avatar_url": "https://avatars.githubusercontent.com/u/61748653?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pietrolesci", "html_url": "https://github.com/pietrolesci", "followers_url": "https://api.github.com/...
[ { "id": 1935892912, "node_id": "MDU6TGFiZWwxOTM1ODkyOTEy", "url": "https://api.github.com/repos/huggingface/datasets/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "Further information is requested" }, { "id": 2067400324, "node_id": "...
open
false
null
[]
null
[ "We're working on adding on-the-fly transforms in datasets.\r\nCurrently the only on-the-fly functions that can be applied are in `set_format` in which we transform the data in either numpy/torch/tf tensors or pandas.\r\nFor example\r\n```python\r\ndataset.set_format(\"torch\")\r\n```\r\napplies `torch.Tensor` to t...
1,607,187,776,000
1,610,534,888,000
null
NONE
null
Hi there, I have a question regarding "on-the-fly" tokenization. This question was elicited by reading the "How to train a new language model from scratch using Transformers and Tokenizers" [here](https://huggingface.co/blog/how-to-train). Towards the end there is this sentence: "If your dataset is very large, you c...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1167/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1167/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1115
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1115/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1115/comments
https://api.github.com/repos/huggingface/datasets/issues/1115/events
https://github.com/huggingface/datasets/issues/1115
757,127,527
MDU6SXNzdWU3NTcxMjc1Mjc=
1,115
Incorrect URL for MRQA SQuAD train subset
{ "login": "jimmycode", "id": 6259768, "node_id": "MDQ6VXNlcjYyNTk3Njg=", "avatar_url": "https://avatars.githubusercontent.com/u/6259768?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jimmycode", "html_url": "https://github.com/jimmycode", "followers_url": "https://api.github.com/users/ji...
[]
closed
false
null
[]
null
[ "good catch !" ]
1,607,090,724,000
1,607,274,862,000
1,607,274,862,000
CONTRIBUTOR
null
https://github.com/huggingface/datasets/blob/4ef4c8f8b7a60e35c6fa21115fca9faae91c9f74/datasets/mrqa/mrqa.py#L53 The URL for `train+SQuAD` subset of MRQA points to the dev set instead of train set. It should be `https://s3.us-east-2.amazonaws.com/mrqa/release/v2/train/SQuAD.jsonl.gz`.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1115/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1115/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1110
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1110/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1110/comments
https://api.github.com/repos/huggingface/datasets/issues/1110/events
https://github.com/huggingface/datasets/issues/1110
757,082,677
MDU6SXNzdWU3NTcwODI2Nzc=
1,110
Using a feature named "_type" fails with certain operations
{ "login": "dcfidalgo", "id": 15979778, "node_id": "MDQ6VXNlcjE1OTc5Nzc4", "avatar_url": "https://avatars.githubusercontent.com/u/15979778?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dcfidalgo", "html_url": "https://github.com/dcfidalgo", "followers_url": "https://api.github.com/users/...
[]
open
false
null
[]
null
[ "Thanks for reporting !\r\n\r\nIndeed this is a keyword in the library that is used to encode/decode features to a python dictionary that we can save/load to json.\r\nWe can probably change `_type` to something that is less likely to collide with user feature names.\r\nIn this case we would want something backward ...
1,607,086,593,000
1,610,535,208,000
null
CONTRIBUTOR
null
A column named `_type` leads to a `TypeError: unhashable type: 'dict'` for certain operations: ```python from datasets import Dataset, concatenate_datasets ds = Dataset.from_dict({"_type": ["whatever"]}).map() concatenate_datasets([ds]) # or simply Dataset(ds._data) ``` Context: We are using datasets to persi...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1110/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1110/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1103
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1103/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1103/comments
https://api.github.com/repos/huggingface/datasets/issues/1103/events
https://github.com/huggingface/datasets/issues/1103
757,016,820
MDU6SXNzdWU3NTcwMTY4MjA=
1,103
Add support to download kaggle datasets
{ "login": "abhishekkrthakur", "id": 1183441, "node_id": "MDQ6VXNlcjExODM0NDE=", "avatar_url": "https://avatars.githubusercontent.com/u/1183441?v=4", "gravatar_id": "", "url": "https://api.github.com/users/abhishekkrthakur", "html_url": "https://github.com/abhishekkrthakur", "followers_url": "https://ap...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Hey, I think this is great idea. Any plan to integrate kaggle private datasets loading to `datasets`?" ]
1,607,080,117,000
1,626,889,093,000
null
MEMBER
null
We can use API key
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1103/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1103/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1102
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1102/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1102/comments
https://api.github.com/repos/huggingface/datasets/issues/1102/events
https://github.com/huggingface/datasets/issues/1102
757,016,515
MDU6SXNzdWU3NTcwMTY1MTU=
1,102
Add retries to download manager
{ "login": "abhishekkrthakur", "id": 1183441, "node_id": "MDQ6VXNlcjExODM0NDE=", "avatar_url": "https://avatars.githubusercontent.com/u/1183441?v=4", "gravatar_id": "", "url": "https://api.github.com/users/abhishekkrthakur", "html_url": "https://github.com/abhishekkrthakur", "followers_url": "https://ap...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
{ "login": "SBrandeis", "id": 33657802, "node_id": "MDQ6VXNlcjMzNjU3ODAy", "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SBrandeis", "html_url": "https://github.com/SBrandeis", "followers_url": "https://api.github.com/users/...
[ { "login": "SBrandeis", "id": 33657802, "node_id": "MDQ6VXNlcjMzNjU3ODAy", "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SBrandeis", "html_url": "https://github.com/SBrandeis", "followers_url": "https://a...
null
[]
1,607,080,091,000
1,608,651,246,000
1,608,651,246,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1102/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1102/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1064
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1064/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1064/comments
https://api.github.com/repos/huggingface/datasets/issues/1064/events
https://github.com/huggingface/datasets/issues/1064
756,382,186
MDU6SXNzdWU3NTYzODIxODY=
1,064
Not support links with 302 redirect
{ "login": "chameleonTK", "id": 6429850, "node_id": "MDQ6VXNlcjY0Mjk4NTA=", "avatar_url": "https://avatars.githubusercontent.com/u/6429850?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chameleonTK", "html_url": "https://github.com/chameleonTK", "followers_url": "https://api.github.com/us...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODk...
closed
false
null
[]
null
[ "Hi !\r\nThis kind of links is now supported by the library since #1316", "> Hi !\r\n> This kind of links is now supported by the library since #1316\r\n\r\nI updated links in TLC datasets to be the github links in this pull request \r\n https://github.com/huggingface/datasets/pull/1737\r\n\r\nEverything works no...
1,607,015,083,000
1,610,592,685,000
1,610,592,685,000
CONTRIBUTOR
null
I have an issue adding this download link https://github.com/jitkapat/thailitcorpus/releases/download/v.2.0/tlc_v.2.0.tar.gz it might be because it is not a direct link (it returns 302 and redirects to aws that returns 403 for head requests). ``` r.head("https://github.com/jitkapat/thailitcorpus/releases/downlo...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1064/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1064/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1046
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1046/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1046/comments
https://api.github.com/repos/huggingface/datasets/issues/1046/events
https://github.com/huggingface/datasets/issues/1046
756,122,709
MDU6SXNzdWU3NTYxMjI3MDk=
1,046
Dataset.map() turns tensors into lists?
{ "login": "tombosc", "id": 5270804, "node_id": "MDQ6VXNlcjUyNzA4MDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/5270804?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tombosc", "html_url": "https://github.com/tombosc", "followers_url": "https://api.github.com/users/tombosc/...
[]
open
false
null
[]
null
[ "A solution is to have the tokenizer return a list instead of a tensor, and then use `dataset_tok.set_format(type = 'torch')` to convert that list into a tensor. Still not sure if bug.", "It is expected behavior, you should set the format to `\"torch\"` as you mentioned to get pytorch tensors back.\r\nBy default ...
1,606,995,826,000
1,608,731,472,000
null
NONE
null
I apply `Dataset.map()` to a function that returns a dict of torch tensors (like a tokenizer from the repo transformers). However, in the mapped dataset, these tensors have turned to lists! ```import datasets import torch from datasets import load_dataset ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1046/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1046/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1027
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1027/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1027/comments
https://api.github.com/repos/huggingface/datasets/issues/1027/events
https://github.com/huggingface/datasets/issues/1027
755,695,420
MDU6SXNzdWU3NTU2OTU0MjA=
1,027
Hi
{ "login": "suemori87", "id": 75398394, "node_id": "MDQ6VXNlcjc1Mzk4Mzk0", "avatar_url": "https://avatars.githubusercontent.com/u/75398394?v=4", "gravatar_id": "", "url": "https://api.github.com/users/suemori87", "html_url": "https://github.com/suemori87", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[]
1,606,952,834,000
1,607,013,761,000
1,607,013,761,000
NONE
null
## Adding a Dataset - **Name:** *name of the dataset* - **Description:** *short description of the dataset (or link to social media or blog post)* - **Paper:** *link to the dataset paper if available* - **Data:** *link to the Github repository or current dataset location* - **Motivation:** *what are some good reasons t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1027/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1027/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1026
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1026/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1026/comments
https://api.github.com/repos/huggingface/datasets/issues/1026/events
https://github.com/huggingface/datasets/issues/1026
755,689,195
MDU6SXNzdWU3NTU2ODkxOTU=
1,026
Lío o
{ "login": "Isaias0", "id": 73465581, "node_id": "MDQ6VXNlcjczNDY1NTgx", "avatar_url": "https://avatars.githubusercontent.com/u/73465581?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Isaias0", "html_url": "https://github.com/Isaias0", "followers_url": "https://api.github.com/users/Isaias...
[]
closed
false
null
[]
null
[]
1,606,951,945,000
1,607,013,767,000
1,607,013,767,000
NONE
null
````l````````` ``` O ``` ````` Ño ``` ```` ```
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1026/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1026/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/1004
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/1004/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/1004/comments
https://api.github.com/repos/huggingface/datasets/issues/1004/events
https://github.com/huggingface/datasets/issues/1004
755,325,368
MDU6SXNzdWU3NTUzMjUzNjg=
1,004
how large datasets are handled under the hood
{ "login": "rabeehkarimimahabadi", "id": 73364383, "node_id": "MDQ6VXNlcjczMzY0Mzgz", "avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehkarimimahabadi", "html_url": "https://github.com/rabeehkarimimahabadi", "followers_url...
[]
open
false
null
[]
null
[ "This library uses Apache Arrow under the hood to store datasets on disk.\r\nThe advantage of Apache Arrow is that it allows to memory map the dataset. This allows to load datasets bigger than memory and with almost no RAM usage. It also offers excellent I/O speed.\r\n\r\nFor example when you access one element or ...
1,606,919,560,000
1,612,175,031,000
null
NONE
null
Hi I want to use multiple large datasets with a mapping style dataloader, where they cannot fit into memory, could you tell me how you handled the datasets under the hood? is this you bring all in memory in case of mapping style ones? or is this some sharding under the hood and you bring in memory when necessary, than...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/1004/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/1004/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/996
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/996/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/996/comments
https://api.github.com/repos/huggingface/datasets/issues/996/events
https://github.com/huggingface/datasets/issues/996
755,176,084
MDU6SXNzdWU3NTUxNzYwODQ=
996
NotADirectoryError while loading the CNN/Dailymail dataset
{ "login": "arc-bu", "id": 75367920, "node_id": "MDQ6VXNlcjc1MzY3OTIw", "avatar_url": "https://avatars.githubusercontent.com/u/75367920?v=4", "gravatar_id": "", "url": "https://api.github.com/users/arc-bu", "html_url": "https://github.com/arc-bu", "followers_url": "https://api.github.com/users/arc-bu/fo...
[]
open
false
null
[]
null
[ "Looks like the google drive download failed.\r\nI'm getting a `Google Drive - Quota exceeded` error while looking at the downloaded file.\r\n\r\nWe should consider finding a better host than google drive for this dataset imo\r\nrelated : #873 #864 ", "It is working now, thank you. \r\n\r\nShould I leave this iss...
1,606,907,276,000
1,639,898,876,000
null
NONE
null
Downloading and preparing dataset cnn_dailymail/3.0.0 (download: 558.32 MiB, generated: 1.28 GiB, post-processed: Unknown size, total: 1.82 GiB) to /root/.cache/huggingface/datasets/cnn_dailymail/3.0.0/3.0.0/0128610a44e10f25b4af6689441c72af86205282d26399642f7db38fa7535602... ---------------------------------------...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/996/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/996/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/993
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/993/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/993/comments
https://api.github.com/repos/huggingface/datasets/issues/993/events
https://github.com/huggingface/datasets/issues/993
755,135,768
MDU6SXNzdWU3NTUxMzU3Njg=
993
Problem downloading amazon_reviews_multi
{ "login": "hfawaz", "id": 29229602, "node_id": "MDQ6VXNlcjI5MjI5NjAy", "avatar_url": "https://avatars.githubusercontent.com/u/29229602?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hfawaz", "html_url": "https://github.com/hfawaz", "followers_url": "https://api.github.com/users/hfawaz/fo...
[]
open
false
null
[]
null
[ "Hi @hfawaz ! This is working fine for me. Is it a repeated occurence? Have you tried from the latest verion?", "Hi, it seems a connection problem. \r\nNow it says: \r\n`ConnectionError: Couldn't reach https://amazon-reviews-ml.s3-us-west-2.amazonaws.com/json/train/dataset_ja_train.json`" ]
1,606,904,157,000
1,607,074,693,000
null
CONTRIBUTOR
null
Thanks for adding the dataset. After trying to load the dataset, I am getting the following error: `ConnectionError: Couldn't reach https://amazon-reviews-ml.s3-us-west-2.amazonaws.com/json/train/dataset_fr_train.json ` I used the following code to load the dataset: `load_dataset( dataset_name, ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/993/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/993/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/988
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/988/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/988/comments
https://api.github.com/repos/huggingface/datasets/issues/988/events
https://github.com/huggingface/datasets/issues/988
755,069,159
MDU6SXNzdWU3NTUwNjkxNTk=
988
making sure datasets are not loaded in memory and distributed training of them
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
open
false
null
[]
null
[ "my implementation of sharding per TPU core: https://github.com/google-research/ruse/blob/d4dd58a2d8efe0ffb1a9e9e77e3228d6824d3c3c/seq2seq/trainers/t5_trainer.py#L316 \r\nmy implementation of dataloader for this case https://github.com/google-research/ruse/blob/d4dd58a2d8efe0ffb1a9e9e77e3228d6824d3c3c/seq2seq/tasks...
1,606,898,715,000
1,606,899,034,000
null
CONTRIBUTOR
null
Hi I am dealing with large-scale datasets which I need to train distributedly, I used the shard function to divide the dataset across the cores, without any sampler, this does not work for distributed training and does not become any faster than 1 TPU core. 1) how I can make sure data is not loaded in memory 2) in cas...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/988/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/988/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/961
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/961/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/961/comments
https://api.github.com/repos/huggingface/datasets/issues/961/events
https://github.com/huggingface/datasets/issues/961
754,434,398
MDU6SXNzdWU3NTQ0MzQzOTg=
961
sample multiple datasets
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
open
false
null
[]
null
[ "here I share my dataloader currently for multiple tasks: https://gist.github.com/rabeehkarimimahabadi/39f9444a4fb6f53dcc4fca5d73bf8195 \r\n\r\nI need to train my model distributedly with this dataloader, \"MultiTasksataloader\", currently this does not work in distributed fasion,\r\nto save on memory I tried to us...
1,606,832,402,000
1,606,872,764,000
null
CONTRIBUTOR
null
Hi I am dealing with multiple datasets, I need to have a dataloader over them with a condition that in each batch data samples are coming from one of the datasets. My main question is: - I need to have a way to sample the datasets first with some weights, lets say 2x dataset1 1x dataset2, could you point me how I c...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/961/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/961/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/942
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/942/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/942/comments
https://api.github.com/repos/huggingface/datasets/issues/942/events
https://github.com/huggingface/datasets/issues/942
754,162,318
MDU6SXNzdWU3NTQxNjIzMTg=
942
D
{ "login": "CryptoMiKKi", "id": 74238514, "node_id": "MDQ6VXNlcjc0MjM4NTE0", "avatar_url": "https://avatars.githubusercontent.com/u/74238514?v=4", "gravatar_id": "", "url": "https://api.github.com/users/CryptoMiKKi", "html_url": "https://github.com/CryptoMiKKi", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[]
1,606,810,630,000
1,607,013,773,000
1,607,013,773,000
NONE
null
## Adding a Dataset - **Name:** *name of the dataset* - **Description:** *short description of the dataset (or link to social media or blog post)* - **Paper:** *link to the dataset paper if available* - **Data:** *link to the Github repository or current dataset location* - **Motivation:** *what are some good reasons t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/942/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/942/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/937
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/937/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/937/comments
https://api.github.com/repos/huggingface/datasets/issues/937/events
https://github.com/huggingface/datasets/issues/937
753,921,078
MDU6SXNzdWU3NTM5MjEwNzg=
937
Local machine/cluster Beam Datasets example/tutorial
{ "login": "shangw-nvidia", "id": 66387198, "node_id": "MDQ6VXNlcjY2Mzg3MTk4", "avatar_url": "https://avatars.githubusercontent.com/u/66387198?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shangw-nvidia", "html_url": "https://github.com/shangw-nvidia", "followers_url": "https://api.githu...
[]
open
false
null
[]
null
[ "I tried to make it run once on the SparkRunner but it seems that this runner has some issues when it is run locally.\r\nFrom my experience the DirectRunner is fine though, even if it's clearly not memory efficient.\r\n\r\nIt would be awesome though to make it work locally on a SparkRunner !\r\nDid you manage to ma...
1,606,785,103,000
1,608,731,696,000
null
NONE
null
Hi, I'm wondering if https://huggingface.co/docs/datasets/beam_dataset.html has an non-GCP or non-Dataflow version example/tutorial? I tried to migrate it to run on DirectRunner and SparkRunner, however, there were way too many runtime errors that I had to fix during the process, and even so I wasn't able to get eit...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/937/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/937/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/927
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/927/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/927/comments
https://api.github.com/repos/huggingface/datasets/issues/927/events
https://github.com/huggingface/datasets/issues/927
753,679,020
MDU6SXNzdWU3NTM2NzkwMjA=
927
Hello
{ "login": "k125-ak", "id": 75259546, "node_id": "MDQ6VXNlcjc1MjU5NTQ2", "avatar_url": "https://avatars.githubusercontent.com/u/75259546?v=4", "gravatar_id": "", "url": "https://api.github.com/users/k125-ak", "html_url": "https://github.com/k125-ak", "followers_url": "https://api.github.com/users/k125-a...
[]
closed
false
null
[]
null
[]
1,606,758,605,000
1,606,758,630,000
1,606,758,630,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/927/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/927/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/919
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/919/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/919/comments
https://api.github.com/repos/huggingface/datasets/issues/919/events
https://github.com/huggingface/datasets/issues/919
753,434,472
MDU6SXNzdWU3NTM0MzQ0NzI=
919
wrong length with datasets
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
closed
false
null
[]
null
[ "Also, I cannot first convert it to torch format, since huggingface seq2seq_trainer codes process the datasets afterwards during datacollector function to make it optimize for TPUs. ", "sorry I misunderstood length of dataset with dataloader, closed. thanks " ]
1,606,739,019,000
1,606,739,847,000
1,606,739,846,000
CONTRIBUTOR
null
Hi I have a MRPC dataset which I convert it to seq2seq format, then this is of this format: `Dataset(features: {'src_texts': Value(dtype='string', id=None), 'tgt_texts': Value(dtype='string', id=None)}, num_rows: 10) ` I feed it to a dataloader: ``` dataloader = DataLoader( train_dataset, ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/919/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/919/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/915
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/915/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/915/comments
https://api.github.com/repos/huggingface/datasets/issues/915/events
https://github.com/huggingface/datasets/issues/915
753,118,481
MDU6SXNzdWU3NTMxMTg0ODE=
915
Shall we change the hashing to encoding to reduce potential replicated cache files?
{ "login": "zhuzilin", "id": 10428324, "node_id": "MDQ6VXNlcjEwNDI4MzI0", "avatar_url": "https://avatars.githubusercontent.com/u/10428324?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zhuzilin", "html_url": "https://github.com/zhuzilin", "followers_url": "https://api.github.com/users/zhu...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 2067400324, "node_id": "MDU6...
open
false
null
[]
null
[ "This is an interesting idea !\r\nDo you have ideas about how to approach the decoding and the normalization ?", "@lhoestq\r\nI think we first need to save the transformation chain to a list in `self._fingerprint`. Then we can\r\n- decode all the current saved datasets to see if there is already one that is equiv...
1,606,708,246,000
1,608,786,709,000
null
NONE
null
Hi there. For now, we are using `xxhash` to hash the transformations to fingerprint and we will save a copy of the processed dataset to disk if there is a new hash value. However, there are some transformations that are idempotent or commutative to each other. I think that encoding the transformation chain as the finge...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/915/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/915/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/911
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/911/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/911/comments
https://api.github.com/repos/huggingface/datasets/issues/911/events
https://github.com/huggingface/datasets/issues/911
752,806,215
MDU6SXNzdWU3NTI4MDYyMTU=
911
datasets module not found
{ "login": "sbassam", "id": 15836274, "node_id": "MDQ6VXNlcjE1ODM2Mjc0", "avatar_url": "https://avatars.githubusercontent.com/u/15836274?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sbassam", "html_url": "https://github.com/sbassam", "followers_url": "https://api.github.com/users/sbassa...
[]
closed
false
null
[]
null
[ "nvm, I'd made an assumption that the library gets installed with transformers. " ]
1,606,613,055,000
1,606,660,389,000
1,606,660,389,000
NONE
null
Currently, running `from datasets import load_dataset` will throw a `ModuleNotFoundError: No module named 'datasets'` error.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/911/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/911/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/910
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/910/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/910/comments
https://api.github.com/repos/huggingface/datasets/issues/910/events
https://github.com/huggingface/datasets/issues/910
752,772,723
MDU6SXNzdWU3NTI3NzI3MjM=
910
Grindr meeting app web.Grindr
{ "login": "jackin34", "id": 75184749, "node_id": "MDQ6VXNlcjc1MTg0NzQ5", "avatar_url": "https://avatars.githubusercontent.com/u/75184749?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jackin34", "html_url": "https://github.com/jackin34", "followers_url": "https://api.github.com/users/jac...
[]
closed
false
null
[]
null
[]
1,606,599,383,000
1,606,644,711,000
1,606,644,711,000
NONE
null
## Adding a Dataset - **Name:** *name of the dataset* - **Description:** *short description of the dataset (or link to social media or blog post)* - **Paper:** *link to the dataset paper if available* - **Data:** *link to the Github repository or current dataset location* - **Motivation:** *what are some good reasons t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/910/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/910/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/900
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/900/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/900/comments
https://api.github.com/repos/huggingface/datasets/issues/900/events
https://github.com/huggingface/datasets/issues/900
752,214,066
MDU6SXNzdWU3NTIyMTQwNjY=
900
datasets.load_dataset() custom chaching directory bug
{ "login": "SapirWeissbuch", "id": 44585792, "node_id": "MDQ6VXNlcjQ0NTg1Nzky", "avatar_url": "https://avatars.githubusercontent.com/u/44585792?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SapirWeissbuch", "html_url": "https://github.com/SapirWeissbuch", "followers_url": "https://api.gi...
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "Thanks for reporting ! I'm looking into it." ]
1,606,479,533,000
1,606,690,133,000
1,606,690,133,000
NONE
null
Hello, I'm having issue with loading a dataset with a custom `cache_dir`. Despite specifying the output dir, it is still downloaded to `~/.cache`. ## Environment info - `datasets` version: 1.1.3 - Platform: Linux-4.19.129-aufs-1-x86_64-with-debian-10.1 - Python version: 3.7.3 ## The code I'm running: ```p...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/900/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/900/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/897
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/897/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/897/comments
https://api.github.com/repos/huggingface/datasets/issues/897/events
https://github.com/huggingface/datasets/issues/897
752,100,256
MDU6SXNzdWU3NTIxMDAyNTY=
897
Dataset viewer issues
{ "login": "BramVanroy", "id": 2779410, "node_id": "MDQ6VXNlcjI3Nzk0MTA=", "avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BramVanroy", "html_url": "https://github.com/BramVanroy", "followers_url": "https://api.github.com/users...
[ { "id": 2107841032, "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer", "name": "nlp-viewer", "color": "94203D", "default": false, "description": "" } ]
closed
false
null
[]
null
[ "Thanks for reporting !\r\ncc @srush for the empty feature list issue and the encoding issue\r\ncc @julien-c maybe we can update the url and just have a redirection from the old url to the new one ?", "Ok, I redirected on our side to a new url. ⚠️ @srush: if you update the Streamlit config too to `/datasets/viewe...
1,606,468,474,000
1,635,671,521,000
1,635,671,521,000
CONTRIBUTOR
null
I was looking through the dataset viewer and I like it a lot. Version numbers, citation information, everything's there! I've spotted a few issues/bugs though: - the URL is still under `nlp`, perhaps an alias for `datasets` can be made - when I remove a **feature** (and the feature list is empty), I get an error. T...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/897/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/897/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/888
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/888/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/888/comments
https://api.github.com/repos/huggingface/datasets/issues/888/events
https://github.com/huggingface/datasets/issues/888
750,944,422
MDU6SXNzdWU3NTA5NDQ0MjI=
888
Nested lists are zipped unexpectedly
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[]
closed
false
null
[]
null
[ "Yes following the Tensorflow Datasets convention, objects with type `Sequence of a Dict` are actually stored as a `dictionary of lists`.\r\nSee the [documentation](https://huggingface.co/docs/datasets/features.html?highlight=features) for more details", "Thanks.\r\nThis is a bit (very) confusing, but I guess if ...
1,606,320,466,000
1,606,325,439,000
1,606,325,439,000
CONTRIBUTOR
null
I might misunderstand something, but I expect that if I define: ```python "top": datasets.features.Sequence({ "middle": datasets.features.Sequence({ "bottom": datasets.Value("int32") }) }) ``` And I then create an example: ```python yield 1, { "top": [{ "middle": [ {"bottom": 1}, ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/888/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/888/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/887
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/887/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/887/comments
https://api.github.com/repos/huggingface/datasets/issues/887/events
https://github.com/huggingface/datasets/issues/887
750,868,831
MDU6SXNzdWU3NTA4Njg4MzE=
887
pyarrow.lib.ArrowNotImplementedError: MakeBuilder: cannot construct builder for type extension<arrow.py_extension_type>
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
[ "Yes right now `ArrayXD` can only be used as a column feature type, not a subtype.\r\nWith the current Arrow limitations I don't think we'll be able to make it work as a subtype, however it should be possible to allow dimensions of dynamic sizes (`Array3D(shape=(None, 137, 2), dtype=\"float32\")` for example since ...
1,606,314,741,000
1,631,207,020,000
null
CONTRIBUTOR
null
I set up a new dataset, with a sequence of arrays (really, I want to have an array of (None, 137, 2), and the first dimension is dynamic) ```python def _info(self): return datasets.DatasetInfo( description=_DESCRIPTION, # This defines the different columns of the dataset and ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/887/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/887/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/885
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/885/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/885/comments
https://api.github.com/repos/huggingface/datasets/issues/885/events
https://github.com/huggingface/datasets/issues/885
750,789,052
MDU6SXNzdWU3NTA3ODkwNTI=
885
Very slow cold-start
{ "login": "AmitMY", "id": 5757359, "node_id": "MDQ6VXNlcjU3NTczNTk=", "avatar_url": "https://avatars.githubusercontent.com/u/5757359?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AmitMY", "html_url": "https://github.com/AmitMY", "followers_url": "https://api.github.com/users/AmitMY/foll...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "Good point!", "Yes indeed. We can probably improve that by using lazy imports", "#1690 added fast start-up of the library " ]
1,606,308,478,000
1,610,537,485,000
1,610,537,485,000
CONTRIBUTOR
null
Hi, I expect when importing `datasets` that nothing major happens in the background, and so the import should be insignificant. When I load a metric, or a dataset, its fine that it takes time. The following ranges from 3 to 9 seconds: ``` python -m timeit -n 1 -r 1 'from datasets import load_dataset' ``` edi...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/885/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/885/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/883
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/883/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/883/comments
https://api.github.com/repos/huggingface/datasets/issues/883/events
https://github.com/huggingface/datasets/issues/883
749,750,801
MDU6SXNzdWU3NDk3NTA4MDE=
883
Downloading/caching only a part of a datasets' dataset.
{ "login": "SapirWeissbuch", "id": 44585792, "node_id": "MDQ6VXNlcjQ0NTg1Nzky", "avatar_url": "https://avatars.githubusercontent.com/u/44585792?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SapirWeissbuch", "html_url": "https://github.com/SapirWeissbuch", "followers_url": "https://api.gi...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 1935892912, "node_id": "MDU6...
open
false
null
[]
null
[ "Not at the moment but we could likely support this feature.", "?", "I think it would be a very helpful feature, because sometimes one only wants to evaluate models on the dev set, and the whole training data may be many times bigger.\r\nThis makes the task impossible with limited memory resources." ]
1,606,227,918,000
1,606,485,115,000
null
NONE
null
Hi, I want to use the validation data *only* (of natural question). I don't want to have the whole dataset cached in my machine, just the dev set. Is this possible? I can't find a way to do it in the docs. Thank you, Sapir
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/883/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/883/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/880
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/880/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/880/comments
https://api.github.com/repos/huggingface/datasets/issues/880/events
https://github.com/huggingface/datasets/issues/880
748,949,606
MDU6SXNzdWU3NDg5NDk2MDY=
880
Add SQA
{ "login": "NielsRogge", "id": 48327001, "node_id": "MDQ6VXNlcjQ4MzI3MDAx", "avatar_url": "https://avatars.githubusercontent.com/u/48327001?v=4", "gravatar_id": "", "url": "https://api.github.com/users/NielsRogge", "html_url": "https://github.com/NielsRogge", "followers_url": "https://api.github.com/use...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "I’ll take this one to test the workflow for the sprint next week cc @yjernite @lhoestq ", "@thomwolf here's a slightly adapted version of the code from the [official Tapas repository](https://github.com/google-research/tapas/blob/master/tapas/utils/interaction_utils.py) that is used to turn the `answer_coordinat...
1,606,149,115,000
1,608,731,904,000
1,608,731,903,000
NONE
null
## Adding a Dataset - **Name:** SQA (Sequential Question Answering) by Microsoft. - **Description:** The SQA dataset was created to explore the task of answering sequences of inter-related questions on HTML tables. It has 6,066 sequences with 17,553 questions in total. - **Paper:** https://www.microsoft.com/en-us/r...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/880/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/880/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/879
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/879/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/879/comments
https://api.github.com/repos/huggingface/datasets/issues/879/events
https://github.com/huggingface/datasets/issues/879
748,848,847
MDU6SXNzdWU3NDg4NDg4NDc=
879
boolq does not load
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
null
[]
null
[ "Hi ! It runs on my side without issues. I tried\r\n```python\r\nfrom datasets import load_dataset\r\nload_dataset(\"boolq\")\r\n```\r\n\r\nWhat version of datasets and tensorflow are your runnning ?\r\nAlso if you manage to get a minimal reproducible script (on google colab for example) that would be useful.", "...
1,606,141,708,000
1,606,485,071,000
null
CONTRIBUTOR
null
Hi I am getting these errors trying to load boolq thanks Traceback (most recent call last): File "test.py", line 5, in <module> data = AutoTask().get("boolq").get_dataset("train", n_obs=10) File "/remote/idiap.svm/user.active/rkarimi/dev/internship/seq2seq/tasks/tasks.py", line 42, in get_dataset d...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/879/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/879/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/878
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/878/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/878/comments
https://api.github.com/repos/huggingface/datasets/issues/878/events
https://github.com/huggingface/datasets/issues/878
748,621,981
MDU6SXNzdWU3NDg2MjE5ODE=
878
Loading Data From S3 Path in Sagemaker
{ "login": "mahesh1amour", "id": 42795522, "node_id": "MDQ6VXNlcjQyNzk1NTIy", "avatar_url": "https://avatars.githubusercontent.com/u/42795522?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mahesh1amour", "html_url": "https://github.com/mahesh1amour", "followers_url": "https://api.github.c...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 1935892912, "node_id": "MDU6...
open
false
null
[]
null
[ "This would be a neat feature", "> neat feature\r\n\r\nI dint get these clearly, can you please elaborate like how to work on these ", "It could maybe work almost out of the box just by using `cached_path` in the text/csv/json scripts, no?", "Thanks thomwolf and julien-c\r\n\r\nI'm still confusion on what you...
1,606,123,042,000
1,608,717,188,000
null
NONE
null
In Sagemaker Im tring to load the data set from S3 path as follows `train_path = 's3://xxxxxxxxxx/xxxxxxxxxx/train.csv' valid_path = 's3://xxxxxxxxxx/xxxxxxxxxx/validation.csv' test_path = 's3://xxxxxxxxxx/xxxxxxxxxx/test.csv' data_files = {} data_files["train"] = train_path data_files...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/878/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/878/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/877
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/877/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/877/comments
https://api.github.com/repos/huggingface/datasets/issues/877/events
https://github.com/huggingface/datasets/issues/877
748,234,438
MDU6SXNzdWU3NDgyMzQ0Mzg=
877
DataLoader(datasets) become more and more slowly within iterations
{ "login": "shexuan", "id": 25664170, "node_id": "MDQ6VXNlcjI1NjY0MTcw", "avatar_url": "https://avatars.githubusercontent.com/u/25664170?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shexuan", "html_url": "https://github.com/shexuan", "followers_url": "https://api.github.com/users/shexua...
[]
closed
false
null
[]
null
[ "Hi ! Thanks for reporting.\r\nDo you have the same slowdown when you iterate through the raw dataset object as well ? (no dataloader)\r\nIt would be nice to know whether it comes from the dataloader or not", "> Hi ! Thanks for reporting.\r\n> Do you have the same slowdown when you iterate through the raw dataset...
1,606,048,870,000
1,606,664,712,000
1,606,664,712,000
NONE
null
Hello, when I for loop my dataloader, the loading speed is becoming more and more slowly! ``` dataset = load_from_disk(dataset_path) # around 21,000,000 lines lineloader = tqdm(DataLoader(dataset, batch_size=1)) for idx, line in enumerate(lineloader): # do some thing for each line ``` In the begining, th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/877/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/877/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/876
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/876/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/876/comments
https://api.github.com/repos/huggingface/datasets/issues/876/events
https://github.com/huggingface/datasets/issues/876
748,195,104
MDU6SXNzdWU3NDgxOTUxMDQ=
876
imdb dataset cannot be loaded
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
closed
false
null
[]
null
[ "It looks like there was an issue while building the imdb dataset.\r\nCould you provide more information about your OS and the version of python and `datasets` ?\r\n\r\nAlso could you try again with \r\n```python\r\ndataset = datasets.load_dataset(\"imdb\", split=\"train\", download_mode=\"force_redownload\")\r\n``...
1,606,033,483,000
1,637,924,836,000
1,608,831,527,000
CONTRIBUTOR
null
Hi I am trying to load the imdb train dataset `dataset = datasets.load_dataset("imdb", split="train")` getting following errors, thanks for your help ``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/876/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/876/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/875
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/875/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/875/comments
https://api.github.com/repos/huggingface/datasets/issues/875/events
https://github.com/huggingface/datasets/issues/875
748,194,311
MDU6SXNzdWU3NDgxOTQzMTE=
875
bug in boolq dataset loading
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
closed
false
null
[]
null
[ "I just opened a PR to fix this.\r\nThanks for reporting !" ]
1,606,033,114,000
1,606,212,753,000
1,606,212,753,000
CONTRIBUTOR
null
Hi I am trying to load boolq dataset: ``` import datasets datasets.load_dataset("boolq") ``` I am getting the following errors, thanks for your help ``` >>> import datasets 2020-11-22 09:16:30.070332: W tensorflow/stream_executor/platform/default/dso_loader.cc:60] Could not load dynamic library 'libcuda...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/875/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/875/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/874
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/874/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/874/comments
https://api.github.com/repos/huggingface/datasets/issues/874/events
https://github.com/huggingface/datasets/issues/874
748,193,140
MDU6SXNzdWU3NDgxOTMxNDA=
874
trec dataset unavailable
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
closed
false
null
[]
null
[ "This was fixed in #740 \r\nCould you try to update `datasets` and try again ?", "This has been fixed in datasets 1.1.3" ]
1,606,032,576,000
1,606,485,402,000
1,606,485,402,000
CONTRIBUTOR
null
Hi when I try to load the trec dataset I am getting these errors, thanks for your help `datasets.load_dataset("trec", split="train") ` ``` File "<stdin>", line 1, in <module> File "/idiap/user/rkarimi/libs/anaconda3/envs/internship/lib/python3.7/site-packages/datasets/load.py", line 611, in load_dataset ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/874/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/874/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/873
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/873/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/873/comments
https://api.github.com/repos/huggingface/datasets/issues/873/events
https://github.com/huggingface/datasets/issues/873
747,959,523
MDU6SXNzdWU3NDc5NTk1MjM=
873
load_dataset('cnn_dalymail', '3.0.0') gives a 'Not a directory' error
{ "login": "vishal-burman", "id": 19861874, "node_id": "MDQ6VXNlcjE5ODYxODc0", "avatar_url": "https://avatars.githubusercontent.com/u/19861874?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vishal-burman", "html_url": "https://github.com/vishal-burman", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "I get the same error. It was fixed some days ago, but again it appears", "Hi @mrm8488 it's working again today without any fix so I am closing this issue.", "I see the issue happening again today - \r\n\r\n[nltk_data] Downloading package stopwords to /root/nltk_data...\r\n[nltk_data] Package stopwords is alr...
1,605,940,245,000
1,606,993,455,000
1,606,047,485,000
NONE
null
``` from datasets import load_dataset dataset = load_dataset('cnn_dailymail', '3.0.0') ``` Stack trace: ``` --------------------------------------------------------------------------- NotADirectoryError Traceback (most recent call last) <ipython-input-6-2e06a8332652> in <module>() ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/873/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/873/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/871
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/871/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/871/comments
https://api.github.com/repos/huggingface/datasets/issues/871/events
https://github.com/huggingface/datasets/issues/871
747,470,136
MDU6SXNzdWU3NDc0NzAxMzY=
871
terminate called after throwing an instance of 'google::protobuf::FatalException'
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
closed
false
null
[]
null
[ "Loading the iwslt2017-en-nl config of iwslt2017 works fine on my side. \r\nMaybe you can open an issue on transformers as well ? And also add more details about your environment (OS, python version, version of transformers and datasets etc.)", "closing now, figured out this is because the max length of decoder w...
1,605,876,984,000
1,607,807,792,000
1,607,807,792,000
CONTRIBUTOR
null
Hi I am using the dataset "iwslt2017-en-nl", and after downloading it I am getting this error when trying to evaluate it on T5-base with seq2seq_trainer.py in the huggingface repo could you assist me please? thanks 100%|█████████████████████████████████████████████████████████████████████████████████████████████...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/871/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/871/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/870
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/870/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/870/comments
https://api.github.com/repos/huggingface/datasets/issues/870/events
https://github.com/huggingface/datasets/issues/870
747,021,996
MDU6SXNzdWU3NDcwMjE5OTY=
870
[Feature Request] Add optional parameter in text loading script to preserve linebreaks
{ "login": "jncasey", "id": 31020859, "node_id": "MDQ6VXNlcjMxMDIwODU5", "avatar_url": "https://avatars.githubusercontent.com/u/31020859?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jncasey", "html_url": "https://github.com/jncasey", "followers_url": "https://api.github.com/users/jncase...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "Hi ! Thanks for your message.\r\nIndeed it's a free feature we can add and that can be useful.\r\nIf you want to contribute, feel free to open a PR to add it to the text dataset script :)" ]
1,605,829,891,000
1,606,484,891,000
null
NONE
null
I'm working on a project about rhyming verse using phonetic poetry and song lyrics, and line breaks are a vital part of the data. I recently switched over to use the datasets library when my various corpora grew larger than my computer's memory. And so far, it is SO great. But the first time I processed all of ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/870/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/870/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/866
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/866/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/866/comments
https://api.github.com/repos/huggingface/datasets/issues/866/events
https://github.com/huggingface/datasets/issues/866
745,719,222
MDU6SXNzdWU3NDU3MTkyMjI=
866
OSCAR from Inria group
{ "login": "jchwenger", "id": 34098722, "node_id": "MDQ6VXNlcjM0MDk4NzIy", "avatar_url": "https://avatars.githubusercontent.com/u/34098722?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jchwenger", "html_url": "https://github.com/jchwenger", "followers_url": "https://api.github.com/users/...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "PR is already open here : #348 \r\nThe only thing remaining is to compute the metadata of each subdataset (one per language + shuffled/unshuffled).\r\nAs soon as #863 is merged we can start computing them. This will take a bit of time though", "Grand, thanks for this!" ]
1,605,710,454,000
1,605,711,690,000
1,605,711,690,000
NONE
null
## Adding a Dataset - **Name:** *OSCAR* (Open Super-large Crawled ALMAnaCH coRpus), multilingual parsing of Common Crawl (separate crawls for many different languages), [here](https://oscar-corpus.com/). - **Description:** *OSCAR or Open Super-large Crawled ALMAnaCH coRpus is a huge multilingual corpus obtained by la...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/866/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/866/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/865
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/865/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/865/comments
https://api.github.com/repos/huggingface/datasets/issues/865/events
https://github.com/huggingface/datasets/issues/865
745,430,497
MDU6SXNzdWU3NDU0MzA0OTc=
865
Have Trouble importing `datasets`
{ "login": "forest1988", "id": 2755894, "node_id": "MDQ6VXNlcjI3NTU4OTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2755894?v=4", "gravatar_id": "", "url": "https://api.github.com/users/forest1988", "html_url": "https://github.com/forest1988", "followers_url": "https://api.github.com/users...
[]
closed
false
null
[]
null
[ "I'm sorry, this was a problem with my environment.\r\nNow that I have identified the cause of environmental dependency, I would like to fix it and try it.\r\nExcuse me for making a noise." ]
1,605,686,681,000
1,605,687,395,000
1,605,687,395,000
CONTRIBUTOR
null
I'm failing to import transformers (v4.0.0-dev), and tracing the cause seems to be failing to import datasets. I cloned the newest version of datasets (master branch), and do `pip install -e .`. Then, `import datasets` causes the error below. ``` ~/workspace/Clone/datasets/src/datasets/utils/file_utils.py in ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/865/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/865/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/864
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/864/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/864/comments
https://api.github.com/repos/huggingface/datasets/issues/864/events
https://github.com/huggingface/datasets/issues/864
745,322,357
MDU6SXNzdWU3NDUzMjIzNTc=
864
Unable to download cnn_dailymail dataset
{ "login": "rohitashwa1907", "id": 46031058, "node_id": "MDQ6VXNlcjQ2MDMxMDU4", "avatar_url": "https://avatars.githubusercontent.com/u/46031058?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rohitashwa1907", "html_url": "https://github.com/rohitashwa1907", "followers_url": "https://api.gi...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "Same error here!\r\n", "Same here! My kaggle notebook stopped working like yesterday. It's strange because I have fixed version of datasets==1.1.2", "I'm looking at it right now", "I couldn't reproduce unfortunately. I tried\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nload_dataset(\"cnn_dailymai...
1,605,674,282,000
1,605,849,731,000
1,605,849,730,000
NONE
null
### Script to reproduce the error ``` from datasets import load_dataset train_dataset = load_dataset("cnn_dailymail", "3.0.0", split= 'train[:10%') valid_dataset = load_dataset("cnn_dailymail","3.0.0", split="validation[:5%]") ``` ### Error ``` -------------------------------------------------------------...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/864/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/864/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/861
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/861/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/861/comments
https://api.github.com/repos/huggingface/datasets/issues/861/events
https://github.com/huggingface/datasets/issues/861
744,753,458
MDU6SXNzdWU3NDQ3NTM0NTg=
861
Possible Bug: Small training/dataset file creates gigantic output
{ "login": "NebelAI", "id": 7240417, "node_id": "MDQ6VXNlcjcyNDA0MTc=", "avatar_url": "https://avatars.githubusercontent.com/u/7240417?v=4", "gravatar_id": "", "url": "https://api.github.com/users/NebelAI", "html_url": "https://github.com/NebelAI", "followers_url": "https://api.github.com/users/NebelAI/...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 1935892912, "node_id": "MDU6...
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "The preprocessing tokenizes the input text. Tokenization outputs `input_ids`, `attention_mask`, `token_type_ids` and `special_tokens_mask`. All those are of length`max_seq_length` because of padding. Therefore for each sample it generate 4 *`max_seq_length` integers. Currently they're all saved as int64. This is w...
1,605,620,939,000
1,617,113,044,000
1,616,414,695,000
NONE
null
Hey guys, I was trying to create a new bert model from scratch via _huggingface transformers + tokenizers + dataets_ (actually using this example script by your team: https://github.com/huggingface/transformers/blob/master/examples/language-modeling/run_mlm.py). It was supposed to be a first test with a small 5 GB r...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/861/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/861/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/860
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/860/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/860/comments
https://api.github.com/repos/huggingface/datasets/issues/860/events
https://github.com/huggingface/datasets/issues/860
744,750,691
MDU6SXNzdWU3NDQ3NTA2OTE=
860
wmt16 cs-en does not donwload
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
null
[]
null
[]
1,605,620,735,000
1,606,484,824,000
null
CONTRIBUTOR
null
Hi I am trying with wmt16, cs-en pair, thanks for the help, perhaps similar to the ro-en issue. thanks split="train", n_obs=data_args.n_train) for task in data_args.task} File "finetune_t5_trainer.py", line 109, in <dictcomp> split="train", n_obs=data_args.n_train) for task in data_args.task} File "/hom...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/860/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/860/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/854
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/854/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/854/comments
https://api.github.com/repos/huggingface/datasets/issues/854/events
https://github.com/huggingface/datasets/issues/854
743,675,376
MDU6SXNzdWU3NDM2NzUzNzY=
854
wmt16 does not download
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
null
[]
null
[ "Hi,I also posted it to the forum, but this is a bug, perhaps it needs to be reported here? thanks ", "It looks like the official OPUS server for WMT16 doesn't provide the data files anymore (503 error).\r\nI searched a bit and couldn't find a mirror except maybe http://nlp.ffzg.hr/resources/corpora/setimes/ (the...
1,605,519,111,000
1,614,222,909,000
null
CONTRIBUTOR
null
Hi, I appreciate your help with the following error, thanks >>> from datasets import load_dataset >>> dataset = load_dataset("wmt16", "ro-en", split="train") Downloading and preparing dataset wmt16/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to /root/...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/854/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/854/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/853
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/853/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/853/comments
https://api.github.com/repos/huggingface/datasets/issues/853/events
https://github.com/huggingface/datasets/issues/853
743,426,583
MDU6SXNzdWU3NDM0MjY1ODM=
853
concatenate_datasets support axis=0 or 1 ?
{ "login": "renqingcolin", "id": 12437751, "node_id": "MDQ6VXNlcjEyNDM3NzUx", "avatar_url": "https://avatars.githubusercontent.com/u/12437751?v=4", "gravatar_id": "", "url": "https://api.github.com/users/renqingcolin", "html_url": "https://github.com/renqingcolin", "followers_url": "https://api.github.c...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 1935892884, "node_id": "MDU6...
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Unfortunately `concatenate_datasets` only supports concatenating the rows, while what you want to achieve is concatenate the columns.\r\nCurrently to add more columns to a dataset, one must use `map`.\r\nWhat you can do is somehting like this:\r\n```python\r\n# suppose you have datasets d1, d2, d3\r\ndef add_colum...
1,605,494,783,000
1,618,848,438,000
1,618,848,438,000
NONE
null
I want to achieve the following result ![image](https://user-images.githubusercontent.com/12437751/99207426-f0c8db80-27f8-11eb-820a-4d9f7287b742.png)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/853/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/853/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/852
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/852/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/852/comments
https://api.github.com/repos/huggingface/datasets/issues/852/events
https://github.com/huggingface/datasets/issues/852
743,396,240
MDU6SXNzdWU3NDMzOTYyNDA=
852
wmt cannot be downloaded
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[]
1,605,488,681,000
1,605,519,118,000
1,605,519,118,000
CONTRIBUTOR
null
Hi, I appreciate your help with the following error, thanks >>> from datasets import load_dataset >>> dataset = load_dataset("wmt16", "ro-en", split="train") Downloading and preparing dataset wmt16/ro-en (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to /root/...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/852/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/852/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/851
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/851/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/851/comments
https://api.github.com/repos/huggingface/datasets/issues/851/events
https://github.com/huggingface/datasets/issues/851
743,343,278
MDU6SXNzdWU3NDMzNDMyNzg=
851
Add support for other languages for rouge
{ "login": "alexyalunin", "id": 23011284, "node_id": "MDQ6VXNlcjIzMDExMjg0", "avatar_url": "https://avatars.githubusercontent.com/u/23011284?v=4", "gravatar_id": "", "url": "https://api.github.com/users/alexyalunin", "html_url": "https://github.com/alexyalunin", "followers_url": "https://api.github.com/...
[ { "id": 2067400959, "node_id": "MDU6TGFiZWwyMDY3NDAwOTU5", "url": "https://api.github.com/repos/huggingface/datasets/labels/Metric%20discussion", "name": "Metric discussion", "color": "d722e8", "default": false, "description": "Discussions on the metrics" } ]
open
false
null
[]
null
[ "@alexyalunin \r\n\r\nI did something similar for others languages.\r\n\r\n[Repo: rouge-metric](https://github.com/m3hrdadfi/rouge-metric)" ]
1,605,473,865,000
1,622,970,472,000
null
NONE
null
I calculate rouge with ``` from datasets import load_metric rouge = load_metric("rouge") rouge_output = rouge.compute(predictions=['тест тест привет'], references=['тест тест пока'], rouge_types=[ "rouge2"])["rouge2"].mid print(rouge_output) ``` the result is `Score(precision=0.0, recall=0.0, fmeasure=0.0)...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/851/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/851/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/849
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/849/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/849/comments
https://api.github.com/repos/huggingface/datasets/issues/849/events
https://github.com/huggingface/datasets/issues/849
742,263,333
MDU6SXNzdWU3NDIyNjMzMzM=
849
Load amazon dataset
{ "login": "bhavitvyamalik", "id": 19718818, "node_id": "MDQ6VXNlcjE5NzE4ODE4", "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bhavitvyamalik", "html_url": "https://github.com/bhavitvyamalik", "followers_url": "https://api.gi...
[]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nWe plan to show information about the different configs of the datasets on the website, with the corresponding `load_dataset` calls.\r\n\r\nAlso I think the bullet points formatting has been fixed" ]
1,605,256,464,000
1,605,597,779,000
1,605,597,779,000
CONTRIBUTOR
null
Hi, I was going through amazon_us_reviews dataset and found that example API usage given on website is different from the API usage while loading dataset. Eg. what API usage is on the [website](https://huggingface.co/datasets/amazon_us_reviews) ``` from datasets import load_dataset dataset = load_dataset("amaz...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/849/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/849/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/848
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/848/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/848/comments
https://api.github.com/repos/huggingface/datasets/issues/848/events
https://github.com/huggingface/datasets/issues/848
742,240,942
MDU6SXNzdWU3NDIyNDA5NDI=
848
Error when concatenate_datasets
{ "login": "shexuan", "id": 25664170, "node_id": "MDQ6VXNlcjI1NjY0MTcw", "avatar_url": "https://avatars.githubusercontent.com/u/25664170?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shexuan", "html_url": "https://github.com/shexuan", "followers_url": "https://api.github.com/users/shexua...
[]
closed
false
null
[]
null
[ "As you can see in the error the test checks if `indices_mappings_in_memory` is True or not, which is different from the test you do in your script. In a dataset, both the data and the indices mapping can be either on disk or in memory.\r\n\r\nThe indices mapping correspond to a mapping on top of the data table tha...
1,605,254,162,000
1,605,289,259,000
1,605,282,910,000
NONE
null
Hello, when I concatenate two dataset loading from disk, I encountered a problem: ``` test_dataset = load_from_disk('data/test_dataset') trn_dataset = load_from_disk('data/train_dataset') train_dataset = concatenate_datasets([trn_dataset, test_dataset]) ``` And it reported ValueError blow: ``` --------------...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/848/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/848/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/847
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/847/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/847/comments
https://api.github.com/repos/huggingface/datasets/issues/847/events
https://github.com/huggingface/datasets/issues/847
742,179,495
MDU6SXNzdWU3NDIxNzk0OTU=
847
multiprocessing in dataset map "can only test a child process"
{ "login": "timothyjlaurent", "id": 2000204, "node_id": "MDQ6VXNlcjIwMDAyMDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/timothyjlaurent", "html_url": "https://github.com/timothyjlaurent", "followers_url": "https://api.g...
[]
open
false
null
[]
null
[ "It looks like an issue with wandb/tqdm here.\r\nWe're using the `multiprocess` library instead of the `multiprocessing` builtin python package to support various types of mapping functions. Maybe there's some sort of incompatibility.\r\n\r\nCould you make a minimal script to reproduce or a google colab ?", "hi f...
1,605,247,264,000
1,612,198,408,000
null
NONE
null
Using a dataset with a single 'text' field and a fast tokenizer in a jupyter notebook. ``` def tokenizer_fn(example): return tokenizer.batch_encode_plus(example['text']) ds_tokenized = text_dataset.map(tokenizer_fn, batched=True, num_proc=6, remove_columns=['text']) ``` ``` -------------------------...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/847/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/847/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/846
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/846/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/846/comments
https://api.github.com/repos/huggingface/datasets/issues/846/events
https://github.com/huggingface/datasets/issues/846
741,885,174
MDU6SXNzdWU3NDE4ODUxNzQ=
846
Add HoVer multi-hop fact verification dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "Hi @yjernite I'm new but wanted to contribute. Has anyone already taken this problem and do you think it is suitable for newbies?", "Hi @tenjjin! This dataset is still up for grabs! Here's the link with the guide to add it. You should play around with the library first (download and look at a few datasets), the...
1,605,210,946,000
1,607,636,853,000
1,607,636,853,000
MEMBER
null
## Adding a Dataset - **Name:** HoVer - **Description:** https://twitter.com/YichenJiang9/status/1326954363806429186 contains 20K claim verification examples - **Paper:** https://arxiv.org/abs/2011.03088 - **Data:** https://hover-nlp.github.io/ - **Motivation:** There are still few multi-hop information extraction...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/846/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/846/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/843
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/843/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/843/comments
https://api.github.com/repos/huggingface/datasets/issues/843/events
https://github.com/huggingface/datasets/issues/843
741,531,121
MDU6SXNzdWU3NDE1MzExMjE=
843
use_custom_baseline still produces errors for bertscore
{ "login": "penatbater", "id": 37921244, "node_id": "MDQ6VXNlcjM3OTIxMjQ0", "avatar_url": "https://avatars.githubusercontent.com/u/37921244?v=4", "gravatar_id": "", "url": "https://api.github.com/users/penatbater", "html_url": "https://github.com/penatbater", "followers_url": "https://api.github.com/use...
[ { "id": 2067393914, "node_id": "MDU6TGFiZWwyMDY3MzkzOTE0", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug", "name": "metric bug", "color": "25b21e", "default": false, "description": "A bug in a metric script" } ]
closed
false
null
[]
null
[ "Thanks for reporting ! That's a bug indeed\r\nIf you want to contribute, feel free to fix this issue and open a PR :)", "This error is because of a mismatch between `datasets` and `bert_score`. With `datasets=1.1.2` and `bert_score>=0.3.6` it works ok. So `pip install -U bert_score` should fix the problem. ", ...
1,605,181,472,000
1,630,404,404,000
1,612,880,508,000
NONE
null
`metric = load_metric('bertscore')` `a1 = "random sentences"` `b1 = "random sentences"` `metric.compute(predictions = [a1], references = [b1], lang = 'en')` `Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/stephen_chan/.local/lib/python3.6/site-packages/datasets/metric.py"...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/843/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/843/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/842
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/842/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/842/comments
https://api.github.com/repos/huggingface/datasets/issues/842/events
https://github.com/huggingface/datasets/issues/842
741,208,428
MDU6SXNzdWU3NDEyMDg0Mjg=
842
How to enable `.map()` pre-processing pipelines to support multi-node parallelism?
{ "login": "shangw-nvidia", "id": 66387198, "node_id": "MDQ6VXNlcjY2Mzg3MTk4", "avatar_url": "https://avatars.githubusercontent.com/u/66387198?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shangw-nvidia", "html_url": "https://github.com/shangw-nvidia", "followers_url": "https://api.githu...
[]
open
false
null
[]
null
[ "Right now multiprocessing only runs on single node.\r\n\r\nHowever it's probably possible to extend it to support multi nodes. Indeed we're using the `multiprocess` library from the `pathos` project to do multiprocessing in `datasets`, and `pathos` is made to support parallelism on several nodes. More info about p...
1,605,146,678,000
1,605,223,707,000
null
NONE
null
Hi, Currently, multiprocessing can be enabled for the `.map()` stages on a single node. However, in the case of multi-node training, (since more than one node would be available) I'm wondering if it's possible to extend the parallel processing among nodes, instead of only 1 node running the `.map()` while the other ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/842/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/842/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/841
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/841/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/841/comments
https://api.github.com/repos/huggingface/datasets/issues/841/events
https://github.com/huggingface/datasets/issues/841
740,737,448
MDU6SXNzdWU3NDA3Mzc0NDg=
841
Can not reuse datasets already downloaded
{ "login": "jc-hou", "id": 30210529, "node_id": "MDQ6VXNlcjMwMjEwNTI5", "avatar_url": "https://avatars.githubusercontent.com/u/30210529?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jc-hou", "html_url": "https://github.com/jc-hou", "followers_url": "https://api.github.com/users/jc-hou/fo...
[]
closed
false
null
[]
null
[ "It seems the process needs '/datasets.huggingface.co/datasets/datasets/wikipedia/wikipedia.py'\r\nWhere and how to assign this ```wikipedia.py``` after I manually download it ?", "\r\ndownload the ```wikipedia.py``` at the working directory and go with ```dataset = load_dataset('wikipedia.py', '20200501.en')``` ...
1,605,098,535,000
1,605,118,636,000
1,605,118,636,000
NONE
null
Hello, I need to connect to a frontal node (with http proxy, no gpu) before connecting to a gpu node (but no http proxy, so can not use wget so on). I successfully downloaded and reuse the wikipedia datasets in a frontal node. When I connect to the gpu node, I supposed to use the downloaded datasets from cache, but...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/841/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/841/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/839
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/839/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/839/comments
https://api.github.com/repos/huggingface/datasets/issues/839/events
https://github.com/huggingface/datasets/issues/839
740,355,270
MDU6SXNzdWU3NDAzNTUyNzA=
839
XSum dataset missing spaces between sentences
{ "login": "loganlebanoff", "id": 10007282, "node_id": "MDQ6VXNlcjEwMDA3Mjgy", "avatar_url": "https://avatars.githubusercontent.com/u/10007282?v=4", "gravatar_id": "", "url": "https://api.github.com/users/loganlebanoff", "html_url": "https://github.com/loganlebanoff", "followers_url": "https://api.githu...
[]
open
false
null
[]
null
[]
1,605,054,883,000
1,605,054,883,000
null
NONE
null
I noticed that the XSum dataset has no space between sentences. This could lead to worse results for anyone training or testing on it. Here's an example (0th entry in the test set): `The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/839/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/839/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/836
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/836/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/836/comments
https://api.github.com/repos/huggingface/datasets/issues/836/events
https://github.com/huggingface/datasets/issues/836
740,187,613
MDU6SXNzdWU3NDAxODc2MTM=
836
load_dataset with 'csv' is not working. while the same file is loading with 'text' mode or with pandas
{ "login": "randubin", "id": 8919490, "node_id": "MDQ6VXNlcjg5MTk0OTA=", "avatar_url": "https://avatars.githubusercontent.com/u/8919490?v=4", "gravatar_id": "", "url": "https://api.github.com/users/randubin", "html_url": "https://github.com/randubin", "followers_url": "https://api.github.com/users/randu...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
null
[]
null
[ "Which version of pyarrow do you have ? Could you try to update pyarrow and try again ?", "Thanks for the fast response. I have the latest version '2.0.0' (I tried to update)\r\nI am working with Python 3.8.5", "I think that the issue is similar to this one:https://issues.apache.org/jira/browse/ARROW-9612\r\nTh...
1,605,036,940,000
1,637,773,159,000
1,605,807,338,000
NONE
null
Hi All I am trying to load a custom dataset and I am trying to load a single file to make sure the file is loading correctly: dataset = load_dataset('csv', data_files=files) When I run it I get: Downloading and preparing dataset csv/default-35575a1051604c88 (download: Unknown size, generated: Unknown size, post-...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/836/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/836/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/835
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/835/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/835/comments
https://api.github.com/repos/huggingface/datasets/issues/835/events
https://github.com/huggingface/datasets/issues/835
740,102,210
MDU6SXNzdWU3NDAxMDIyMTA=
835
Wikipedia postprocessing
{ "login": "bminixhofer", "id": 13353204, "node_id": "MDQ6VXNlcjEzMzUzMjA0", "avatar_url": "https://avatars.githubusercontent.com/u/13353204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bminixhofer", "html_url": "https://github.com/bminixhofer", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[ "Hi @bminixhofer ! Parsing WikiMedia is notoriously difficult: this processing used [mwparserfromhell](https://github.com/earwig/mwparserfromhell) which is pretty good but not perfect.\r\n\r\nAs an alternative, you can also use the Wiki40b dataset which was pre-processed using an un-released Google internal tool", ...
1,605,029,198,000
1,605,032,600,000
1,605,030,561,000
NONE
null
Hi, thanks for this library! Running this code: ```py import datasets wikipedia = datasets.load_dataset("wikipedia", "20200501.de") print(wikipedia['train']['text'][0]) ``` I get: ``` mini|Ricardo Flores Magón mini|Mexikanische Revolutionäre, Magón in der Mitte anführend, gegen die Diktatur von Porfir...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/835/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/835/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/834
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/834/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/834/comments
https://api.github.com/repos/huggingface/datasets/issues/834/events
https://github.com/huggingface/datasets/issues/834
740,082,890
MDU6SXNzdWU3NDAwODI4OTA=
834
[GEM] add WikiLingua cross-lingual abstractive summarization dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "Hey @yjernite. This is a very interesting dataset. Would love to work on adding it but I see that the link to the data is to a gdrive folder. Can I just confirm wether dlmanager can handle gdrive urls or would this have to be a manual dl?", "Hi @KMFODA ! A version of WikiLingua is actually already accessible in ...
1,605,027,643,000
1,618,488,249,000
1,618,488,098,000
MEMBER
null
## Adding a Dataset - **Name:** WikiLingua - **Description:** The dataset includes ~770k article and summary pairs in 18 languages from WikiHow. The gold-standard article-summary alignments across languages were extracted by aligning the images that are used to describe each how-to step in an article. - **Paper:** h...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/834/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/834/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/833
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/833/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/833/comments
https://api.github.com/repos/huggingface/datasets/issues/833/events
https://github.com/huggingface/datasets/issues/833
740,079,692
MDU6SXNzdWU3NDAwNzk2OTI=
833
[GEM] add ASSET text simplification dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[]
1,605,027,390,000
1,607,002,695,000
1,607,002,695,000
MEMBER
null
## Adding a Dataset - **Name:** ASSET - **Description:** ASSET is a crowdsourced multi-reference corpus for assessing sentence simplification in English where each simplification was produced by executing several rewriting transformations. - **Paper:** https://www.aclweb.org/anthology/2020.acl-main.424.pdf - **Dat...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/833/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/833/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/832
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/832/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/832/comments
https://api.github.com/repos/huggingface/datasets/issues/832/events
https://github.com/huggingface/datasets/issues/832
740,077,228
MDU6SXNzdWU3NDAwNzcyMjg=
832
[GEM] add WikiAuto text simplification dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[]
1,605,027,203,000
1,607,002,688,000
1,607,002,688,000
MEMBER
null
## Adding a Dataset - **Name:** WikiAuto - **Description:** Sentences in English Wikipedia and their corresponding sentences in Simple English Wikipedia that are written with simpler grammar and word choices. A lot of lexical and syntactic paraphrasing. - **Paper:** https://www.aclweb.org/anthology/2020.acl-main.70...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/832/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/832/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/831
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/831/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/831/comments
https://api.github.com/repos/huggingface/datasets/issues/831/events
https://github.com/huggingface/datasets/issues/831
740,071,697
MDU6SXNzdWU3NDAwNzE2OTc=
831
[GEM] Add WebNLG dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[]
1,605,026,808,000
1,607,002,681,000
1,607,002,681,000
MEMBER
null
## Adding a Dataset - **Name:** WebNLG - **Description:** WebNLG consists of Data/Text pairs where the data is a set of triples extracted from DBpedia and the text is a verbalisation of these triples (16,095 data inputs and 42,873 data-text pairs). The data is available in English and Russian - **Paper:** https://ww...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/831/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/831/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/830
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/830/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/830/comments
https://api.github.com/repos/huggingface/datasets/issues/830/events
https://github.com/huggingface/datasets/issues/830
740,065,376
MDU6SXNzdWU3NDAwNjUzNzY=
830
[GEM] add ToTTo Table-to-text dataset
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
null
[]
null
[ "closed via #1098 " ]
1,605,026,314,000
1,607,605,562,000
1,607,605,561,000
MEMBER
null
## Adding a Dataset - **Name:** ToTTo - **Description:** ToTTo is an open-domain English table-to-text dataset with over 120,000 training examples that proposes a controlled generation task: given a Wikipedia table and a set of highlighted table cells, produce a one-sentence description. - **Paper:** https://arxiv.o...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/830/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/830/timeline
null
null
null
false