url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1.28B
node_id
stringlengths
18
32
number
int64
1
4.53k
title
stringlengths
1
276
user
dict
labels
list
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
list
milestone
dict
comments
list
created_at
int64
1,587B
1,656B
updated_at
int64
1,587B
1,656B
closed_at
int64
1,587B
1,656B
author_association
stringclasses
3 values
active_lock_reason
null
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
state_reason
stringclasses
1 value
draft
bool
2 classes
pull_request
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/2805
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2805/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2805/comments
https://api.github.com/repos/huggingface/datasets/issues/2805/events
https://github.com/huggingface/datasets/pull/2805
971,436,456
MDExOlB1bGxSZXF1ZXN0NzEzMTc3MTI4
2,805
Fix streaming zip files from canonical datasets
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,629,097,900,000
1,629,110,040,000
1,629,110,040,000
MEMBER
null
Previous PR #2798 fixed streaming remote zip files when passing the parameter `data_files`. However, that broke streaming zip files used in canonical `datasets` scripts, which normally have a subsequent `join()` (patched with `xjoin()`) after the `StreamingDownloadManager.download_and_extract()` is called. This P...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2805/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2805/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2805", "html_url": "https://github.com/huggingface/datasets/pull/2805", "diff_url": "https://github.com/huggingface/datasets/pull/2805.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2805.patch", "merged_at": 1629110040000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2804
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2804/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2804/comments
https://api.github.com/repos/huggingface/datasets/issues/2804/events
https://github.com/huggingface/datasets/pull/2804
971,353,437
MDExOlB1bGxSZXF1ZXN0NzEzMTA2NTMw
2,804
Add Food-101
{ "login": "nateraw", "id": 32437151, "node_id": "MDQ6VXNlcjMyNDM3MTUx", "avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nateraw", "html_url": "https://github.com/nateraw", "followers_url": "https://api.github.com/users/natera...
[]
closed
false
null
[]
null
[]
1,629,087,975,000
1,629,469,893,000
1,629,377,286,000
CONTRIBUTOR
null
Adds image classification dataset [Food-101](https://data.vision.ee.ethz.ch/cvl/datasets_extra/food-101/).
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2804/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2804/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2804", "html_url": "https://github.com/huggingface/datasets/pull/2804", "diff_url": "https://github.com/huggingface/datasets/pull/2804.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2804.patch", "merged_at": 1629377286000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2803
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2803/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2803/comments
https://api.github.com/repos/huggingface/datasets/issues/2803/events
https://github.com/huggingface/datasets/pull/2803
970,858,928
MDExOlB1bGxSZXF1ZXN0NzEyNzQxODMz
2,803
add stack exchange
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Hi ! Merging this one since it's all good :)\r\n\r\nHowever I think it would also be better to actually rename it `the_pile_stack_exchange` to make things clearer and to avoid name collisions in the future. I would like to do the same for `books3` as well.\r\n\r\nIf you don't mind I'll open a PR to do the renaming...
1,628,928,662,000
1,629,367,653,000
1,629,360,458,000
CONTRIBUTOR
null
stack exchange is part of EleutherAI/The Pile, but AFAIK, The Pile dataset blend all sub datasets together thus we are not able to use just one of its sub dataset from The Pile data. So I create an independent dataset using The Pile preliminary components. I also change default `timeout` to 100 seconds instead of 10...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2803/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2803/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2803", "html_url": "https://github.com/huggingface/datasets/pull/2803", "diff_url": "https://github.com/huggingface/datasets/pull/2803.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2803.patch", "merged_at": 1629360458000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2802
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2802/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2802/comments
https://api.github.com/repos/huggingface/datasets/issues/2802/events
https://github.com/huggingface/datasets/pull/2802
970,848,302
MDExOlB1bGxSZXF1ZXN0NzEyNzM0MTc3
2,802
add openwebtext2
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "It seems we need to `pip install jsonlines` to pass the checks ?", "Hi ! Do you really need `jsonlines` ? I think it simply uses `json.loads` under the hood.\r\n\r\nCurrently the test are failing because `jsonlines` is not part of the extra requirements `TESTS_REQUIRE` in setup.py\r\n\r\nSo either you can replac...
1,628,924,943,000
1,629,727,574,000
1,629,727,574,000
CONTRIBUTOR
null
openwebtext2 is part of EleutherAI/The Pile, but AFAIK, The Pile dataset blend all sub datasets together thus we are not able to use just one of its sub dataset from The Pile data. So I create an independent dataset using The Pile preliminary components. When I was creating dataset card. I found there is room for cr...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2802/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2802/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2802", "html_url": "https://github.com/huggingface/datasets/pull/2802", "diff_url": "https://github.com/huggingface/datasets/pull/2802.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2802.patch", "merged_at": 1629727574000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2801
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2801/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2801/comments
https://api.github.com/repos/huggingface/datasets/issues/2801/events
https://github.com/huggingface/datasets/pull/2801
970,844,617
MDExOlB1bGxSZXF1ZXN0NzEyNzMwODEz
2,801
add books3
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "> When I was creating dataset card. I found there is room for creating / editing dataset card. I've made it an issue. #2797\r\n\r\nThanks for the message, we'll definitely improve this\r\n\r\n> Also I am wondering whether the import of The Pile dataset is actively undertaken (because I may need it recently)? #1675...
1,628,924,665,000
1,629,391,389,000
1,629,301,019,000
CONTRIBUTOR
null
books3 is part of EleutherAI/The Pile, but AFAIK, The Pile dataset blend all sub datasets together thus we are not able to use just one of its sub dataset from The Pile data. So I create an independent dataset using The Pile preliminary components. When I was creating dataset card. I found there is room for creating...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2801/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2801/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2801", "html_url": "https://github.com/huggingface/datasets/pull/2801", "diff_url": "https://github.com/huggingface/datasets/pull/2801.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2801.patch", "merged_at": 1629301019000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2800
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2800/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2800/comments
https://api.github.com/repos/huggingface/datasets/issues/2800/events
https://github.com/huggingface/datasets/pull/2800
970,819,988
MDExOlB1bGxSZXF1ZXN0NzEyNzExNTcx
2,800
Support streaming tar files
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "Hi ! Why do we need the custom `readline` for exactly ? feel free to add a comment to say why it's needed" ]
1,628,916,017,000
1,629,972,150,000
1,628,916,957,000
MEMBER
null
This PR adds support to stream tar files by using the `fsspec` tar protocol. It also uses the custom `readline` implemented in PR #2786. The corresponding test is implemented in PR #2786.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2800/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2800/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2800", "html_url": "https://github.com/huggingface/datasets/pull/2800", "diff_url": "https://github.com/huggingface/datasets/pull/2800.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2800.patch", "merged_at": 1628916957000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2799
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2799/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2799/comments
https://api.github.com/repos/huggingface/datasets/issues/2799/events
https://github.com/huggingface/datasets/issues/2799
970,507,351
MDU6SXNzdWU5NzA1MDczNTE=
2,799
Loading JSON throws ArrowNotImplementedError
{ "login": "lewtun", "id": 26859204, "node_id": "MDQ6VXNlcjI2ODU5MjA0", "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lewtun", "html_url": "https://github.com/lewtun", "followers_url": "https://api.github.com/users/lewtun/fo...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi @lewtun, thanks for reporting.\r\n\r\nApparently, `pyarrow.json` tries to cast timestamp-like fields in your JSON file to pyarrow timestamp type, and it fails with `ArrowNotImplementedError`.\r\n\r\nI will investigate if there is a way to tell pyarrow not to try that timestamp casting.", "I think the issue is...
1,628,868,708,000
1,641,841,172,000
1,641,841,172,000
MEMBER
null
## Describe the bug I have created a [dataset](https://huggingface.co/datasets/lewtun/github-issues-test) of GitHub issues in line-separated JSON format and am finding that I cannot load it with the `json` loading script (see stack trace below). Curiously, there is no problem loading the dataset with `pandas` which...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2799/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2799/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2798
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2798/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2798/comments
https://api.github.com/repos/huggingface/datasets/issues/2798/events
https://github.com/huggingface/datasets/pull/2798
970,493,126
MDExOlB1bGxSZXF1ZXN0NzEyNDM3ODc2
2,798
Fix streaming zip files
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "Hi ! I don't fully understand this change @albertvillanova \r\nThe `_extract` method used to return the compound URL that points to the root of the inside of the archive.\r\nThis way users can use the usual os.path.join or other functions to point to the relevant files. I don't see why you're using a glob pattern ...
1,628,867,821,000
1,629,123,410,000
1,628,869,108,000
MEMBER
null
Currently, streaming remote zip data files gives `FileNotFoundError` message: ```python data_files = f"https://huggingface.co/datasets/albertvillanova/datasets-tests-compression/resolve/main/sample.zip" ds = load_dataset("json", split="train", data_files=data_files, streaming=True) next(iter(ds)) ``` This PR fi...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2798/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2798/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2798", "html_url": "https://github.com/huggingface/datasets/pull/2798", "diff_url": "https://github.com/huggingface/datasets/pull/2798.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2798.patch", "merged_at": 1628869108000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2797
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2797/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2797/comments
https://api.github.com/repos/huggingface/datasets/issues/2797/events
https://github.com/huggingface/datasets/issues/2797
970,331,634
MDU6SXNzdWU5NzAzMzE2MzQ=
2,797
Make creating/editing dataset cards easier, by editing on site and dumping info from test command.
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.c...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[]
1,628,855,689,000
1,628,930,529,000
null
CONTRIBUTOR
null
**Is your feature request related to a problem? Please describe.** Creating and editing dataset cards should be but not that easy - If other else know Some information I don't know (bias of dataset, dataset curation, supported dataset, ...), he/she should know the description on hf.co comes from README.md under git...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2797/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2797/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2796
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2796/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2796/comments
https://api.github.com/repos/huggingface/datasets/issues/2796/events
https://github.com/huggingface/datasets/pull/2796
970,235,846
MDExOlB1bGxSZXF1ZXN0NzEyMjE1ODM2
2,796
add cedr dataset
{ "login": "naumov-al", "id": 22640075, "node_id": "MDQ6VXNlcjIyNjQwMDc1", "avatar_url": "https://avatars.githubusercontent.com/u/22640075?v=4", "gravatar_id": "", "url": "https://api.github.com/users/naumov-al", "html_url": "https://github.com/naumov-al", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "> Hi ! Thanks a lot for adding this one :)\r\n> \r\n> Good job with the dataset card and the dataset script !\r\n> \r\n> I left a few suggestions\r\n\r\nThank you very much for your helpful suggestions. I have tried to carry them all out." ]
1,628,847,455,000
1,630,080,096,000
1,630,080,096,000
CONTRIBUTOR
null
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2796/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2796/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2796", "html_url": "https://github.com/huggingface/datasets/pull/2796", "diff_url": "https://github.com/huggingface/datasets/pull/2796.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2796.patch", "merged_at": 1630080095000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2794
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2794/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2794/comments
https://api.github.com/repos/huggingface/datasets/issues/2794/events
https://github.com/huggingface/datasets/issues/2794
969,728,545
MDU6SXNzdWU5Njk3Mjg1NDU=
2,794
Warnings and documentation about pickling incorrect
{ "login": "mbforbes", "id": 1170062, "node_id": "MDQ6VXNlcjExNzAwNjI=", "avatar_url": "https://avatars.githubusercontent.com/u/1170062?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mbforbes", "html_url": "https://github.com/mbforbes", "followers_url": "https://api.github.com/users/mbfor...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
[]
1,628,809,753,000
1,628,809,771,000
null
NONE
null
## Describe the bug I have a docs bug and a closely related docs enhancement suggestion! ### Bug The warning and documentation say "either `dill` or `pickle`" for fingerprinting. But it seems that `dill`, which is installed by `datasets` by default, _must_ work, or else the fingerprinting fails. Warning: ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2794/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2794/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2793
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2793/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2793/comments
https://api.github.com/repos/huggingface/datasets/issues/2793/events
https://github.com/huggingface/datasets/pull/2793
968,967,773
MDExOlB1bGxSZXF1ZXN0NzExMDQ4NDY2
2,793
Fix type hint for data_files
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,628,779,357,000
1,628,782,529,000
1,628,782,529,000
MEMBER
null
Fix type hint for `data_files` in signatures and docstrings.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2793/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2793/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2793", "html_url": "https://github.com/huggingface/datasets/pull/2793", "diff_url": "https://github.com/huggingface/datasets/pull/2793.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2793.patch", "merged_at": 1628782529000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2792
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2792/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2792/comments
https://api.github.com/repos/huggingface/datasets/issues/2792/events
https://github.com/huggingface/datasets/pull/2792
968,650,274
MDExOlB1bGxSZXF1ZXN0NzEwNzUyMjc0
2,792
Update: GooAQ - add train/val/test splits
{ "login": "bhavitvyamalik", "id": 19718818, "node_id": "MDQ6VXNlcjE5NzE4ODE4", "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bhavitvyamalik", "html_url": "https://github.com/bhavitvyamalik", "followers_url": "https://api.gi...
[]
closed
false
null
[]
null
[ "@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l...
1,628,768,418,000
1,630,079,925,000
1,630,079,894,000
CONTRIBUTOR
null
[GooAQ](https://github.com/allenai/gooaq) dataset was recently updated after splits were added for the same. This PR contains new updated GooAQ with train/val/test splits and updated README as well.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2792/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2792/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2792", "html_url": "https://github.com/huggingface/datasets/pull/2792", "diff_url": "https://github.com/huggingface/datasets/pull/2792.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2792.patch", "merged_at": 1630079894000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2791
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2791/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2791/comments
https://api.github.com/repos/huggingface/datasets/issues/2791/events
https://github.com/huggingface/datasets/pull/2791
968,360,314
MDExOlB1bGxSZXF1ZXN0NzEwNDgxNDAy
2,791
Fix typo in cnn_dailymail
{ "login": "omaralsayed", "id": 42531544, "node_id": "MDQ6VXNlcjQyNTMxNTQ0", "avatar_url": "https://avatars.githubusercontent.com/u/42531544?v=4", "gravatar_id": "", "url": "https://api.github.com/users/omaralsayed", "html_url": "https://github.com/omaralsayed", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[]
1,628,757,522,000
1,628,767,079,000
1,628,767,079,000
CONTRIBUTOR
null
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2791/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2791/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2791", "html_url": "https://github.com/huggingface/datasets/pull/2791", "diff_url": "https://github.com/huggingface/datasets/pull/2791.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2791.patch", "merged_at": 1628767079000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2790
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2790/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2790/comments
https://api.github.com/repos/huggingface/datasets/issues/2790/events
https://github.com/huggingface/datasets/pull/2790
967,772,181
MDExOlB1bGxSZXF1ZXN0NzA5OTI3NjM2
2,790
Fix typo in test_dataset_common
{ "login": "nateraw", "id": 32437151, "node_id": "MDQ6VXNlcjMyNDM3MTUx", "avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nateraw", "html_url": "https://github.com/nateraw", "followers_url": "https://api.github.com/users/natera...
[]
closed
false
null
[]
null
[]
1,628,730,629,000
1,628,767,889,000
1,628,767,889,000
CONTRIBUTOR
null
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2790/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2790/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2790", "html_url": "https://github.com/huggingface/datasets/pull/2790", "diff_url": "https://github.com/huggingface/datasets/pull/2790.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2790.patch", "merged_at": 1628767889000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2789
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2789/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2789/comments
https://api.github.com/repos/huggingface/datasets/issues/2789/events
https://github.com/huggingface/datasets/pull/2789
967,361,934
MDExOlB1bGxSZXF1ZXN0NzA5NTQwMzY5
2,789
Updated dataset description of DaNE
{ "login": "KennethEnevoldsen", "id": 23721977, "node_id": "MDQ6VXNlcjIzNzIxOTc3", "avatar_url": "https://avatars.githubusercontent.com/u/23721977?v=4", "gravatar_id": "", "url": "https://api.github.com/users/KennethEnevoldsen", "html_url": "https://github.com/KennethEnevoldsen", "followers_url": "https...
[]
closed
false
null
[]
null
[ "Thanks for finishing it @albertvillanova " ]
1,628,711,928,000
1,628,784,659,000
1,628,784,361,000
CONTRIBUTOR
null
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2789/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2789/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2789", "html_url": "https://github.com/huggingface/datasets/pull/2789", "diff_url": "https://github.com/huggingface/datasets/pull/2789.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2789.patch", "merged_at": 1628784361000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2788
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2788/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2788/comments
https://api.github.com/repos/huggingface/datasets/issues/2788/events
https://github.com/huggingface/datasets/issues/2788
967,149,389
MDU6SXNzdWU5NjcxNDkzODk=
2,788
How to sample every file in a list of files making up a split in a dataset when loading?
{ "login": "brijow", "id": 11220949, "node_id": "MDQ6VXNlcjExMjIwOTQ5", "avatar_url": "https://avatars.githubusercontent.com/u/11220949?v=4", "gravatar_id": "", "url": "https://api.github.com/users/brijow", "html_url": "https://github.com/brijow", "followers_url": "https://api.github.com/users/brijow/fo...
[]
open
false
null
[]
null
[ "Hi ! This is not possible just with `load_dataset`.\r\n\r\nYou can do something like this instead:\r\n```python\r\nseed=42\r\ndata_files_dict = {\r\n \"train\": [train_file1, train_file2],\r\n \"test\": [test_file1, test_file2],\r\n \"val\": [val_file1, val_file2]\r\n}\r\ndataset = datasets.load_dataset(\...
1,628,703,801,000
1,629,738,742,000
null
NONE
null
I am loading a dataset with multiple train, test, and validation files like this: ``` data_files_dict = { "train": [train_file1, train_file2], "test": [test_file1, test_file2], "val": [val_file1, val_file2] } dataset = datasets.load_dataset( "csv", data_files=data_files_dict, split=[...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2788/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2788/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2787
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2787/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2787/comments
https://api.github.com/repos/huggingface/datasets/issues/2787/events
https://github.com/huggingface/datasets/issues/2787
967,018,406
MDU6SXNzdWU5NjcwMTg0MDY=
2,787
ConnectionError: Couldn't reach https://raw.githubusercontent.com
{ "login": "jinec", "id": 39627475, "node_id": "MDQ6VXNlcjM5NjI3NDc1", "avatar_url": "https://avatars.githubusercontent.com/u/39627475?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jinec", "html_url": "https://github.com/jinec", "followers_url": "https://api.github.com/users/jinec/follow...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "the bug code locate in :\r\n if data_args.task_name is not None:\r\n # Downloading and loading a dataset from the hub.\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)", "Hi @jinec,\r\n\r\nFrom time to time we get this kind of `ConnectionError` coming fr...
1,628,698,741,000
1,637,735,138,000
1,629,299,358,000
NONE
null
Hello, I am trying to run run_glue.py and it gives me this error - Traceback (most recent call last): File "E:/BERT/pytorch_hugging/transformers/examples/pytorch/text-classification/run_glue.py", line 546, in <module> main() File "E:/BERT/pytorch_hugging/transformers/examples/pytorch/text-classification/...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2787/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2787/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2786
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2786/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2786/comments
https://api.github.com/repos/huggingface/datasets/issues/2786/events
https://github.com/huggingface/datasets/pull/2786
966,282,934
MDExOlB1bGxSZXF1ZXN0NzA4NTQwMzU0
2,786
Support streaming compressed files
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,628,672,526,000
1,629,178,119,000
1,629,095,779,000
MEMBER
null
Add support to stream compressed files (current options in fsspec): - bz2 - lz4 - xz - zstd cc: @lewtun
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2786/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2786/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2786", "html_url": "https://github.com/huggingface/datasets/pull/2786", "diff_url": "https://github.com/huggingface/datasets/pull/2786.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2786.patch", "merged_at": 1629095779000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2783
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2783/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2783/comments
https://api.github.com/repos/huggingface/datasets/issues/2783/events
https://github.com/huggingface/datasets/pull/2783
965,461,382
MDExOlB1bGxSZXF1ZXN0NzA3NzcxOTM3
2,783
Add KS task to SUPERB
{ "login": "anton-l", "id": 26864830, "node_id": "MDQ6VXNlcjI2ODY0ODMw", "avatar_url": "https://avatars.githubusercontent.com/u/26864830?v=4", "gravatar_id": "", "url": "https://api.github.com/users/anton-l", "html_url": "https://github.com/anton-l", "followers_url": "https://api.github.com/users/anton-...
[]
closed
false
null
[]
null
[ "thanks a lot for implementing this @anton-l !!\r\n\r\ni won't have time to review this while i'm away, so happy for @albertvillanova and @patrickvonplaten to decide when to merge :)", "@albertvillanova thanks! Everything should be ready now :)", "> The _background_noise_/_silence_ audio files are much longer t...
1,628,633,647,000
1,628,786,701,000
1,628,713,157,000
MEMBER
null
Add the KS (keyword spotting) task as described in the [SUPERB paper](https://arxiv.org/abs/2105.01051). - [s3prl instructions](https://github.com/s3prl/s3prl/blob/master/s3prl/downstream/README.md#ks-keyword-spotting) - [s3prl implementation](https://github.com/s3prl/s3prl/blob/master/s3prl/downstream/speech_comma...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2783/reactions", "total_count": 3, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 3, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2783/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2783", "html_url": "https://github.com/huggingface/datasets/pull/2783", "diff_url": "https://github.com/huggingface/datasets/pull/2783.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2783.patch", "merged_at": 1628713157000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2782
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2782/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2782/comments
https://api.github.com/repos/huggingface/datasets/issues/2782/events
https://github.com/huggingface/datasets/pull/2782
964,858,439
MDExOlB1bGxSZXF1ZXN0NzA3MjQ5NDE5
2,782
Fix renaming of corpus_bleu args
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,628,593,354,000
1,628,594,167,000
1,628,594,167,000
MEMBER
null
Last `sacrebleu` release (v2.0.0) has renamed `sacrebleu.corpus_bleu` args from `(sys_stream, ref_streams)` to `(hipotheses, references)`: https://github.com/mjpost/sacrebleu/pull/152/files#diff-2553a315bb1f7e68c9c1b00d56eaeb74f5205aeb3a189bc3e527b122c6078795L17-R15 This PR passes the args without parameter names, s...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2782/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2782/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2782", "html_url": "https://github.com/huggingface/datasets/pull/2782", "diff_url": "https://github.com/huggingface/datasets/pull/2782.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2782.patch", "merged_at": 1628594167000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2781
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2781/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2781/comments
https://api.github.com/repos/huggingface/datasets/issues/2781/events
https://github.com/huggingface/datasets/issues/2781
964,805,351
MDU6SXNzdWU5NjQ4MDUzNTE=
2,781
Latest v2.0.0 release of sacrebleu has broken some metrics
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[]
1,628,589,581,000
1,628,594,167,000
1,628,594,167,000
MEMBER
null
## Describe the bug After `sacrebleu` v2.0.0 release (see changes here: https://github.com/mjpost/sacrebleu/pull/152/files#diff-2553a315bb1f7e68c9c1b00d56eaeb74f5205aeb3a189bc3e527b122c6078795L17-R15), some of `datasets` metrics are broken: - Default tokenizer `sacrebleu.DEFAULT_TOKENIZER` no longer exists: - #273...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2781/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2781/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2780
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2780/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2780/comments
https://api.github.com/repos/huggingface/datasets/issues/2780/events
https://github.com/huggingface/datasets/pull/2780
964,794,764
MDExOlB1bGxSZXF1ZXN0NzA3MTk2NjA3
2,780
VIVOS dataset for Vietnamese ASR
{ "login": "binh234", "id": 57580923, "node_id": "MDQ6VXNlcjU3NTgwOTIz", "avatar_url": "https://avatars.githubusercontent.com/u/57580923?v=4", "gravatar_id": "", "url": "https://api.github.com/users/binh234", "html_url": "https://github.com/binh234", "followers_url": "https://api.github.com/users/binh23...
[]
closed
false
null
[]
null
[]
1,628,588,856,000
1,628,766,570,000
1,628,766,570,000
CONTRIBUTOR
null
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2780/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2780/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2780", "html_url": "https://github.com/huggingface/datasets/pull/2780", "diff_url": "https://github.com/huggingface/datasets/pull/2780.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2780.patch", "merged_at": 1628766570000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2779
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2779/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2779/comments
https://api.github.com/repos/huggingface/datasets/issues/2779/events
https://github.com/huggingface/datasets/pull/2779
964,775,085
MDExOlB1bGxSZXF1ZXN0NzA3MTgwNTgw
2,779
Fix sacrebleu tokenizers
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,628,587,467,000
1,628,593,388,000
1,628,593,074,000
MEMBER
null
Last `sacrebleu` release (v2.0.0) has removed `sacrebleu.TOKENIZERS`: https://github.com/mjpost/sacrebleu/pull/152/files#diff-2553a315bb1f7e68c9c1b00d56eaeb74f5205aeb3a189bc3e527b122c6078795L17-R15 This PR makes a hot fix of the bug by using a private function in `sacrebleu`: `sacrebleu.metrics.bleu._get_tokenizer()...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2779/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2779/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2779", "html_url": "https://github.com/huggingface/datasets/pull/2779", "diff_url": "https://github.com/huggingface/datasets/pull/2779.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2779.patch", "merged_at": 1628593074000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2778
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2778/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2778/comments
https://api.github.com/repos/huggingface/datasets/issues/2778/events
https://github.com/huggingface/datasets/pull/2778
964,737,422
MDExOlB1bGxSZXF1ZXN0NzA3MTQ5MTk2
2,778
Do not pass tokenize to sacrebleu
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,628,584,837,000
1,628,589,817,000
1,628,589,817,000
MEMBER
null
Last `sacrebleu` release (v2.0.0) has removed `sacrebleu.DEFAULT_TOKENIZER`: https://github.com/mjpost/sacrebleu/pull/152/files#diff-2553a315bb1f7e68c9c1b00d56eaeb74f5205aeb3a189bc3e527b122c6078795L17-R15 This PR does not pass `tokenize` to `sacrebleu` (note that the user cannot pass it anyway) and `sacrebleu` will ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2778/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2778/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2778", "html_url": "https://github.com/huggingface/datasets/pull/2778", "diff_url": "https://github.com/huggingface/datasets/pull/2778.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2778.patch", "merged_at": 1628589817000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2777
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2777/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2777/comments
https://api.github.com/repos/huggingface/datasets/issues/2777/events
https://github.com/huggingface/datasets/pull/2777
964,696,380
MDExOlB1bGxSZXF1ZXN0NzA3MTEzNzg3
2,777
Use packaging to handle versions
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,628,581,899,000
1,629,294,987,000
1,629,294,987,000
MEMBER
null
Use packaging module to handle/validate/check versions of Python packages. Related to #2769.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2777/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2777/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2777", "html_url": "https://github.com/huggingface/datasets/pull/2777", "diff_url": "https://github.com/huggingface/datasets/pull/2777.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2777.patch", "merged_at": 1629294987000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2776
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2776/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2776/comments
https://api.github.com/repos/huggingface/datasets/issues/2776/events
https://github.com/huggingface/datasets/issues/2776
964,400,596
MDU6SXNzdWU5NjQ0MDA1OTY=
2,776
document `config.HF_DATASETS_OFFLINE` and precedence
{ "login": "stas00", "id": 10676103, "node_id": "MDQ6VXNlcjEwNjc2MTAz", "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stas00", "html_url": "https://github.com/stas00", "followers_url": "https://api.github.com/users/stas00/fo...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[]
1,628,544,197,000
1,628,544,197,000
null
MEMBER
null
https://github.com/huggingface/datasets/pull/1976 implemented `HF_DATASETS_OFFLINE`, but: 1. `config.HF_DATASETS_OFFLINE` is not documented 2. the precedence is not documented (env, config) I'm thinking it probably should be similar to what it says https://huggingface.co/docs/datasets/loading_datasets.html#from-th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2776/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2776/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2775
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2775/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2775/comments
https://api.github.com/repos/huggingface/datasets/issues/2775/events
https://github.com/huggingface/datasets/issues/2775
964,303,626
MDU6SXNzdWU5NjQzMDM2MjY=
2,775
`generate_random_fingerprint()` deterministic with 🤗Transformers' `set_seed()`
{ "login": "mbforbes", "id": 1170062, "node_id": "MDQ6VXNlcjExNzAwNjI=", "avatar_url": "https://avatars.githubusercontent.com/u/1170062?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mbforbes", "html_url": "https://github.com/mbforbes", "followers_url": "https://api.github.com/users/mbfor...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
[ "I dug into what I believe is the root of this issue and added a repro in my comment. If this is better addressed as a cross-team issue, let me know and I can open an issue in the Transformers repo", "Hi !\r\n\r\nIMO we shouldn't try to modify `set_seed` from transformers but maybe make `datasets` have its own RN...
1,628,537,331,000
1,629,966,654,000
null
NONE
null
## Describe the bug **Update:** I dug into this to try to reproduce the underlying issue, and I believe it's that `set_seed()` from the `transformers` library makes the "random" fingerprint identical each time. I believe this is still a bug, because `datasets` is used exactly this way in `transformers` after `set_se...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2775/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2775/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2774
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2774/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2774/comments
https://api.github.com/repos/huggingface/datasets/issues/2774/events
https://github.com/huggingface/datasets/pull/2774
963,932,199
MDExOlB1bGxSZXF1ZXN0NzA2NDY2MDc0
2,774
Prevent .map from using multiprocessing when loading from cache
{ "login": "thomasw21", "id": 24695242, "node_id": "MDQ6VXNlcjI0Njk1MjQy", "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomasw21", "html_url": "https://github.com/thomasw21", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "I'm guessing tests are failling, because this was pushed before https://github.com/huggingface/datasets/pull/2779 was merged? cc @albertvillanova ", "Hi @thomasw21, yes you are right: those failing tests were fixed with #2779.\r\n\r\nWould you mind to merge current upstream master branch and push again?\r\n```\r...
1,628,511,098,000
1,631,182,828,000
1,631,182,828,000
MEMBER
null
## Context On our setup, we use different setup to train vs proprocessing datasets. Usually we are able to obtain a high number of cpus to preprocess, which allows us to use `num_proc` however we can't use as many during training phase. Currently if we use `num_proc={whatever the preprocessing value was}` we load fr...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2774/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2774/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2774", "html_url": "https://github.com/huggingface/datasets/pull/2774", "diff_url": "https://github.com/huggingface/datasets/pull/2774.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2774.patch", "merged_at": 1631182828000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2773
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2773/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2773/comments
https://api.github.com/repos/huggingface/datasets/issues/2773/events
https://github.com/huggingface/datasets/issues/2773
963,730,497
MDU6SXNzdWU5NjM3MzA0OTc=
2,773
Remove dataset_infos.json
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 2067400324, "node_id": "MDU6...
open
false
null
[]
null
[]
1,628,494,999,000
1,628,494,999,000
null
MEMBER
null
**Is your feature request related to a problem? Please describe.** As discussed, there are infos in the `dataset_infos.json` which are redundant and we could have them only in the README file. Others could be migrated to the README, like: "dataset_size", "size_in_bytes", "download_size", "splits.split_name.[num_byt...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2773/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2773/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2772
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2772/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2772/comments
https://api.github.com/repos/huggingface/datasets/issues/2772/events
https://github.com/huggingface/datasets/issues/2772
963,348,834
MDU6SXNzdWU5NjMzNDg4MzQ=
2,772
Remove returned feature constrain
{ "login": "PosoSAgapo", "id": 33200481, "node_id": "MDQ6VXNlcjMzMjAwNDgx", "avatar_url": "https://avatars.githubusercontent.com/u/33200481?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PosoSAgapo", "html_url": "https://github.com/PosoSAgapo", "followers_url": "https://api.github.com/use...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[]
1,628,395,290,000
1,628,412,481,000
null
NONE
null
In the current version, the returned value of the map function has to be list or ndarray. However, this makes it unsuitable for many tasks. In NLP, many features are sparse like verb words, noun chunks, if we want to assign different values to different words, which will result in a large sparse matrix if we only score...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2772/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2772/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2771
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2771/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2771/comments
https://api.github.com/repos/huggingface/datasets/issues/2771/events
https://github.com/huggingface/datasets/pull/2771
963,257,036
MDExOlB1bGxSZXF1ZXN0NzA1OTExMDMw
2,771
[WIP][Common Voice 7] Add common voice 7.0
{ "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "followers_url": "https://...
[]
closed
false
null
[]
null
[ "Hi ! I think the name `common_voice_7` is fine :)\r\nMoreover if the dataset_infos.json is missing I'm pretty sure you don't need to specify `ignore_verifications=True`", "Hi, how about to add a new parameter \"version\" in the function load_dataset, something like: \r\n`load_dataset(\"common_voice\", \"lg\", ve...
1,628,352,070,000
1,638,833,042,000
1,638,833,042,000
MEMBER
null
This PR allows to load the new common voice dataset manually as explained when doing: ```python from datasets import load_dataset ds = load_dataset("./datasets/datasets/common_voice_7", "ab") ``` => ``` Please follow the manual download instructions: You need t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2771/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2771/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2771", "html_url": "https://github.com/huggingface/datasets/pull/2771", "diff_url": "https://github.com/huggingface/datasets/pull/2771.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2771.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/2770
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2770/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2770/comments
https://api.github.com/repos/huggingface/datasets/issues/2770/events
https://github.com/huggingface/datasets/pull/2770
963,246,512
MDExOlB1bGxSZXF1ZXN0NzA1OTAzMzIy
2,770
Add support for fast tokenizer in BertScore
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[]
1,628,348,403,000
1,628,512,483,000
1,628,507,785,000
CONTRIBUTOR
null
This PR adds support for a fast tokenizer in BertScore, which has been added recently to the lib. Fixes #2765
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2770/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2770/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2770", "html_url": "https://github.com/huggingface/datasets/pull/2770", "diff_url": "https://github.com/huggingface/datasets/pull/2770.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2770.patch", "merged_at": 1628507785000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2769
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2769/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2769/comments
https://api.github.com/repos/huggingface/datasets/issues/2769/events
https://github.com/huggingface/datasets/pull/2769
963,240,802
MDExOlB1bGxSZXF1ZXN0NzA1ODk5MTYy
2,769
Allow PyArrow from source
{ "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "followers_url": "https://...
[]
closed
false
null
[]
null
[]
1,628,346,404,000
1,628,523,519,000
1,628,523,519,000
MEMBER
null
When installing pyarrow from source the version is: ```python >>> import pyarrow; pyarrow.__version__ '2.1.0.dev612' ``` -> however this breaks the install check at init of `datasets`. This PR makes sure that everything coming after the last `'.'` is removed.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2769/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2769/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2769", "html_url": "https://github.com/huggingface/datasets/pull/2769", "diff_url": "https://github.com/huggingface/datasets/pull/2769.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2769.patch", "merged_at": 1628523519000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2768
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2768/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2768/comments
https://api.github.com/repos/huggingface/datasets/issues/2768/events
https://github.com/huggingface/datasets/issues/2768
963,229,173
MDU6SXNzdWU5NjMyMjkxNzM=
2,768
`ArrowInvalid: Added column's length must match table's length.` after using `select`
{ "login": "lvwerra", "id": 8264887, "node_id": "MDQ6VXNlcjgyNjQ4ODc=", "avatar_url": "https://avatars.githubusercontent.com/u/8264887?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lvwerra", "html_url": "https://github.com/lvwerra", "followers_url": "https://api.github.com/users/lvwerra/...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi,\r\n\r\nthe `select` method creates an indices mapping and doesn't modify the underlying PyArrow table by default for better performance. To modify the underlying table after the `select` call, call `flatten_indices` on the dataset object as follows:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds =...
1,628,342,249,000
1,628,508,403,000
1,628,508,403,000
MEMBER
null
## Describe the bug I would like to add a column to a downsampled dataset. However I get an error message saying the length don't match with the length of the unsampled dataset indicated. I suspect that the dataset size is not updated when calling `select`. ## Steps to reproduce the bug ```python from datasets im...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2768/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2768/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2767
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2767/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2767/comments
https://api.github.com/repos/huggingface/datasets/issues/2767/events
https://github.com/huggingface/datasets/issues/2767
963,002,120
MDU6SXNzdWU5NjMwMDIxMjA=
2,767
equal operation to perform unbatch for huggingface datasets
{ "login": "dorooddorood606", "id": 79288051, "node_id": "MDQ6VXNlcjc5Mjg4MDUx", "avatar_url": "https://avatars.githubusercontent.com/u/79288051?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dorooddorood606", "html_url": "https://github.com/dorooddorood606", "followers_url": "https://api...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi @lhoestq \r\nMaybe this is clearer to explain like this, currently map function, map one example to \"one\" modified one, lets assume we want to map one example to \"multiple\" examples, in which we do not know in advance how many examples they would be per each entry. I greatly appreciate telling me how I can ...
1,628,279,152,000
1,646,661,480,000
1,646,661,480,000
NONE
null
Hi I need to use "unbatch" operation in tensorflow on a huggingface dataset, I could not find this operation, could you kindly direct me how I can do it, here is the problem I am trying to solve: I am considering "record" dataset in SuperGlue and I need to replicate each entery of the dataset for each answer, to ma...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2767/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2767/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2766
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2766/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2766/comments
https://api.github.com/repos/huggingface/datasets/issues/2766/events
https://github.com/huggingface/datasets/pull/2766
962,994,198
MDExOlB1bGxSZXF1ZXN0NzA1NzAyNjM5
2,766
fix typo (ShuffingConfig -> ShufflingConfig)
{ "login": "daleevans", "id": 4944007, "node_id": "MDQ6VXNlcjQ5NDQwMDc=", "avatar_url": "https://avatars.githubusercontent.com/u/4944007?v=4", "gravatar_id": "", "url": "https://api.github.com/users/daleevans", "html_url": "https://github.com/daleevans", "followers_url": "https://api.github.com/users/da...
[]
closed
false
null
[]
null
[]
1,628,278,300,000
1,628,605,023,000
1,628,605,022,000
CONTRIBUTOR
null
pretty straightforward, it should be Shuffling instead of Shuffing
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2766/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2766/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2766", "html_url": "https://github.com/huggingface/datasets/pull/2766", "diff_url": "https://github.com/huggingface/datasets/pull/2766.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2766.patch", "merged_at": 1628605022000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2765
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2765/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2765/comments
https://api.github.com/repos/huggingface/datasets/issues/2765/events
https://github.com/huggingface/datasets/issues/2765
962,861,395
MDU6SXNzdWU5NjI4NjEzOTU=
2,765
BERTScore Error
{ "login": "gagan3012", "id": 49101362, "node_id": "MDQ6VXNlcjQ5MTAxMzYy", "avatar_url": "https://avatars.githubusercontent.com/u/49101362?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gagan3012", "html_url": "https://github.com/gagan3012", "followers_url": "https://api.github.com/users/...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi,\r\n\r\nThe `use_fast_tokenizer` argument has been recently added to the bert-score lib. I've opened a PR with the fix. In the meantime, you can try to downgrade the version of bert-score with the following command to make the code work:\r\n```\r\npip uninstall bert-score\r\npip install \"bert-score<0.3.10\"\r\...
1,628,265,537,000
1,628,507,785,000
1,628,507,785,000
NONE
null
## Describe the bug A clear and concise description of what the bug is. ## Steps to reproduce the bug ```python predictions = ["hello there", "general kenobi"] references = ["hello there", "general kenobi"] bert = load_metric('bertscore') bert.compute(predictions=predictions, references=references,lang='en') ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2765/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2765/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2764
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2764/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2764/comments
https://api.github.com/repos/huggingface/datasets/issues/2764/events
https://github.com/huggingface/datasets/pull/2764
962,554,799
MDExOlB1bGxSZXF1ZXN0NzA1MzI3MDQ5
2,764
Add DER metric for SUPERB speaker diarization task
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
open
false
null
[]
null
[]
1,628,241,156,000
1,628,244,413,000
null
MEMBER
null
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2764/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2764/timeline
null
null
true
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2764", "html_url": "https://github.com/huggingface/datasets/pull/2764", "diff_url": "https://github.com/huggingface/datasets/pull/2764.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2764.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/2763
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2763/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2763/comments
https://api.github.com/repos/huggingface/datasets/issues/2763/events
https://github.com/huggingface/datasets/issues/2763
961,895,523
MDU6SXNzdWU5NjE4OTU1MjM=
2,763
English wikipedia datasets is not clean
{ "login": "lucadiliello", "id": 23355969, "node_id": "MDQ6VXNlcjIzMzU1OTY5", "avatar_url": "https://avatars.githubusercontent.com/u/23355969?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lucadiliello", "html_url": "https://github.com/lucadiliello", "followers_url": "https://api.github.c...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
[ "Hi ! Certain users might need these data (for training or simply to explore/index the dataset).\r\n\r\nFeel free to implement a map function that gets rid of these paragraphs and process the wikipedia dataset with it before training" ]
1,628,174,244,000
1,629,738,016,000
null
CONTRIBUTOR
null
## Describe the bug Wikipedia english dumps contain many wikipedia paragraphs like "References", "Category:" and "See Also" that should not be used for training. ## Steps to reproduce the bug ```python # Sample code to reproduce the bug from datasets import load_dataset w = load_dataset('wikipedia', '20200501.e...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2763/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2763/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2762
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2762/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2762/comments
https://api.github.com/repos/huggingface/datasets/issues/2762/events
https://github.com/huggingface/datasets/issues/2762
961,652,046
MDU6SXNzdWU5NjE2NTIwNDY=
2,762
Add RVL-CDIP dataset
{ "login": "NielsRogge", "id": 48327001, "node_id": "MDQ6VXNlcjQ4MzI3MDAx", "avatar_url": "https://avatars.githubusercontent.com/u/48327001?v=4", "gravatar_id": "", "url": "https://api.github.com/users/NielsRogge", "html_url": "https://github.com/NielsRogge", "followers_url": "https://api.github.com/use...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" }, { "id": 3608941089, ...
closed
false
{ "login": "dnaveenr", "id": 17746528, "node_id": "MDQ6VXNlcjE3NzQ2NTI4", "avatar_url": "https://avatars.githubusercontent.com/u/17746528?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dnaveenr", "html_url": "https://github.com/dnaveenr", "followers_url": "https://api.github.com/users/dna...
[ { "login": "dnaveenr", "id": 17746528, "node_id": "MDQ6VXNlcjE3NzQ2NTI4", "avatar_url": "https://avatars.githubusercontent.com/u/17746528?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dnaveenr", "html_url": "https://github.com/dnaveenr", "followers_url": "https://api....
null
[ "cc @nateraw ", "#self-assign", "[labels_only.tar.gz](https://docs.google.com/uc?authuser=0&id=0B0NKIRwUL9KYcXo3bV9LU0t3SGs&export=download) on the RVL-CDIP website does not work for me.\r\n\r\n> 404. That’s an error. The requested URL was not found on this server.\r\n\r\nI contacted the author ( Adam Harley) r...
1,628,157,425,000
1,650,561,341,000
1,650,561,341,000
CONTRIBUTOR
null
## Adding a Dataset - **Name:** RVL-CDIP - **Description:** The RVL-CDIP (Ryerson Vision Lab Complex Document Information Processing) dataset consists of 400,000 grayscale images in 16 classes, with 25,000 images per class. There are 320,000 training images, 40,000 validation images, and 40,000 test images. The image...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2762/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2762/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2761
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2761/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2761/comments
https://api.github.com/repos/huggingface/datasets/issues/2761/events
https://github.com/huggingface/datasets/issues/2761
961,568,287
MDU6SXNzdWU5NjE1NjgyODc=
2,761
Error loading C4 realnewslike dataset
{ "login": "danshirron", "id": 32061512, "node_id": "MDQ6VXNlcjMyMDYxNTEy", "avatar_url": "https://avatars.githubusercontent.com/u/32061512?v=4", "gravatar_id": "", "url": "https://api.github.com/users/danshirron", "html_url": "https://github.com/danshirron", "followers_url": "https://api.github.com/use...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi @danshirron, \r\n`c4` was updated few days back by @lhoestq. The new configs are `['en', 'en.noclean', 'en.realnewslike', 'en.webtextlike'].` You'll need to remove any older version of this dataset you previously downloaded and then run `load_dataset` again with new configuration.", "@bhavitvyamalik @lhoestq ...
1,628,151,418,000
1,628,451,874,000
1,628,451,874,000
NONE
null
## Describe the bug Error loading C4 realnewslike dataset. Validation part mismatch ## Steps to reproduce the bug ```python raw_datasets = load_dataset('c4', 'realnewslike', cache_dir=model_args.cache_dir) ## Expected results success on data loading ## Actual results Downloading: 100%|███████████████████████...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2761/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2761/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2760
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2760/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2760/comments
https://api.github.com/repos/huggingface/datasets/issues/2760/events
https://github.com/huggingface/datasets/issues/2760
961,372,667
MDU6SXNzdWU5NjEzNzI2Njc=
2,760
Add Nuswide dataset
{ "login": "shivangibithel", "id": 19774925, "node_id": "MDQ6VXNlcjE5Nzc0OTI1", "avatar_url": "https://avatars.githubusercontent.com/u/19774925?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shivangibithel", "html_url": "https://github.com/shivangibithel", "followers_url": "https://api.gi...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" }, { "id": 3608941089, ...
open
false
null
[]
null
[]
1,628,132,441,000
1,638,965,183,000
null
NONE
null
## Adding a Dataset - **Name:** *NUSWIDE* - **Description:** *[A Real-World Web Image Dataset from National University of Singapore](https://lms.comp.nus.edu.sg/wp-content/uploads/2019/research/nuswide/NUS-WIDE.html)* - **Paper:** *[here](https://lms.comp.nus.edu.sg/wp-content/uploads/2019/research/nuswide/nuswide-c...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2760/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2760/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2758
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2758/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2758/comments
https://api.github.com/repos/huggingface/datasets/issues/2758/events
https://github.com/huggingface/datasets/pull/2758
960,206,575
MDExOlB1bGxSZXF1ZXN0NzAzMjQ5Nzky
2,758
Raise ManualDownloadError when loading a dataset that requires previous manual download
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,628,072,395,000
1,628,076,990,000
1,628,076,990,000
MEMBER
null
This PR implements the raising of a `ManualDownloadError` when loading a dataset that requires previous manual download, and this is missing. The `ManualDownloadError` is raised whether the dataset is loaded in normal or streaming mode. Close #2749. cc: @severo
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2758/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2758/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2758", "html_url": "https://github.com/huggingface/datasets/pull/2758", "diff_url": "https://github.com/huggingface/datasets/pull/2758.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2758.patch", "merged_at": 1628076990000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2757
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2757/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2757/comments
https://api.github.com/repos/huggingface/datasets/issues/2757/events
https://github.com/huggingface/datasets/issues/2757
959,984,081
MDU6SXNzdWU5NTk5ODQwODE=
2,757
Unexpected type after `concatenate_datasets`
{ "login": "JulesBelveze", "id": 32683010, "node_id": "MDQ6VXNlcjMyNjgzMDEw", "avatar_url": "https://avatars.githubusercontent.com/u/32683010?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JulesBelveze", "html_url": "https://github.com/JulesBelveze", "followers_url": "https://api.github.c...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Hi @JulesBelveze, thanks for your question.\r\n\r\nNote that 🤗 `datasets` internally store their data in Apache Arrow format.\r\n\r\nHowever, when accessing dataset columns, by default they are returned as native Python objects (lists in this case).\r\n\r\nIf you would like their columns to be returned in a more...
1,628,061,039,000
1,628,092,884,000
1,628,092,883,000
NONE
null
## Describe the bug I am trying to concatenate two `Dataset` using `concatenate_datasets` but it turns out that after concatenation the features are casted from `torch.Tensor` to `list`. It then leads to a weird tensors when trying to convert it to a `DataLoader`. However, if I use each `Dataset` separately everythi...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2757/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2757/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2756
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2756/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2756/comments
https://api.github.com/repos/huggingface/datasets/issues/2756/events
https://github.com/huggingface/datasets/pull/2756
959,255,646
MDExOlB1bGxSZXF1ZXN0NzAyMzk4Mjk1
2,756
Fix metadata JSON for ubuntu_dialogs_corpus dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,628,005,739,000
1,628,070,205,000
1,628,070,205,000
MEMBER
null
Related to #2743.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2756/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2756/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2756", "html_url": "https://github.com/huggingface/datasets/pull/2756", "diff_url": "https://github.com/huggingface/datasets/pull/2756.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2756.patch", "merged_at": 1628070205000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2755
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2755/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2755/comments
https://api.github.com/repos/huggingface/datasets/issues/2755/events
https://github.com/huggingface/datasets/pull/2755
959,115,888
MDExOlB1bGxSZXF1ZXN0NzAyMjgwMjI4
2,755
Fix metadata JSON for turkish_movie_sentiment dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,997,144,000
1,628,068,014,000
1,628,068,013,000
MEMBER
null
Related to #2743.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2755/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2755/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2755", "html_url": "https://github.com/huggingface/datasets/pull/2755", "diff_url": "https://github.com/huggingface/datasets/pull/2755.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2755.patch", "merged_at": 1628068013000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2754
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2754/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2754/comments
https://api.github.com/repos/huggingface/datasets/issues/2754/events
https://github.com/huggingface/datasets/pull/2754
959,105,577
MDExOlB1bGxSZXF1ZXN0NzAyMjcxMjM4
2,754
Generate metadata JSON for telugu_books dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,996,492,000
1,628,066,942,000
1,628,066,942,000
MEMBER
null
Related to #2743.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2754/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2754/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2754", "html_url": "https://github.com/huggingface/datasets/pull/2754", "diff_url": "https://github.com/huggingface/datasets/pull/2754.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2754.patch", "merged_at": 1628066941000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2753
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2753/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2753/comments
https://api.github.com/repos/huggingface/datasets/issues/2753/events
https://github.com/huggingface/datasets/pull/2753
959,036,995
MDExOlB1bGxSZXF1ZXN0NzAyMjEyMjMz
2,753
Generate metadata JSON for reclor dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,991,549,000
1,628,064,435,000
1,628,064,435,000
MEMBER
null
Related to #2743.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2753/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2753/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2753", "html_url": "https://github.com/huggingface/datasets/pull/2753", "diff_url": "https://github.com/huggingface/datasets/pull/2753.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2753.patch", "merged_at": 1628064435000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2752
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2752/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2752/comments
https://api.github.com/repos/huggingface/datasets/issues/2752/events
https://github.com/huggingface/datasets/pull/2752
959,023,608
MDExOlB1bGxSZXF1ZXN0NzAyMjAxMjAy
2,752
Generate metadata JSON for lm1b dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,990,496,000
1,628,059,240,000
1,628,059,239,000
MEMBER
null
Related to #2743.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2752/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2752/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2752", "html_url": "https://github.com/huggingface/datasets/pull/2752", "diff_url": "https://github.com/huggingface/datasets/pull/2752.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2752.patch", "merged_at": 1628059239000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2751
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2751/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2751/comments
https://api.github.com/repos/huggingface/datasets/issues/2751/events
https://github.com/huggingface/datasets/pull/2751
959,021,262
MDExOlB1bGxSZXF1ZXN0NzAyMTk5MjA5
2,751
Update metadata for wikihow dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,990,317,000
1,628,005,929,000
1,628,005,929,000
MEMBER
null
Update metadata for wikihow dataset: - Remove leading new line character in description and citation - Update metadata JSON - Remove no longer necessary `urls_checksums/checksums.txt` file Related to #2748.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2751/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2751/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2751", "html_url": "https://github.com/huggingface/datasets/pull/2751", "diff_url": "https://github.com/huggingface/datasets/pull/2751.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2751.patch", "merged_at": 1628005929000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2750
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2750/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2750/comments
https://api.github.com/repos/huggingface/datasets/issues/2750/events
https://github.com/huggingface/datasets/issues/2750
958,984,730
MDU6SXNzdWU5NTg5ODQ3MzA=
2,750
Second concatenation of datasets produces errors
{ "login": "Aktsvigun", "id": 36672861, "node_id": "MDQ6VXNlcjM2NjcyODYx", "avatar_url": "https://avatars.githubusercontent.com/u/36672861?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Aktsvigun", "html_url": "https://github.com/Aktsvigun", "followers_url": "https://api.github.com/users/...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "@albertvillanova ", "Hi @Aktsvigun, thanks for reporting.\r\n\r\nI'm investigating this.", "Hi @albertvillanova ,\r\nany update on this? Can I probably help in some way?", "Hi @Aktsvigun! We are planning to address this issue before our next release, in a couple of weeks at most. 😅 \r\n\r\nIn the meantime, ...
1,627,987,624,000
1,642,602,223,000
1,642,601,945,000
NONE
null
Hi, I am need to concatenate my dataset with others several times, and after I concatenate it for the second time, the features of features (e.g. tags names) are collapsed. This hinders, for instance, the usage of tokenize function with `data.map`. ``` from datasets import load_dataset, concatenate_datasets d...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2750/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2750/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2749
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2749/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2749/comments
https://api.github.com/repos/huggingface/datasets/issues/2749/events
https://github.com/huggingface/datasets/issues/2749
958,968,748
MDU6SXNzdWU5NTg5Njg3NDg=
2,749
Raise a proper exception when trying to stream a dataset that requires to manually download files
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/foll...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Hi @severo, thanks for reporting.\r\n\r\nAs discussed, datasets requiring manual download should be:\r\n- programmatically identifiable\r\n- properly handled with more clear error message when trying to load them with streaming\r\n\r\nIn relation with programmatically identifiability, note that for datasets requir...
1,627,986,387,000
1,628,499,215,000
1,628,076,990,000
CONTRIBUTOR
null
## Describe the bug At least for 'reclor', 'telugu_books', 'turkish_movie_sentiment', 'ubuntu_dialogs_corpus', 'wikihow', trying to `load_dataset` in streaming mode raises a `TypeError` without any detail about why it fails. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2749/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2749/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2748
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2748/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2748/comments
https://api.github.com/repos/huggingface/datasets/issues/2748/events
https://github.com/huggingface/datasets/pull/2748
958,889,041
MDExOlB1bGxSZXF1ZXN0NzAyMDg4NTk4
2,748
Generate metadata JSON for wikihow dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,980,940,000
1,627,985,871,000
1,627,985,871,000
MEMBER
null
Related to #2743.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2748/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2748/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2748", "html_url": "https://github.com/huggingface/datasets/pull/2748", "diff_url": "https://github.com/huggingface/datasets/pull/2748.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2748.patch", "merged_at": 1627985871000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2747
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2747/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2747/comments
https://api.github.com/repos/huggingface/datasets/issues/2747/events
https://github.com/huggingface/datasets/pull/2747
958,867,627
MDExOlB1bGxSZXF1ZXN0NzAyMDcwOTgy
2,747
add multi-proc in `to_json`
{ "login": "bhavitvyamalik", "id": 19718818, "node_id": "MDQ6VXNlcjE5NzE4ODE4", "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bhavitvyamalik", "html_url": "https://github.com/bhavitvyamalik", "followers_url": "https://api.gi...
[]
closed
false
null
[]
null
[ "Thank you for working on this, @bhavitvyamalik \r\n\r\n10% is not solving the issue, we want 5-10x faster on a machine that has lots of resources, but limited processing time.\r\n\r\nSo let's benchmark it on an instance with many more cores, I can test with 12 on my dev box and 40 on JZ. \r\n\r\nCould you please s...
1,627,979,413,000
1,634,667,861,000
1,631,541,397,000
CONTRIBUTOR
null
Closes #2663. I've tried adding multiprocessing in `to_json`. Here's some benchmarking I did to compare the timings of current version (say v1) and multi-proc version (say v2). I did this with `cpu_count` 4 (2015 Macbook Air) 1. Dataset name: `ascent_kb` - 8.9M samples (all samples were used, reporting this for a si...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2747/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2747/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2747", "html_url": "https://github.com/huggingface/datasets/pull/2747", "diff_url": "https://github.com/huggingface/datasets/pull/2747.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2747.patch", "merged_at": 1631541397000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2746
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2746/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2746/comments
https://api.github.com/repos/huggingface/datasets/issues/2746/events
https://github.com/huggingface/datasets/issues/2746
958,551,619
MDU6SXNzdWU5NTg1NTE2MTk=
2,746
Cannot load `few-nerd` dataset
{ "login": "Mehrad0711", "id": 28717374, "node_id": "MDQ6VXNlcjI4NzE3Mzc0", "avatar_url": "https://avatars.githubusercontent.com/u/28717374?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Mehrad0711", "html_url": "https://github.com/Mehrad0711", "followers_url": "https://api.github.com/use...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi @Mehrad0711,\r\n\r\nI'm afraid there is no \"canonical\" Hugging Face dataset named \"few-nerd\".\r\n\r\nThere are 2 kinds of datasets hosted at the Hugging Face Hub:\r\n- canonical datasets (their identifier contains no slash \"/\"): we, the Hugging Face team, supervise their implementation and we make sure th...
1,627,942,737,000
1,637,052,694,000
1,628,019,943,000
NONE
null
## Describe the bug Cannot load `few-nerd` dataset. ## Steps to reproduce the bug ```python from datasets import load_dataset load_dataset('few-nerd', 'supervised') ``` ## Actual results Executing above code will give the following error: ``` Using the latest cached version of the module from /Users...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2746/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2746/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2745
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2745/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2745/comments
https://api.github.com/repos/huggingface/datasets/issues/2745/events
https://github.com/huggingface/datasets/pull/2745
958,269,579
MDExOlB1bGxSZXF1ZXN0NzAxNTc0Mjcz
2,745
added semeval18_emotion_classification dataset
{ "login": "maxpel", "id": 31095360, "node_id": "MDQ6VXNlcjMxMDk1MzYw", "avatar_url": "https://avatars.githubusercontent.com/u/31095360?v=4", "gravatar_id": "", "url": "https://api.github.com/users/maxpel", "html_url": "https://github.com/maxpel", "followers_url": "https://api.github.com/users/maxpel/fo...
[]
closed
false
null
[]
null
[ "For training the multilabel classifier, I would combine the labels into a list, for example for the English dataset:\r\n\r\n```\r\ndfpre=pd.read_csv(path+\"2018-E-c-En-train.txt\",sep=\"\\t\")\r\ndfpre['list'] = dfpre[dfpre.columns[2:]].values.tolist()\r\ndf = dfpre[['Tweet', 'list']].copy()\r\ndf.rename(columns={...
1,627,918,795,000
1,635,499,325,000
1,632,217,715,000
CONTRIBUTOR
null
I added the data set of SemEval 2018 Task 1 (Subtask 5) for emotion detection in three languages. ``` datasets-cli test datasets/semeval18_emotion_classification/ --save_infos --all_configs RUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_semeval18_emotion_classification ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2745/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2745/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2745", "html_url": "https://github.com/huggingface/datasets/pull/2745", "diff_url": "https://github.com/huggingface/datasets/pull/2745.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2745.patch", "merged_at": 1632217715000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2744
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2744/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2744/comments
https://api.github.com/repos/huggingface/datasets/issues/2744/events
https://github.com/huggingface/datasets/pull/2744
958,146,637
MDExOlB1bGxSZXF1ZXN0NzAxNDY4NDcz
2,744
Fix key by recreating metadata JSON for journalists_questions dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,910,873,000
1,627,982,734,000
1,627,982,733,000
MEMBER
null
Close #2743.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2744/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2744/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2744", "html_url": "https://github.com/huggingface/datasets/pull/2744", "diff_url": "https://github.com/huggingface/datasets/pull/2744.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2744.patch", "merged_at": 1627982733000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2743
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2743/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2743/comments
https://api.github.com/repos/huggingface/datasets/issues/2743/events
https://github.com/huggingface/datasets/issues/2743
958,119,251
MDU6SXNzdWU5NTgxMTkyNTE=
2,743
Dataset JSON is incorrect
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/foll...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "As discussed, the metadata JSON files must be regenerated because the keys were nor properly generated and they will not be read by the builder:\r\n> Indeed there is some problem/bug while reading the datasets_info.json file: there is a mismatch with the config.name keys in the file...\r\nIn the meanwhile, in orde...
1,627,909,286,000
1,627,985,217,000
1,627,982,733,000
CONTRIBUTOR
null
## Describe the bug The JSON file generated for https://github.com/huggingface/datasets/blob/573f3d35081cee239d1b962878206e9abe6cde91/datasets/journalists_questions/journalists_questions.py is https://github.com/huggingface/datasets/blob/573f3d35081cee239d1b962878206e9abe6cde91/datasets/journalists_questions/dataset...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2743/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2743/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2742
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2742/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2742/comments
https://api.github.com/repos/huggingface/datasets/issues/2742/events
https://github.com/huggingface/datasets/issues/2742
958,114,064
MDU6SXNzdWU5NTgxMTQwNjQ=
2,742
Improve detection of streamable file types
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/foll...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 3470211881, "node_id": "LA_k...
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "maybe we should rather attempt to download a `Range` from the server and see if it works?" ]
1,627,908,909,000
1,636,737,490,000
1,636,737,490,000
CONTRIBUTOR
null
**Is your feature request related to a problem? Please describe.** ```python from datasets import load_dataset_builder from datasets.utils.streaming_download_manager import StreamingDownloadManager builder = load_dataset_builder("journalists_questions", name="plain_text") builder._split_generators(StreamingDownl...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2742/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2742/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2741
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2741/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2741/comments
https://api.github.com/repos/huggingface/datasets/issues/2741/events
https://github.com/huggingface/datasets/issues/2741
957,979,559
MDU6SXNzdWU5NTc5Nzk1NTk=
2,741
Add Hypersim dataset
{ "login": "osanseviero", "id": 7246357, "node_id": "MDQ6VXNlcjcyNDYzNTc=", "avatar_url": "https://avatars.githubusercontent.com/u/7246357?v=4", "gravatar_id": "", "url": "https://api.github.com/users/osanseviero", "html_url": "https://github.com/osanseviero", "followers_url": "https://api.github.com/us...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" }, { "id": 3608941089, ...
open
false
null
[]
null
[]
1,627,898,810,000
1,638,965,211,000
null
MEMBER
null
## Adding a Dataset - **Name:** Hypersim - **Description:** photorealistic synthetic dataset for holistic indoor scene understanding - **Paper:** *link to the dataset paper if available* - **Data:** https://github.com/apple/ml-hypersim Instructions to add a new dataset can be found [here](https://github.com/hugg...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2741/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2741/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2740
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2740/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2740/comments
https://api.github.com/repos/huggingface/datasets/issues/2740/events
https://github.com/huggingface/datasets/pull/2740
957,911,035
MDExOlB1bGxSZXF1ZXN0NzAxMjY0NTI3
2,740
Update release instructions
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,893,960,000
1,627,915,196,000
1,627,915,196,000
MEMBER
null
Update release instructions.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2740/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2740/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2740", "html_url": "https://github.com/huggingface/datasets/pull/2740", "diff_url": "https://github.com/huggingface/datasets/pull/2740.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2740.patch", "merged_at": 1627915196000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2739
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2739/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2739/comments
https://api.github.com/repos/huggingface/datasets/issues/2739/events
https://github.com/huggingface/datasets/pull/2739
957,751,260
MDExOlB1bGxSZXF1ZXN0NzAxMTI0ODQ3
2,739
Pass tokenize to sacrebleu only if explicitly passed by user
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,880,945,000
1,627,964,617,000
1,627,964,617,000
MEMBER
null
Next `sacrebleu` release (v2.0.0) will remove `sacrebleu.DEFAULT_TOKENIZER`: https://github.com/mjpost/sacrebleu/pull/152/files#diff-2553a315bb1f7e68c9c1b00d56eaeb74f5205aeb3a189bc3e527b122c6078795L17-R15 This PR passes `tokenize` to `sacrebleu` only if explicitly passed by the user, otherwise it will not pass it (a...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2739/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2739/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2739", "html_url": "https://github.com/huggingface/datasets/pull/2739", "diff_url": "https://github.com/huggingface/datasets/pull/2739.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2739.patch", "merged_at": 1627964617000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2738
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2738/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2738/comments
https://api.github.com/repos/huggingface/datasets/issues/2738/events
https://github.com/huggingface/datasets/pull/2738
957,517,746
MDExOlB1bGxSZXF1ZXN0NzAwOTI5NzA4
2,738
Sunbird AI Ugandan low resource language dataset
{ "login": "ak3ra", "id": 12105163, "node_id": "MDQ6VXNlcjEyMTA1MTYz", "avatar_url": "https://avatars.githubusercontent.com/u/12105163?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ak3ra", "html_url": "https://github.com/ak3ra", "followers_url": "https://api.github.com/users/ak3ra/follow...
[]
open
false
null
[]
null
[ "Hi @ak3ra , have you had a chance to take my comments into account ?\r\n\r\nLet me know if you have questions or if I can help :)", "@lhoestq Working on this, thanks for the detailed review :) ", "Hi ! Cool thanks :)\r\nFeel free to merge master into your branch to fix the CI issues\r\n\r\nLet me know if you ...
1,627,831,080,000
1,654,847,202,000
null
NONE
null
Multi-way parallel text corpus of 5 key Ugandan languages for the task of machine translation.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2738/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2738/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2738", "html_url": "https://github.com/huggingface/datasets/pull/2738", "diff_url": "https://github.com/huggingface/datasets/pull/2738.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2738.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/2737
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2737/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2737/comments
https://api.github.com/repos/huggingface/datasets/issues/2737/events
https://github.com/huggingface/datasets/issues/2737
957,124,881
MDU6SXNzdWU5NTcxMjQ4ODE=
2,737
SacreBLEU update
{ "login": "devrimcavusoglu", "id": 46989091, "node_id": "MDQ6VXNlcjQ2OTg5MDkx", "avatar_url": "https://avatars.githubusercontent.com/u/46989091?v=4", "gravatar_id": "", "url": "https://api.github.com/users/devrimcavusoglu", "html_url": "https://github.com/devrimcavusoglu", "followers_url": "https://api...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi @devrimcavusoglu, \r\nI tried your code with latest version of `datasets`and `sacrebleu==1.5.1` and it's running fine after changing one small thing:\r\n```\r\nsacrebleu = datasets.load_metric('sacrebleu')\r\npredictions = [\"It is a guide to action which ensures that the military always obeys the commands of t...
1,627,689,188,000
1,632,307,661,000
1,627,964,617,000
NONE
null
With the latest release of [sacrebleu](https://github.com/mjpost/sacrebleu), `datasets.metrics.sacrebleu` is broken, and getting error. AttributeError: module 'sacrebleu' has no attribute 'DEFAULT_TOKENIZER' this happens since in new version of sacrebleu there is no `DEFAULT_TOKENIZER`, but sacrebleu.py tries...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2737/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2737/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2736
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2736/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2736/comments
https://api.github.com/repos/huggingface/datasets/issues/2736/events
https://github.com/huggingface/datasets/issues/2736
956,895,199
MDU6SXNzdWU5NTY4OTUxOTk=
2,736
Add Microsoft Building Footprints dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" }, { "id": 3608941089, ...
open
false
null
[]
null
[ "Motivation: this can be a useful dataset for researchers working on climate change adaptation, urban studies, geography, etc. I'll see if I can figure out how to add it!" ]
1,627,661,828,000
1,638,965,343,000
null
MEMBER
null
## Adding a Dataset - **Name:** Microsoft Building Footprints - **Description:** With the goal to increase the coverage of building footprint data available as open data for OpenStreetMap and humanitarian efforts, we have released millions of building footprints as open data available to download free of charge. - *...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2736/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2736/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2735
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2735/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2735/comments
https://api.github.com/repos/huggingface/datasets/issues/2735/events
https://github.com/huggingface/datasets/issues/2735
956,889,365
MDU6SXNzdWU5NTY4ODkzNjU=
2,735
Add Open Buildings dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[]
1,627,661,319,000
1,627,707,685,000
null
MEMBER
null
## Adding a Dataset - **Name:** Open Buildings - **Description:** A dataset of building footprints to support social good applications. Building footprints are useful for a range of important applications, from population estimation, urban planning and humanitarian response, to environmental and climate science....
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2735/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2735/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2734
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2734/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2734/comments
https://api.github.com/repos/huggingface/datasets/issues/2734/events
https://github.com/huggingface/datasets/pull/2734
956,844,874
MDExOlB1bGxSZXF1ZXN0NzAwMzc4NjI4
2,734
Update BibTeX entry
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,658,571,000
1,627,660,078,000
1,627,660,078,000
MEMBER
null
Update BibTeX entry.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2734/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2734/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2734", "html_url": "https://github.com/huggingface/datasets/pull/2734", "diff_url": "https://github.com/huggingface/datasets/pull/2734.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2734.patch", "merged_at": 1627660078000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2733
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2733/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2733/comments
https://api.github.com/repos/huggingface/datasets/issues/2733/events
https://github.com/huggingface/datasets/pull/2733
956,725,476
MDExOlB1bGxSZXF1ZXN0NzAwMjc1NDMy
2,733
Add missing parquet known extension
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,627,650,080,000
1,627,651,471,000
1,627,651,470,000
MEMBER
null
This code was failing because the parquet extension wasn't recognized: ```python from datasets import load_dataset base_url = "https://storage.googleapis.com/huggingface-nlp/cache/datasets/wikipedia/20200501.en/1.0.0/" data_files = {"train": base_url + "wikipedia-train.parquet"} wiki = load_dataset("parquet", da...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2733/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2733/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2733", "html_url": "https://github.com/huggingface/datasets/pull/2733", "diff_url": "https://github.com/huggingface/datasets/pull/2733.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2733.patch", "merged_at": 1627651470000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2732
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2732/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2732/comments
https://api.github.com/repos/huggingface/datasets/issues/2732/events
https://github.com/huggingface/datasets/pull/2732
956,676,360
MDExOlB1bGxSZXF1ZXN0NzAwMjMzMzQy
2,732
Updated TTC4900 Dataset
{ "login": "yavuzKomecoglu", "id": 5150963, "node_id": "MDQ6VXNlcjUxNTA5NjM=", "avatar_url": "https://avatars.githubusercontent.com/u/5150963?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yavuzKomecoglu", "html_url": "https://github.com/yavuzKomecoglu", "followers_url": "https://api.gith...
[]
closed
false
null
[]
null
[ "@lhoestq, lütfen bu PR'ı gözden geçirebilir misiniz?", "> Thanks ! This looks all good now :)\r\n\r\nThanks" ]
1,627,645,934,000
1,627,660,851,000
1,627,660,694,000
CONTRIBUTOR
null
- The source address of the TTC4900 dataset of [@savasy](https://github.com/savasy) has been updated for direct download. - Updated readme.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2732/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2732/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2732", "html_url": "https://github.com/huggingface/datasets/pull/2732", "diff_url": "https://github.com/huggingface/datasets/pull/2732.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2732.patch", "merged_at": 1627660694000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2731
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2731/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2731/comments
https://api.github.com/repos/huggingface/datasets/issues/2731/events
https://github.com/huggingface/datasets/pull/2731
956,087,452
MDExOlB1bGxSZXF1ZXN0Njk5NzQwMjg5
2,731
Adding to_tf_dataset method
{ "login": "Rocketknight1", "id": 12866554, "node_id": "MDQ6VXNlcjEyODY2NTU0", "avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Rocketknight1", "html_url": "https://github.com/Rocketknight1", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "This seems to be working reasonably well in testing, and performance is way better. `tf.py_function` has been dropped for an input generator, but I moved as much of the code as possible outside the generator to allow TF to compile it correctly. I also avoid `tf.RaggedTensor` at all costs, and do the shuffle in the...
1,627,582,225,000
1,631,800,254,000
1,631,800,254,000
MEMBER
null
Oh my **god** do not merge this yet, it's just a draft. I've added a method (via a mixin) to the `arrow_dataset.Dataset` class that automatically converts our Dataset classes to TF Dataset classes ready for training. It hopefully has most of the features we want, including streaming from disk (no need to load the wh...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2731/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2731/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2731", "html_url": "https://github.com/huggingface/datasets/pull/2731", "diff_url": "https://github.com/huggingface/datasets/pull/2731.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2731.patch", "merged_at": 1631800253000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2730
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2730/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2730/comments
https://api.github.com/repos/huggingface/datasets/issues/2730/events
https://github.com/huggingface/datasets/issues/2730
955,987,834
MDU6SXNzdWU5NTU5ODc4MzQ=
2,730
Update CommonVoice with new release
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[ "cc @patrickvonplaten?", "Does anybody know if there is a bundled link, which would allow direct data download instead of manual? \r\nSomething similar to: `https://voice-prod-bundler-ee1969a6ce8178826482b88e843c335139bd3fb4.s3.amazonaws.com/cv-corpus-6.1-2020-12-11/ab.tar.gz` ? cc @patil-suraj \r\n", "Also see...
1,627,574,399,000
1,628,353,159,000
null
MEMBER
null
## Adding a Dataset - **Name:** CommonVoice mid-2021 release - **Description:** more data in CommonVoice: Languages that have increased the most by percentage are Thai (almost 20x growth, from 12 hours to 250 hours), Luganda (almost 9x growth, from 8 to 80), Esperanto (7x growth, from 100 to 840), and Tamil (almost 8...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2730/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2730/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2729
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2729/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2729/comments
https://api.github.com/repos/huggingface/datasets/issues/2729/events
https://github.com/huggingface/datasets/pull/2729
955,920,489
MDExOlB1bGxSZXF1ZXN0Njk5NTk5MjA4
2,729
Fix IndexError while loading Arabic Billion Words dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[]
1,627,570,022,000
1,627,650,235,000
1,627,650,235,000
MEMBER
null
Catch `IndexError` and ignore that record. Close #2727.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2729/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2729/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2729", "html_url": "https://github.com/huggingface/datasets/pull/2729", "diff_url": "https://github.com/huggingface/datasets/pull/2729.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2729.patch", "merged_at": 1627650235000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2728
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2728/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2728/comments
https://api.github.com/repos/huggingface/datasets/issues/2728/events
https://github.com/huggingface/datasets/issues/2728
955,892,970
MDU6SXNzdWU5NTU4OTI5NzA=
2,728
Concurrent use of same dataset (already downloaded)
{ "login": "PierreColombo", "id": 22492839, "node_id": "MDQ6VXNlcjIyNDkyODM5", "avatar_url": "https://avatars.githubusercontent.com/u/22492839?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PierreColombo", "html_url": "https://github.com/PierreColombo", "followers_url": "https://api.githu...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
[ "Launching simultaneous job relying on the same datasets try some writing issue. I guess it is unexpected since I only need to load some already downloaded file.", "If i have two jobs that use the same dataset. I got :\r\n\r\n\r\n File \"compute_measures.py\", line 181, in <module>\r\n train_loader, val_loade...
1,627,568,318,000
1,627,889,157,000
null
CONTRIBUTOR
null
## Describe the bug When launching several jobs at the same time loading the same dataset trigger some errors see (last comments). ## Steps to reproduce the bug export HF_DATASETS_CACHE=/gpfswork/rech/toto/datasets for MODEL in "bert-base-uncased" "roberta-base" "distilbert-base-cased"; do # "bert-base-uncased" ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2728/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2728/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2727
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2727/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2727/comments
https://api.github.com/repos/huggingface/datasets/issues/2727/events
https://github.com/huggingface/datasets/issues/2727
955,812,149
MDU6SXNzdWU5NTU4MTIxNDk=
2,727
Error in loading the Arabic Billion Words Corpus
{ "login": "M-Salti", "id": 9285264, "node_id": "MDQ6VXNlcjkyODUyNjQ=", "avatar_url": "https://avatars.githubusercontent.com/u/9285264?v=4", "gravatar_id": "", "url": "https://api.github.com/users/M-Salti", "html_url": "https://github.com/M-Salti", "followers_url": "https://api.github.com/users/M-Salti/...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "I modified the dataset loading script to catch the `IndexError` and inspect the records at which the error is happening, and I found this:\r\nFor the `Techreen` config, the error happens in 36 records when trying to find the `Text` or `Dateline` tags. All these 36 records look something like:\r\n```\r\n<Techreen>\...
1,627,563,189,000
1,627,650,235,000
1,627,650,235,000
CONTRIBUTOR
null
## Describe the bug I get `IndexError: list index out of range` when trying to load the `Techreen` and `Almustaqbal` configs of the dataset. ## Steps to reproduce the bug ```python load_dataset("arabic_billion_words", "Techreen") load_dataset("arabic_billion_words", "Almustaqbal") ``` ## Expected results Th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2727/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2727/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2726
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2726/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2726/comments
https://api.github.com/repos/huggingface/datasets/issues/2726/events
https://github.com/huggingface/datasets/pull/2726
955,674,388
MDExOlB1bGxSZXF1ZXN0Njk5Mzg5MDk1
2,726
Typo fix `tokenize_exemple`
{ "login": "shabie", "id": 30535146, "node_id": "MDQ6VXNlcjMwNTM1MTQ2", "avatar_url": "https://avatars.githubusercontent.com/u/30535146?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shabie", "html_url": "https://github.com/shabie", "followers_url": "https://api.github.com/users/shabie/fo...
[]
closed
false
null
[]
null
[]
1,627,553,017,000
1,627,560,025,000
1,627,560,025,000
CONTRIBUTOR
null
There is a small typo in the main README.md
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2726/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2726/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2726", "html_url": "https://github.com/huggingface/datasets/pull/2726", "diff_url": "https://github.com/huggingface/datasets/pull/2726.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2726.patch", "merged_at": 1627560025000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2725
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2725/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2725/comments
https://api.github.com/repos/huggingface/datasets/issues/2725/events
https://github.com/huggingface/datasets/pull/2725
955,020,776
MDExOlB1bGxSZXF1ZXN0Njk4ODMwNjYw
2,725
Pass use_auth_token to request_etags
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,488,809,000
1,627,490,282,000
1,627,490,282,000
MEMBER
null
Fix #2724.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2725/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2725/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2725", "html_url": "https://github.com/huggingface/datasets/pull/2725", "diff_url": "https://github.com/huggingface/datasets/pull/2725.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2725.patch", "merged_at": 1627490281000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2724
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2724/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2724/comments
https://api.github.com/repos/huggingface/datasets/issues/2724/events
https://github.com/huggingface/datasets/issues/2724
954,919,607
MDU6SXNzdWU5NTQ5MTk2MDc=
2,724
404 Error when loading remote data files from private repo
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "I guess the issue is when computing the ETags of the remote files. Indeed `use_auth_token` must be passed to `request_etags` here:\r\n\r\nhttps://github.com/huggingface/datasets/blob/35b5e4bc0cb2ed896e40f3eb2a4aa3de1cb1a6c5/src/datasets/builder.py#L160-L160", "Yes, I remember having properly implemented that: \r...
1,627,482,263,000
1,627,534,729,000
1,627,490,281,000
MEMBER
null
## Describe the bug When loading remote data files from a private repo, a 404 error is raised. ## Steps to reproduce the bug ```python url = hf_hub_url("lewtun/asr-preds-test", "preds.jsonl", repo_type="dataset") dset = load_dataset("json", data_files=url, use_auth_token=True) # HTTPError: 404 Client Error: Not...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2724/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2724/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2723
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2723/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2723/comments
https://api.github.com/repos/huggingface/datasets/issues/2723/events
https://github.com/huggingface/datasets/pull/2723
954,864,104
MDExOlB1bGxSZXF1ZXN0Njk4Njk0NDMw
2,723
Fix en subset by modifying dataset_info with correct validation infos
{ "login": "thomasw21", "id": 24695242, "node_id": "MDQ6VXNlcjI0Njk1MjQy", "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomasw21", "html_url": "https://github.com/thomasw21", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[]
1,627,479,379,000
1,627,485,743,000
1,627,485,743,000
MEMBER
null
- Related to: #2682 We correct the values of `en` subset concerning the expected validation values (both `num_bytes` and `num_examples`. Instead of having: `{"name": "validation", "num_bytes": 828589180707, "num_examples": 364868892, "dataset_name": "c4"}` We replace with correct values: `{"name": "vali...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2723/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2723/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2723", "html_url": "https://github.com/huggingface/datasets/pull/2723", "diff_url": "https://github.com/huggingface/datasets/pull/2723.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2723.patch", "merged_at": 1627485743000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2722
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2722/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2722/comments
https://api.github.com/repos/huggingface/datasets/issues/2722/events
https://github.com/huggingface/datasets/issues/2722
954,446,053
MDU6SXNzdWU5NTQ0NDYwNTM=
2,722
Missing cache file
{ "login": "PosoSAgapo", "id": 33200481, "node_id": "MDQ6VXNlcjMzMjAwNDgx", "avatar_url": "https://avatars.githubusercontent.com/u/33200481?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PosoSAgapo", "html_url": "https://github.com/PosoSAgapo", "followers_url": "https://api.github.com/use...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "This could be solved by going to the glue/ directory and delete sst2 directory, then load the dataset again will help you redownload the dataset.", "Hi ! Not sure why this file was missing, but yes the way to fix this is to delete the sst2 directory and to reload the dataset" ]
1,627,444,327,000
1,647,851,271,000
1,647,851,271,000
NONE
null
Strangely missing cache file after I restart my program again. `glue_dataset = datasets.load_dataset('glue', 'sst2')` `FileNotFoundError: [Errno 2] No such file or directory: /Users/chris/.cache/huggingface/datasets/glue/sst2/1.0.0/dacbe3125aa31d7f70367a07a8a9e72a5a0bfeb5fc42e75c9db75b96d6053ad/dataset_info.json...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2722/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2722/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2721
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2721/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2721/comments
https://api.github.com/repos/huggingface/datasets/issues/2721/events
https://github.com/huggingface/datasets/pull/2721
954,238,230
MDExOlB1bGxSZXF1ZXN0Njk4MTY0Njg3
2,721
Deal with the bad check in test_load.py
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "Hi ! I did a change for this test already in #2662 :\r\n\r\nhttps://github.com/huggingface/datasets/blob/00686c46b7aaf6bfcd4102cec300a3c031284a5a/tests/test_load.py#L312-L316\r\n\r\n(though I have to change the variable name `m_combined_path` to `m_url` or something)\r\n\r\nI guess it's ok to remove this check for...
1,627,417,403,000
1,627,466,314,000
1,627,462,398,000
CONTRIBUTOR
null
This PR removes a check that's been added in #2684. My intention with this check was to capture an URL in the error message, but instead, it captures a substring of the previous regex match in the test function. Another option would be to replace this check with: ```python m_paths = re.findall(r"\S*_dummy/_dummy.py\b...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2721/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2721/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2721", "html_url": "https://github.com/huggingface/datasets/pull/2721", "diff_url": "https://github.com/huggingface/datasets/pull/2721.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2721.patch", "merged_at": 1627462398000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2720
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2720/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2720/comments
https://api.github.com/repos/huggingface/datasets/issues/2720/events
https://github.com/huggingface/datasets/pull/2720
954,024,426
MDExOlB1bGxSZXF1ZXN0Njk3OTgxNjMx
2,720
fix: 🐛 fix two typos
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/foll...
[]
closed
false
null
[]
null
[]
1,627,401,017,000
1,627,411,097,000
1,627,411,096,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2720/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2720/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2720", "html_url": "https://github.com/huggingface/datasets/pull/2720", "diff_url": "https://github.com/huggingface/datasets/pull/2720.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2720.patch", "merged_at": 1627411096000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2719
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2719/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2719/comments
https://api.github.com/repos/huggingface/datasets/issues/2719/events
https://github.com/huggingface/datasets/issues/2719
953,932,416
MDU6SXNzdWU5NTM5MzI0MTY=
2,719
Use ETag in streaming mode to detect resource updates
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/foll...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 3470211881, "node_id": "LA_k...
open
false
null
[]
null
[]
1,627,395,429,000
1,634,895,368,000
null
CONTRIBUTOR
null
**Is your feature request related to a problem? Please describe.** I want to cache data I generate from processing a dataset I've loaded in streaming mode, but I've currently no way to know if the remote data has been updated or not, thus I don't know when to invalidate my cache. **Describe the solution you'd lik...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2719/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2719/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2718
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2718/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2718/comments
https://api.github.com/repos/huggingface/datasets/issues/2718/events
https://github.com/huggingface/datasets/pull/2718
953,360,663
MDExOlB1bGxSZXF1ZXN0Njk3NDE0NTQy
2,718
New documentation structure
{ "login": "stevhliu", "id": 59462357, "node_id": "MDQ6VXNlcjU5NDYyMzU3", "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stevhliu", "html_url": "https://github.com/stevhliu", "followers_url": "https://api.github.com/users/ste...
[]
closed
false
null
[]
null
[ "I just did some minor changes + added some content in these sections: share, about arrow, about cache\r\n\r\nFeel free to mark this PR as ready for review ! :)", "I just separated the `Share` How-to page into three pages: share, dataset_script and dataset_card.\r\n\r\nThis way in the share page we can explain in...
1,627,341,313,000
1,631,553,653,000
1,631,553,652,000
MEMBER
null
Organize Datasets documentation into four documentation types to improve clarity and discoverability of content. **Content to add in the very short term (feel free to add anything I'm missing):** - A discussion on why Datasets uses Arrow that includes some context and background about why we use Arrow. Would also b...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2718/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2718/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2718", "html_url": "https://github.com/huggingface/datasets/pull/2718", "diff_url": "https://github.com/huggingface/datasets/pull/2718.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2718.patch", "merged_at": 1631553652000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2717
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2717/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2717/comments
https://api.github.com/repos/huggingface/datasets/issues/2717/events
https://github.com/huggingface/datasets/pull/2717
952,979,976
MDExOlB1bGxSZXF1ZXN0Njk3MDkzNDEx
2,717
Fix shuffle on IterableDataset that disables batching in case any functions were mapped
{ "login": "amankhandelia", "id": 7098967, "node_id": "MDQ6VXNlcjcwOTg5Njc=", "avatar_url": "https://avatars.githubusercontent.com/u/7098967?v=4", "gravatar_id": "", "url": "https://api.github.com/users/amankhandelia", "html_url": "https://github.com/amankhandelia", "followers_url": "https://api.github....
[]
closed
false
null
[]
null
[]
1,627,310,542,000
1,627,322,654,000
1,627,317,006,000
CONTRIBUTOR
null
Made a very minor change to fix the issue#2716. Added the missing argument in the constructor call. As discussed in the bug report, the change is made to prevent the `shuffle` method call from resetting the value of `batched` attribute in `MappedExamplesIterable` Fix #2716.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2717/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2717/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2717", "html_url": "https://github.com/huggingface/datasets/pull/2717", "diff_url": "https://github.com/huggingface/datasets/pull/2717.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2717.patch", "merged_at": 1627317005000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2716
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2716/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2716/comments
https://api.github.com/repos/huggingface/datasets/issues/2716/events
https://github.com/huggingface/datasets/issues/2716
952,902,778
MDU6SXNzdWU5NTI5MDI3Nzg=
2,716
Calling shuffle on IterableDataset will disable batching in case any functions were mapped
{ "login": "amankhandelia", "id": 7098967, "node_id": "MDQ6VXNlcjcwOTg5Njc=", "avatar_url": "https://avatars.githubusercontent.com/u/7098967?v=4", "gravatar_id": "", "url": "https://api.github.com/users/amankhandelia", "html_url": "https://github.com/amankhandelia", "followers_url": "https://api.github....
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi :) Good catch ! Feel free to open a PR if you want to contribute, this would be very welcome ;)", "Have raised the PR [here](https://github.com/huggingface/datasets/pull/2717)", "Fixed by #2717." ]
1,627,305,899,000
1,627,322,683,000
1,627,322,683,000
CONTRIBUTOR
null
When using dataset in streaming mode, if one applies `shuffle` method on the dataset and `map` method for which `batched=True` than the batching operation will not happen, instead `batched` will be set to `False` I did RCA on the dataset codebase, the problem is emerging from [this line of code](https://github.com/h...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2716/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2716/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2715
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2715/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2715/comments
https://api.github.com/repos/huggingface/datasets/issues/2715/events
https://github.com/huggingface/datasets/pull/2715
952,845,229
MDExOlB1bGxSZXF1ZXN0Njk2OTc5MjQ1
2,715
Update PAN-X data URL in XTREME dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "Merging since the CI is just about missing infos in the dataset card" ]
1,627,302,077,000
1,627,306,079,000
1,627,306,079,000
MEMBER
null
Related to #2710, #2691.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2715/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2715/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2715", "html_url": "https://github.com/huggingface/datasets/pull/2715", "diff_url": "https://github.com/huggingface/datasets/pull/2715.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2715.patch", "merged_at": 1627306079000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2714
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2714/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2714/comments
https://api.github.com/repos/huggingface/datasets/issues/2714/events
https://github.com/huggingface/datasets/issues/2714
952,580,820
MDU6SXNzdWU5NTI1ODA4MjA=
2,714
add more precise information for size
{ "login": "pennyl67", "id": 1493902, "node_id": "MDQ6VXNlcjE0OTM5MDI=", "avatar_url": "https://avatars.githubusercontent.com/u/1493902?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pennyl67", "html_url": "https://github.com/pennyl67", "followers_url": "https://api.github.com/users/penny...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "We already have this information in the dataset_infos.json files of each dataset.\r\nMaybe we can parse these files in the backend to return their content with the endpoint at huggingface.co/api/datasets\r\n\r\nFor now if you want to access this info you have to load the json for each dataset. For example:\r\n- fo...
1,627,283,463,000
1,627,290,985,000
null
NONE
null
For the import into ELG, we would like a more precise description of the size of the dataset, instead of the current size categories. The size can be expressed in bytes, or any other preferred size unit. As suggested in the slack channel, perhaps this could be computed with a regex for existing datasets.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2714/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2714/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2713
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2713/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2713/comments
https://api.github.com/repos/huggingface/datasets/issues/2713/events
https://github.com/huggingface/datasets/pull/2713
952,515,256
MDExOlB1bGxSZXF1ZXN0Njk2Njk3MzU0
2,713
Enumerate all ner_tags values in WNUT 17 dataset
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,627,276,936,000
1,627,291,855,000
1,627,291,855,000
MEMBER
null
This PR does: - Enumerate all ner_tags in dataset card Data Fields section - Add all metadata tags to dataset card Close #2709.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2713/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2713/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2713", "html_url": "https://github.com/huggingface/datasets/pull/2713", "diff_url": "https://github.com/huggingface/datasets/pull/2713.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2713.patch", "merged_at": 1627291854000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2710
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2710/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2710/comments
https://api.github.com/repos/huggingface/datasets/issues/2710/events
https://github.com/huggingface/datasets/pull/2710
951,723,326
MDExOlB1bGxSZXF1ZXN0Njk2MDYyNjAy
2,710
Update WikiANN data URL
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "We have to update the URL in the XTREME benchmark as well:\r\n\r\nhttps://github.com/huggingface/datasets/blob/0dfc639cec450ed8762a997789a2ed63e63cdcf2/datasets/xtreme/xtreme.py#L411-L411\r\n\r\n" ]
1,627,057,761,000
1,627,292,063,000
1,627,292,063,000
MEMBER
null
WikiANN data source URL is no longer accessible: 404 error from Dropbox. We have decided to host it at Hugging Face. This PR updates the data source URL, the metadata JSON file and the dataset card. Close #2691.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2710/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2710/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2710", "html_url": "https://github.com/huggingface/datasets/pull/2710", "diff_url": "https://github.com/huggingface/datasets/pull/2710.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2710.patch", "merged_at": 1627292062000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2709
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2709/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2709/comments
https://api.github.com/repos/huggingface/datasets/issues/2709/events
https://github.com/huggingface/datasets/issues/2709
951,534,757
MDU6SXNzdWU5NTE1MzQ3NTc=
2,709
Missing documentation for wnut_17 (ner_tags)
{ "login": "maxpel", "id": 31095360, "node_id": "MDQ6VXNlcjMxMDk1MzYw", "avatar_url": "https://avatars.githubusercontent.com/u/31095360?v=4", "gravatar_id": "", "url": "https://api.github.com/users/maxpel", "html_url": "https://github.com/maxpel", "followers_url": "https://api.github.com/users/maxpel/fo...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Hi @maxpel, thanks for reporting this issue.\r\n\r\nIndeed, the documentation in the dataset card is not complete. I’m opening a Pull Request to fix it.\r\n\r\nAs the paper explains, there are 6 entity types and we have ordered them alphabetically: `corporation`, `creative-work`, `group`, `location`, `person` and ...
1,627,043,132,000
1,627,291,855,000
1,627,291,855,000
CONTRIBUTOR
null
On the info page of the wnut_17 data set (https://huggingface.co/datasets/wnut_17), the model output of ner-tags is only documented for these 5 cases: `ner_tags: a list of classification labels, with possible values including O (0), B-corporation (1), I-corporation (2), B-creative-work (3), I-creative-work (4).` ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2709/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2709/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2708
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2708/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2708/comments
https://api.github.com/repos/huggingface/datasets/issues/2708/events
https://github.com/huggingface/datasets/issues/2708
951,092,660
MDU6SXNzdWU5NTEwOTI2NjA=
2,708
QASC: incomplete training set
{ "login": "danyaljj", "id": 2441454, "node_id": "MDQ6VXNlcjI0NDE0NTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2441454?v=4", "gravatar_id": "", "url": "https://api.github.com/users/danyaljj", "html_url": "https://github.com/danyaljj", "followers_url": "https://api.github.com/users/danya...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi @danyaljj, thanks for reporting.\r\n\r\nUnfortunately, I have not been able to reproduce your problem. My train split has 8134 examples:\r\n```ipython\r\nIn [10]: ds[\"train\"]\r\nOut[10]:\r\nDataset({\r\n features: ['id', 'question', 'choices', 'answerKey', 'fact1', 'fact2', 'combinedfact', 'formatted_quest...
1,626,991,184,000
1,627,047,007,000
1,627,047,007,000
CONTRIBUTOR
null
## Describe the bug The training instances are not loaded properly. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("qasc", script_version='1.10.2') def load_instances(split): instances = dataset[split] print(f"split: {split} - size: {len(instanc...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2708/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2708/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2707
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2707/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2707/comments
https://api.github.com/repos/huggingface/datasets/issues/2707/events
https://github.com/huggingface/datasets/issues/2707
950,812,945
MDU6SXNzdWU5NTA4MTI5NDU=
2,707
404 Not Found Error when loading LAMA dataset
{ "login": "dwil2444", "id": 26467159, "node_id": "MDQ6VXNlcjI2NDY3MTU5", "avatar_url": "https://avatars.githubusercontent.com/u/26467159?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dwil2444", "html_url": "https://github.com/dwil2444", "followers_url": "https://api.github.com/users/dwi...
[]
closed
false
null
[]
null
[ "Hi @dwil2444! I was able to reproduce your error when I downgraded to v1.1.2. Updating to the latest version of Datasets fixed the error for me :)", "Hi @dwil2444, thanks for reporting.\r\n\r\nCould you please confirm which `datasets` version you were using and if the problem persists after you update it to the ...
1,626,969,153,000
1,627,309,747,000
1,627,309,747,000
NONE
null
The [LAMA](https://huggingface.co/datasets/viewer/?dataset=lama) probing dataset is not available for download: Steps to Reproduce: 1. `from datasets import load_dataset` 2. `dataset = load_dataset('lama', 'trex')`. Results: `FileNotFoundError: Couldn't find file locally at lama/lama.py, or remotely ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2707/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2707/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2706
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2706/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2706/comments
https://api.github.com/repos/huggingface/datasets/issues/2706/events
https://github.com/huggingface/datasets/pull/2706
950,606,561
MDExOlB1bGxSZXF1ZXN0Njk1MTI3ODgz
2,706
Update BibTeX entry
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,626,956,969,000
1,626,957,780,000
1,626,957,780,000
MEMBER
null
Update BibTeX entry.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2706/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2706/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2706", "html_url": "https://github.com/huggingface/datasets/pull/2706", "diff_url": "https://github.com/huggingface/datasets/pull/2706.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2706.patch", "merged_at": 1626957780000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2705
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2705/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2705/comments
https://api.github.com/repos/huggingface/datasets/issues/2705/events
https://github.com/huggingface/datasets/issues/2705
950,488,583
MDU6SXNzdWU5NTA0ODg1ODM=
2,705
404 not found error on loading WIKIANN dataset
{ "login": "ronbutan", "id": 39296659, "node_id": "MDQ6VXNlcjM5Mjk2NjU5", "avatar_url": "https://avatars.githubusercontent.com/u/39296659?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ronbutan", "html_url": "https://github.com/ronbutan", "followers_url": "https://api.github.com/users/ron...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi @ronbutan, thanks for reporting.\r\n\r\nYou are right: we have recently found that the link to the original PAN-X dataset (also called WikiANN), hosted at Dropbox, is no longer working.\r\n\r\nWe have opened an issue in the GitHub repository of the original dataset (afshinrahimi/mmner#4) and we have also contac...
1,626,947,750,000
1,627,027,652,000
1,627,027,652,000
NONE
null
## Describe the bug Unable to retreive wikiann English dataset ## Steps to reproduce the bug ```python from datasets import list_datasets, load_dataset, list_metrics, load_metric WIKIANN = load_dataset("wikiann","en") ``` ## Expected results Colab notebook should display successful download status ## Act...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2705/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2705/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2704
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2704/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2704/comments
https://api.github.com/repos/huggingface/datasets/issues/2704/events
https://github.com/huggingface/datasets/pull/2704
950,483,980
MDExOlB1bGxSZXF1ZXN0Njk1MDIzMTEz
2,704
Fix pick default config name message
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,626,947,383,000
1,626,948,161,000
1,626,948,160,000
MEMBER
null
The error message to tell which config name to load is not displayed. This is because in the code it was considering the config kwargs to be non-empty, which is a special case for custom configs created on the fly. It appears after this change: https://github.com/huggingface/datasets/pull/2659 I fixed that by ma...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2704/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2704/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2704", "html_url": "https://github.com/huggingface/datasets/pull/2704", "diff_url": "https://github.com/huggingface/datasets/pull/2704.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2704.patch", "merged_at": 1626948160000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2703
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2703/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2703/comments
https://api.github.com/repos/huggingface/datasets/issues/2703/events
https://github.com/huggingface/datasets/issues/2703
950,482,284
MDU6SXNzdWU5NTA0ODIyODQ=
2,703
Bad message when config name is missing
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[]
1,626,947,243,000
1,626,948,160,000
1,626,948,160,000
MEMBER
null
When loading a dataset that have several configurations, we expect to see an error message if the user doesn't specify a config name. However in `datasets` 1.10.0 and 1.10.1 it doesn't show the right message: ```python import datasets datasets.load_dataset("glue") ``` raises ```python AttributeError: 'Bui...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2703/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2703/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/2702
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2702/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2702/comments
https://api.github.com/repos/huggingface/datasets/issues/2702/events
https://github.com/huggingface/datasets/pull/2702
950,448,159
MDExOlB1bGxSZXF1ZXN0Njk0OTkyOTc1
2,702
Update BibTeX entry
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[]
1,626,944,679,000
1,626,945,459,000
1,626,945,458,000
MEMBER
null
Update BibTeX entry.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2702/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2702/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2702", "html_url": "https://github.com/huggingface/datasets/pull/2702", "diff_url": "https://github.com/huggingface/datasets/pull/2702.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2702.patch", "merged_at": 1626945458000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2701
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2701/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2701/comments
https://api.github.com/repos/huggingface/datasets/issues/2701/events
https://github.com/huggingface/datasets/pull/2701
950,422,403
MDExOlB1bGxSZXF1ZXN0Njk0OTcxMzM3
2,701
Fix download_mode docstrings
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
closed
false
null
[]
null
[]
1,626,942,625,000
1,626,946,411,000
1,626,946,411,000
MEMBER
null
Fix `download_mode` docstrings.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2701/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2701/timeline
null
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2701", "html_url": "https://github.com/huggingface/datasets/pull/2701", "diff_url": "https://github.com/huggingface/datasets/pull/2701.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2701.patch", "merged_at": 1626946411000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2700
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2700/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2700/comments
https://api.github.com/repos/huggingface/datasets/issues/2700/events
https://github.com/huggingface/datasets/issues/2700
950,276,325
MDU6SXNzdWU5NTAyNzYzMjU=
2,700
from datasets import Dataset is failing
{ "login": "kswamy15", "id": 5582286, "node_id": "MDQ6VXNlcjU1ODIyODY=", "avatar_url": "https://avatars.githubusercontent.com/u/5582286?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kswamy15", "html_url": "https://github.com/kswamy15", "followers_url": "https://api.github.com/users/kswam...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Hi @kswamy15, thanks for reporting.\r\n\r\nWe are fixing this critical issue and making an urgent patch release of the `datasets` library today.\r\n\r\nIn the meantime, you can circumvent this issue by updating the `tqdm` library: `!pip install -U tqdm`" ]
1,626,925,883,000
1,626,938,625,000
1,626,937,747,000
NONE
null
## Describe the bug A clear and concise description of what the bug is. ## Steps to reproduce the bug ```python # Sample code to reproduce the bug from datasets import Dataset ``` ## Expected results A clear and concise description of the expected results. ## Actual results Specify the actual results or...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2700/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2700/timeline
null
completed
null
null
false