url stringlengths 58 61 | repository_url stringclasses 1
value | labels_url stringlengths 72 75 | comments_url stringlengths 67 70 | events_url stringlengths 65 68 | html_url stringlengths 46 51 | id int64 599M 1.07B | node_id stringlengths 18 32 | number int64 1 3.39k | title stringlengths 1 276 | user dict | labels list | state stringclasses 1
value | locked bool 1
class | assignee dict | assignees list | milestone dict | comments list | created_at int64 1,587B 1,639B | updated_at int64 1,587B 1,639B | closed_at int64 1,587B 1,639B | author_association stringclasses 3
values | active_lock_reason null | body stringlengths 0 228k ⌀ | reactions dict | timeline_url stringlengths 67 70 | performed_via_github_app null | draft bool 2
classes | pull_request dict | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/2916 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2916/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2916/comments | https://api.github.com/repos/huggingface/datasets/issues/2916/events | https://github.com/huggingface/datasets/pull/2916 | 997,003,661 | PR_kwDODunzps4rx5ua | 2,916 | Add OpenAI's pass@k code evaluation metric | {
"login": "lvwerra",
"id": 8264887,
"node_id": "MDQ6VXNlcjgyNjQ4ODc=",
"avatar_url": "https://avatars.githubusercontent.com/u/8264887?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lvwerra",
"html_url": "https://github.com/lvwerra",
"followers_url": "https://api.github.com/users/lvwerra/... | [] | closed | false | null | [] | null | [
"> The implementation makes heavy use of multiprocessing which this PR does not touch. Is this conflicting with multiprocessing natively integrated in datasets?\r\n\r\nIt should work normally, but feel free to test it.\r\nThere is some documentation about using metrics in a distributed setup that uses multiprocessi... | 1,631,707,543,000 | 1,636,726,791,000 | 1,636,726,790,000 | CONTRIBUTOR | null | This PR introduces the `code_eval` metric which implements [OpenAI's code evaluation harness](https://github.com/openai/human-eval) introduced in the [Codex paper](https://arxiv.org/abs/2107.03374). It is heavily based on the original implementation and just adapts the interface to follow the `predictions`/`references`... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2916/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 1,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2916/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2916",
"html_url": "https://github.com/huggingface/datasets/pull/2916",
"diff_url": "https://github.com/huggingface/datasets/pull/2916.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2916.patch",
"merged_at": 1636726790000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2915 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2915/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2915/comments | https://api.github.com/repos/huggingface/datasets/issues/2915/events | https://github.com/huggingface/datasets/pull/2915 | 996,870,071 | PR_kwDODunzps4rxfWb | 2,915 | Fix fsspec AbstractFileSystem access | {
"login": "pierre-godard",
"id": 3969168,
"node_id": "MDQ6VXNlcjM5NjkxNjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/3969168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pierre-godard",
"html_url": "https://github.com/pierre-godard",
"followers_url": "https://api.github.... | [] | closed | false | null | [] | null | [] | 1,631,698,760,000 | 1,631,705,724,000 | 1,631,705,724,000 | CONTRIBUTOR | null | This addresses the issue from #2914 by changing the way fsspec's AbstractFileSystem is accessed. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2915/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2915/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2915",
"html_url": "https://github.com/huggingface/datasets/pull/2915",
"diff_url": "https://github.com/huggingface/datasets/pull/2915.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2915.patch",
"merged_at": 1631705724000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2914 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2914/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2914/comments | https://api.github.com/repos/huggingface/datasets/issues/2914/events | https://github.com/huggingface/datasets/issues/2914 | 996,770,168 | I_kwDODunzps47aYF4 | 2,914 | Having a dependency defining fsspec entrypoint raises an AttributeError when importing datasets | {
"login": "pierre-godard",
"id": 3969168,
"node_id": "MDQ6VXNlcjM5NjkxNjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/3969168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pierre-godard",
"html_url": "https://github.com/pierre-godard",
"followers_url": "https://api.github.... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Closed by #2915."
] | 1,631,692,446,000 | 1,631,724,557,000 | 1,631,724,556,000 | CONTRIBUTOR | null | ## Describe the bug
In one of my project, I defined a custom fsspec filesystem with an entrypoint.
My guess is that by doing so, a variable named `spec` is created in the module `fsspec` (created by entering a for loop as there are entrypoints defined, see the loop in question [here](https://github.com/intake/filesys... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2914/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2914/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2913 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2913/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2913/comments | https://api.github.com/repos/huggingface/datasets/issues/2913/events | https://github.com/huggingface/datasets/issues/2913 | 996,436,368 | I_kwDODunzps47ZGmQ | 2,913 | timit_asr dataset only includes one text phrase | {
"login": "margotwagner",
"id": 39107794,
"node_id": "MDQ6VXNlcjM5MTA3Nzk0",
"avatar_url": "https://avatars.githubusercontent.com/u/39107794?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/margotwagner",
"html_url": "https://github.com/margotwagner",
"followers_url": "https://api.github.c... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Hi @margotwagner, \r\nThis bug was fixed in #1995. Upgrading the datasets should work (min v1.8.0 ideally)",
"Hi @margotwagner,\r\n\r\nYes, as @bhavitvyamalik has commented, this bug was fixed in `datasets` version 1.5.0. You need to update it, as your current version is 1.4.1:\r\n> Environment info\r\n> - `data... | 1,631,653,567,000 | 1,631,693,119,000 | 1,631,693,118,000 | NONE | null | ## Describe the bug
The dataset 'timit_asr' only includes one text phrase. It only includes the transcription "Would such an act of refusal be useful?" multiple times rather than different phrases.
## Steps to reproduce the bug
Note: I am following the tutorial https://huggingface.co/blog/fine-tune-wav2vec2-englis... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2913/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2913/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2912 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2912/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2912/comments | https://api.github.com/repos/huggingface/datasets/issues/2912/events | https://github.com/huggingface/datasets/pull/2912 | 996,256,005 | PR_kwDODunzps4rvhgp | 2,912 | Update link to Blog in docs footer | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,631,640,194,000 | 1,631,692,763,000 | 1,631,692,763,000 | MEMBER | null | Update link. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2912/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2912/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2912",
"html_url": "https://github.com/huggingface/datasets/pull/2912",
"diff_url": "https://github.com/huggingface/datasets/pull/2912.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2912.patch",
"merged_at": 1631692763000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2911 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2911/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2911/comments | https://api.github.com/repos/huggingface/datasets/issues/2911/events | https://github.com/huggingface/datasets/pull/2911 | 996,202,598 | PR_kwDODunzps4rvW7Y | 2,911 | Fix exception chaining | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,631,636,369,000 | 1,631,804,684,000 | 1,631,804,684,000 | MEMBER | null | Fix exception chaining to avoid tracebacks with message: `During handling of the above exception, another exception occurred:` | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2911/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2911/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2911",
"html_url": "https://github.com/huggingface/datasets/pull/2911",
"diff_url": "https://github.com/huggingface/datasets/pull/2911.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2911.patch",
"merged_at": 1631804684000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2910 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2910/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2910/comments | https://api.github.com/repos/huggingface/datasets/issues/2910/events | https://github.com/huggingface/datasets/pull/2910 | 996,149,632 | PR_kwDODunzps4rvL9N | 2,910 | feat: 🎸 pass additional arguments to get private configs + info | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [] | closed | false | null | [] | null | [
"Included in https://github.com/huggingface/datasets/pull/2906"
] | 1,631,633,059,000 | 1,631,722,749,000 | 1,631,722,746,000 | CONTRIBUTOR | null | `use_auth_token` can now be passed to the functions to get the configs
or infos of private datasets on the hub | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2910/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2910/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2910",
"html_url": "https://github.com/huggingface/datasets/pull/2910",
"diff_url": "https://github.com/huggingface/datasets/pull/2910.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2910.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2909 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2909/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2909/comments | https://api.github.com/repos/huggingface/datasets/issues/2909/events | https://github.com/huggingface/datasets/pull/2909 | 996,002,180 | PR_kwDODunzps4rutdo | 2,909 | fix anli splits | {
"login": "zaidalyafeai",
"id": 15667714,
"node_id": "MDQ6VXNlcjE1NjY3NzE0",
"avatar_url": "https://avatars.githubusercontent.com/u/15667714?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zaidalyafeai",
"html_url": "https://github.com/zaidalyafeai",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | [] | 1,631,625,035,000 | 1,634,124,469,000 | 1,634,124,469,000 | CONTRIBUTOR | null | I can't run the tests for dummy data, facing this error
`ImportError while loading conftest '/home/zaid/tmp/fix_anli_splits/datasets/tests/conftest.py'.
tests/conftest.py:10: in <module>
from datasets import config
E ImportError: cannot import name 'config' from 'datasets' (unknown location)` | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2909/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2909/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2909",
"html_url": "https://github.com/huggingface/datasets/pull/2909",
"diff_url": "https://github.com/huggingface/datasets/pull/2909.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2909.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2908 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2908/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2908/comments | https://api.github.com/repos/huggingface/datasets/issues/2908/events | https://github.com/huggingface/datasets/pull/2908 | 995,970,612 | PR_kwDODunzps4rumwW | 2,908 | Update Zenodo metadata with creator names and affiliation | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,631,623,177,000 | 1,631,629,765,000 | 1,631,629,765,000 | MEMBER | null | This PR helps in prefilling author data when automatically generating the DOI after each release. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2908/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2908/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2908",
"html_url": "https://github.com/huggingface/datasets/pull/2908",
"diff_url": "https://github.com/huggingface/datasets/pull/2908.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2908.patch",
"merged_at": 1631629765000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2907 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2907/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2907/comments | https://api.github.com/repos/huggingface/datasets/issues/2907/events | https://github.com/huggingface/datasets/pull/2907 | 995,968,152 | PR_kwDODunzps4rumOy | 2,907 | add story_cloze dataset | {
"login": "zaidalyafeai",
"id": 15667714,
"node_id": "MDQ6VXNlcjE1NjY3NzE0",
"avatar_url": "https://avatars.githubusercontent.com/u/15667714?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zaidalyafeai",
"html_url": "https://github.com/zaidalyafeai",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | [
"Will create a new one, this one seems to be missed up. "
] | 1,631,623,013,000 | 1,633,729,302,000 | 1,633,729,301,000 | CONTRIBUTOR | null | @lhoestq I have spent some time but I still I can't succeed in correctly testing the dummy_data. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2907/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2907/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2907",
"html_url": "https://github.com/huggingface/datasets/pull/2907",
"diff_url": "https://github.com/huggingface/datasets/pull/2907.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2907.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2906 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2906/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2906/comments | https://api.github.com/repos/huggingface/datasets/issues/2906/events | https://github.com/huggingface/datasets/pull/2906 | 995,962,905 | PR_kwDODunzps4rulH- | 2,906 | feat: 🎸 add a function to get a dataset config's split names | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [] | closed | false | null | [] | null | [
"> Should I add a section in https://github.com/huggingface/datasets/blob/master/docs/source/load_hub.rst? (there is no section for get_dataset_infos)\r\n\r\nYes totally :) This tutorial should indeed mention this, given how fundamental it is"
] | 1,631,622,682,000 | 1,633,341,338,000 | 1,633,341,337,000 | CONTRIBUTOR | null | Also: pass additional arguments (use_auth_token) to get private configs + info of private datasets on the hub
Questions:
- [x] I'm not sure how the versions work: I changed 1.12.1.dev0 to 1.12.1.dev1, was it correct?
-> no: reverted
- [x] Should I add a section in https://github.com/huggingface/datasets/blo... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2906/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2906/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2906",
"html_url": "https://github.com/huggingface/datasets/pull/2906",
"diff_url": "https://github.com/huggingface/datasets/pull/2906.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2906.patch",
"merged_at": 1633341337000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2905 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2905/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2905/comments | https://api.github.com/repos/huggingface/datasets/issues/2905/events | https://github.com/huggingface/datasets/pull/2905 | 995,843,964 | PR_kwDODunzps4ruL5X | 2,905 | Update BibTeX entry | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,631,614,577,000 | 1,631,622,337,000 | 1,631,622,337,000 | MEMBER | null | Update BibTeX entry. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2905/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2905/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2905",
"html_url": "https://github.com/huggingface/datasets/pull/2905",
"diff_url": "https://github.com/huggingface/datasets/pull/2905.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2905.patch",
"merged_at": 1631622337000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2903 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2903/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2903/comments | https://api.github.com/repos/huggingface/datasets/issues/2903/events | https://github.com/huggingface/datasets/pull/2903 | 995,715,191 | PR_kwDODunzps4rtxxV | 2,903 | Fix xpathopen to accept positional arguments | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"thanks!"
] | 1,631,606,570,000 | 1,631,609,481,000 | 1,631,608,847,000 | MEMBER | null | Fix `xpathopen()` so that it also accepts positional arguments.
Fix #2901. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2903/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2903/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2903",
"html_url": "https://github.com/huggingface/datasets/pull/2903",
"diff_url": "https://github.com/huggingface/datasets/pull/2903.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2903.patch",
"merged_at": 1631608847000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2901 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2901/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2901/comments | https://api.github.com/repos/huggingface/datasets/issues/2901/events | https://github.com/huggingface/datasets/issues/2901 | 995,232,844 | MDU6SXNzdWU5OTUyMzI4NDQ= | 2,901 | Incompatibility with pytest | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Sorry, my bad... When implementing `xpathopen`, I just considered the use case in the COUNTER dataset... I'm fixing it!"
] | 1,631,560,337,000 | 1,631,608,847,000 | 1,631,608,847,000 | CONTRIBUTOR | null | ## Describe the bug
pytest complains about xpathopen / path.open("w")
## Steps to reproduce the bug
Create a test file, `test.py`:
```python
import datasets as ds
def load_dataset():
ds.load_dataset("counter", split="train", streaming=True)
```
And launch it with pytest:
```bash
python -m pyt... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2901/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2901/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2900 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2900/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2900/comments | https://api.github.com/repos/huggingface/datasets/issues/2900/events | https://github.com/huggingface/datasets/pull/2900 | 994,922,580 | MDExOlB1bGxSZXF1ZXN0NzMyNzczNDkw | 2,900 | Fix null sequence encoding | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [] | 1,631,541,308,000 | 1,631,542,663,000 | 1,631,542,662,000 | MEMBER | null | The Sequence feature encoding was failing when a `None` sequence was used in a dataset.
Fix https://github.com/huggingface/datasets/issues/2892 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2900/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2900/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2900",
"html_url": "https://github.com/huggingface/datasets/pull/2900",
"diff_url": "https://github.com/huggingface/datasets/pull/2900.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2900.patch",
"merged_at": 1631542662000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2899 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2899/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2899/comments | https://api.github.com/repos/huggingface/datasets/issues/2899/events | https://github.com/huggingface/datasets/issues/2899 | 994,082,432 | MDU6SXNzdWU5OTQwODI0MzI= | 2,899 | Dataset | {
"login": "rcacho172",
"id": 90449239,
"node_id": "MDQ6VXNlcjkwNDQ5MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/90449239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rcacho172",
"html_url": "https://github.com/rcacho172",
"followers_url": "https://api.github.com/users/... | [
{
"id": 2067376369,
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request",
"name": "dataset request",
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset"
}
] | closed | false | null | [] | null | [] | 1,631,432,333,000 | 1,631,463,135,000 | 1,631,463,135,000 | NONE | null | ## Adding a Dataset
- **Name:** *name of the dataset*
- **Description:** *short description of the dataset (or link to social media or blog post)*
- **Paper:** *link to the dataset paper if available*
- **Data:** *link to the Github repository or current dataset location*
- **Motivation:** *what are some good reasons t... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2899/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2899/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2898 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2898/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2898/comments | https://api.github.com/repos/huggingface/datasets/issues/2898/events | https://github.com/huggingface/datasets/issues/2898 | 994,032,814 | MDU6SXNzdWU5OTQwMzI4MTQ= | 2,898 | Hug emoji | {
"login": "Jackg-08",
"id": 90539794,
"node_id": "MDQ6VXNlcjkwNTM5Nzk0",
"avatar_url": "https://avatars.githubusercontent.com/u/90539794?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Jackg-08",
"html_url": "https://github.com/Jackg-08",
"followers_url": "https://api.github.com/users/Jac... | [
{
"id": 2067376369,
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request",
"name": "dataset request",
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset"
}
] | closed | false | null | [] | null | [] | 1,631,417,271,000 | 1,631,463,193,000 | 1,631,463,193,000 | NONE | null | ## Adding a Dataset
- **Name:** *name of the dataset*
- **Description:** *short description of the dataset (or link to social media or blog post)*
- **Paper:** *link to the dataset paper if available*
- **Data:** *link to the Github repository or current dataset location*
- **Motivation:** *what are some good reasons t... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2898/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2898/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2897 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2897/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2897/comments | https://api.github.com/repos/huggingface/datasets/issues/2897/events | https://github.com/huggingface/datasets/pull/2897 | 993,798,386 | MDExOlB1bGxSZXF1ZXN0NzMxOTA0ODk4 | 2,897 | Add OpenAI's HumanEval dataset | {
"login": "lvwerra",
"id": 8264887,
"node_id": "MDQ6VXNlcjgyNjQ4ODc=",
"avatar_url": "https://avatars.githubusercontent.com/u/8264887?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lvwerra",
"html_url": "https://github.com/lvwerra",
"followers_url": "https://api.github.com/users/lvwerra/... | [] | closed | false | null | [] | null | [
"I just fixed the class name, and added `[More Information Needed]` in empty sections in case people want to complete the dataset card :)"
] | 1,631,353,067,000 | 1,631,804,531,000 | 1,631,804,531,000 | CONTRIBUTOR | null | This PR adds OpenAI's [HumanEval](https://github.com/openai/human-eval) dataset. The dataset consists of 164 handcrafted programming problems with solutions and unittests to verify solution. This dataset is useful to evaluate code generation models. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2897/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2897/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2897",
"html_url": "https://github.com/huggingface/datasets/pull/2897",
"diff_url": "https://github.com/huggingface/datasets/pull/2897.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2897.patch",
"merged_at": 1631804531000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2896 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2896/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2896/comments | https://api.github.com/repos/huggingface/datasets/issues/2896/events | https://github.com/huggingface/datasets/pull/2896 | 993,613,113 | MDExOlB1bGxSZXF1ZXN0NzMxNzcwMTE3 | 2,896 | add multi-proc in `to_csv` | {
"login": "bhavitvyamalik",
"id": 19718818,
"node_id": "MDQ6VXNlcjE5NzE4ODE4",
"avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bhavitvyamalik",
"html_url": "https://github.com/bhavitvyamalik",
"followers_url": "https://api.gi... | [] | closed | false | null | [] | null | [
"I think you can just add a test `test_dataset_to_csv_multiproc` in `tests/io/test_csv.py` and we'll be good",
"Hi @lhoestq, \r\nI've added `test_dataset_to_csv` apart from `test_dataset_to_csv_multiproc` as no test was there to check generated CSV file when `num_proc=1`. Please let me know if anything is also re... | 1,631,309,709,000 | 1,635,400,053,000 | 1,635,264,042,000 | CONTRIBUTOR | null | This PR extends the multi-proc method used in #2747 for`to_json` to `to_csv` as well.
Results on my machine post benchmarking on `ascent_kb` dataset (giving ~45% improvement when compared to num_proc = 1):
```
Time taken on 1 num_proc, 10000 batch_size 674.2055702209473
Time taken on 4 num_proc, 10000 batch_siz... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2896/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2896/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2896",
"html_url": "https://github.com/huggingface/datasets/pull/2896",
"diff_url": "https://github.com/huggingface/datasets/pull/2896.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2896.patch",
"merged_at": 1635264041000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2895 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2895/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2895/comments | https://api.github.com/repos/huggingface/datasets/issues/2895/events | https://github.com/huggingface/datasets/pull/2895 | 993,462,274 | MDExOlB1bGxSZXF1ZXN0NzMxNjQ0NTY2 | 2,895 | Use pyarrow.Table.replace_schema_metadata instead of pyarrow.Table.cast | {
"login": "arsarabi",
"id": 12345848,
"node_id": "MDQ6VXNlcjEyMzQ1ODQ4",
"avatar_url": "https://avatars.githubusercontent.com/u/12345848?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/arsarabi",
"html_url": "https://github.com/arsarabi",
"followers_url": "https://api.github.com/users/ars... | [] | closed | false | null | [] | null | [] | 1,631,296,617,000 | 1,632,264,601,000 | 1,632,212,315,000 | CONTRIBUTOR | null | This PR partially addresses #2252.
``update_metadata_with_features`` uses ``Table.cast`` which slows down ``load_from_disk`` (and possibly other methods that use it) for very large datasets. Since ``update_metadata_with_features`` is only updating the schema metadata, it makes more sense to use ``pyarrow.Table.repla... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2895/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2895/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2895",
"html_url": "https://github.com/huggingface/datasets/pull/2895",
"diff_url": "https://github.com/huggingface/datasets/pull/2895.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2895.patch",
"merged_at": 1632212315000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2894 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2894/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2894/comments | https://api.github.com/repos/huggingface/datasets/issues/2894/events | https://github.com/huggingface/datasets/pull/2894 | 993,375,654 | MDExOlB1bGxSZXF1ZXN0NzMxNTcxODc5 | 2,894 | Fix COUNTER dataset | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,631,290,049,000 | 1,631,291,265,000 | 1,631,291,264,000 | MEMBER | null | Fix filename generating `FileNotFoundError`.
Related to #2866.
CC: @severo. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2894/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2894/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2894",
"html_url": "https://github.com/huggingface/datasets/pull/2894",
"diff_url": "https://github.com/huggingface/datasets/pull/2894.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2894.patch",
"merged_at": 1631291264000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2893 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2893/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2893/comments | https://api.github.com/repos/huggingface/datasets/issues/2893/events | https://github.com/huggingface/datasets/pull/2893 | 993,342,781 | MDExOlB1bGxSZXF1ZXN0NzMxNTQ0NDQz | 2,893 | add mbpp dataset | {
"login": "lvwerra",
"id": 8264887,
"node_id": "MDQ6VXNlcjgyNjQ4ODc=",
"avatar_url": "https://avatars.githubusercontent.com/u/8264887?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lvwerra",
"html_url": "https://github.com/lvwerra",
"followers_url": "https://api.github.com/users/lvwerra/... | [] | closed | false | null | [] | null | [
"I think it's fine to have the original schema"
] | 1,631,287,650,000 | 1,631,784,942,000 | 1,631,784,942,000 | CONTRIBUTOR | null | This PR adds the mbpp dataset introduced by Google [here](https://github.com/google-research/google-research/tree/master/mbpp) as mentioned in #2816.
The dataset contain two versions: a full and a sanitized one. They have a slightly different schema and it is current state the loading preserves the original schema. ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2893/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2893/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2893",
"html_url": "https://github.com/huggingface/datasets/pull/2893",
"diff_url": "https://github.com/huggingface/datasets/pull/2893.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2893.patch",
"merged_at": 1631784942000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2892 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2892/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2892/comments | https://api.github.com/repos/huggingface/datasets/issues/2892/events | https://github.com/huggingface/datasets/issues/2892 | 993,274,572 | MDU6SXNzdWU5OTMyNzQ1NzI= | 2,892 | Error when encoding a dataset with None objects with a Sequence feature | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"This has been fixed by https://github.com/huggingface/datasets/pull/2900\r\nWe're doing a new release 1.12 today to make the fix available :)"
] | 1,631,283,103,000 | 1,631,542,693,000 | 1,631,542,662,000 | MEMBER | null | There is an error when encoding a dataset with None objects with a Sequence feature
To reproduce:
```python
from datasets import Dataset, Features, Value, Sequence
data = {"a": [[0], None]}
features = Features({"a": Sequence(Value("int32"))})
dataset = Dataset.from_dict(data, features=features)
```
raises
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2892/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2892/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2891 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2891/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2891/comments | https://api.github.com/repos/huggingface/datasets/issues/2891/events | https://github.com/huggingface/datasets/pull/2891 | 993,161,984 | MDExOlB1bGxSZXF1ZXN0NzMxMzkwNjM2 | 2,891 | Allow dynamic first dimension for ArrayXD | {
"login": "rpowalski",
"id": 10357417,
"node_id": "MDQ6VXNlcjEwMzU3NDE3",
"avatar_url": "https://avatars.githubusercontent.com/u/10357417?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rpowalski",
"html_url": "https://github.com/rpowalski",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | [
"@lhoestq, thanks for your review.\r\n\r\nI added test for `to_pylist`, I didn't do that for `to_numpy` because this method shouldn't be called for dynamic dimension ArrayXD - this method will try to make a single numpy array for the whole column which cannot be done for dynamic arrays.\r\n\r\nI dig into `to_pandas... | 1,631,274,772,000 | 1,637,681,593,000 | 1,635,500,237,000 | CONTRIBUTOR | null | Add support for dynamic first dimension for ArrayXD features. See issue [#887](https://github.com/huggingface/datasets/issues/887).
Following changes allow for `to_pylist` method of `ArrayExtensionArray` to return a list of numpy arrays where fist dimension can vary.
@lhoestq Could you suggest how you want to exten... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2891/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2891/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2891",
"html_url": "https://github.com/huggingface/datasets/pull/2891",
"diff_url": "https://github.com/huggingface/datasets/pull/2891.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2891.patch",
"merged_at": 1635500237000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2890 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2890/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2890/comments | https://api.github.com/repos/huggingface/datasets/issues/2890/events | https://github.com/huggingface/datasets/issues/2890 | 993,074,102 | MDU6SXNzdWU5OTMwNzQxMDI= | 2,890 | 0x290B112ED1280537B24Ee6C268a004994a16e6CE | {
"login": "rcacho172",
"id": 90449239,
"node_id": "MDQ6VXNlcjkwNDQ5MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/90449239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rcacho172",
"html_url": "https://github.com/rcacho172",
"followers_url": "https://api.github.com/users/... | [
{
"id": 2067376369,
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request",
"name": "dataset request",
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset"
}
] | closed | false | null | [] | null | [] | 1,631,267,477,000 | 1,631,274,329,000 | 1,631,274,329,000 | NONE | null | ## Adding a Dataset
- **Name:** *name of the dataset*
- **Description:** *short description of the dataset (or link to social media or blog post)*
- **Paper:** *link to the dataset paper if available*
- **Data:** *link to the Github repository or current dataset location*
- **Motivation:** *what are some good reasons t... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2890/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2890/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2889 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2889/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2889/comments | https://api.github.com/repos/huggingface/datasets/issues/2889/events | https://github.com/huggingface/datasets/issues/2889 | 992,968,382 | MDU6SXNzdWU5OTI5NjgzODI= | 2,889 | Coc | {
"login": "Bwiggity",
"id": 90444264,
"node_id": "MDQ6VXNlcjkwNDQ0MjY0",
"avatar_url": "https://avatars.githubusercontent.com/u/90444264?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Bwiggity",
"html_url": "https://github.com/Bwiggity",
"followers_url": "https://api.github.com/users/Bwi... | [
{
"id": 2067376369,
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request",
"name": "dataset request",
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset"
}
] | closed | false | null | [] | null | [] | 1,631,259,127,000 | 1,631,274,354,000 | 1,631,274,354,000 | NONE | null | ## Adding a Dataset
- **Name:** *name of the dataset*
- **Description:** *short description of the dataset (or link to social media or blog post)*
- **Paper:** *link to the dataset paper if available*
- **Data:** *link to the Github repository or current dataset location*
- **Motivation:** *what are some good reasons t... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2889/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2889/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2888 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2888/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2888/comments | https://api.github.com/repos/huggingface/datasets/issues/2888/events | https://github.com/huggingface/datasets/issues/2888 | 992,676,535 | MDU6SXNzdWU5OTI2NzY1MzU= | 2,888 | v1.11.1 release date | {
"login": "fcakyon",
"id": 34196005,
"node_id": "MDQ6VXNlcjM0MTk2MDA1",
"avatar_url": "https://avatars.githubusercontent.com/u/34196005?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fcakyon",
"html_url": "https://github.com/fcakyon",
"followers_url": "https://api.github.com/users/fcakyo... | [
{
"id": 1935892912,
"node_id": "MDU6TGFiZWwxOTM1ODkyOTEy",
"url": "https://api.github.com/repos/huggingface/datasets/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "Further information is requested"
}
] | closed | false | null | [] | null | [
"Hi ! Probably 1.12 on monday :)\r\n",
"@albertvillanova i think this issue is still valid and should not be closed till `>1.11.0` is published :)"
] | 1,631,224,395,000 | 1,631,477,915,000 | 1,631,463,339,000 | NONE | null | Hello, i need to use latest features in one of my packages but there have been no new datasets release since 2 months ago.
When do you plan to publush v1.11.1 release? | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2888/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/huggingface/datasets/issues/2888/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2887 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2887/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2887/comments | https://api.github.com/repos/huggingface/datasets/issues/2887/events | https://github.com/huggingface/datasets/pull/2887 | 992,576,305 | MDExOlB1bGxSZXF1ZXN0NzMwODg4MTU3 | 2,887 | #2837 Use cache folder for lockfile | {
"login": "Dref360",
"id": 8976546,
"node_id": "MDQ6VXNlcjg5NzY1NDY=",
"avatar_url": "https://avatars.githubusercontent.com/u/8976546?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Dref360",
"html_url": "https://github.com/Dref360",
"followers_url": "https://api.github.com/users/Dref360/... | [] | closed | false | null | [] | null | [
"The CI fail about the meteor metric is unrelated to this PR "
] | 1,631,217,356,000 | 1,633,456,702,000 | 1,633,456,702,000 | CONTRIBUTOR | null | Fixes #2837
Use a cache folder directory to store the FileLock.
The issue was that the lock file was in a readonly folder.
| {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2887/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2887/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2887",
"html_url": "https://github.com/huggingface/datasets/pull/2887",
"diff_url": "https://github.com/huggingface/datasets/pull/2887.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2887.patch",
"merged_at": 1633456702000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2886 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2886/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2886/comments | https://api.github.com/repos/huggingface/datasets/issues/2886/events | https://github.com/huggingface/datasets/issues/2886 | 992,534,632 | MDU6SXNzdWU5OTI1MzQ2MzI= | 2,886 | Hj | {
"login": "Noorasri",
"id": 90416328,
"node_id": "MDQ6VXNlcjkwNDE2MzI4",
"avatar_url": "https://avatars.githubusercontent.com/u/90416328?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Noorasri",
"html_url": "https://github.com/Noorasri",
"followers_url": "https://api.github.com/users/Noo... | [] | closed | false | null | [] | null | [] | 1,631,213,932,000 | 1,631,274,389,000 | 1,631,274,389,000 | NONE | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2886/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2886/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2884 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2884/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2884/comments | https://api.github.com/repos/huggingface/datasets/issues/2884/events | https://github.com/huggingface/datasets/pull/2884 | 992,135,698 | MDExOlB1bGxSZXF1ZXN0NzMwNTA4MTE1 | 2,884 | Add IC, SI, ER tasks to SUPERB | {
"login": "anton-l",
"id": 26864830,
"node_id": "MDQ6VXNlcjI2ODY0ODMw",
"avatar_url": "https://avatars.githubusercontent.com/u/26864830?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/anton-l",
"html_url": "https://github.com/anton-l",
"followers_url": "https://api.github.com/users/anton-... | [] | closed | false | null | [] | null | [
"Sorry for the late PR, uploading 10+GB files to the hub through a VPN was an adventure :sweat_smile: ",
"Thank you so much for adding these subsets @anton-l! \r\n\r\n> These datasets either require manual downloads or have broken/unstable links. You can get all necessary archives in this repo: https://huggingfac... | 1,631,188,563,000 | 1,632,129,478,000 | 1,632,128,449,000 | CONTRIBUTOR | null | This PR adds 3 additional classification tasks to SUPERB
#### Intent Classification
Dataset URL seems to be down at the moment :( See the note below.
S3PRL source: https://github.com/s3prl/s3prl/blob/master/s3prl/downstream/fluent_commands/dataset.py
Instructions: https://github.com/s3prl/s3prl/tree/master/s3prl/... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2884/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2884/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2884",
"html_url": "https://github.com/huggingface/datasets/pull/2884",
"diff_url": "https://github.com/huggingface/datasets/pull/2884.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2884.patch",
"merged_at": 1632128449000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2883 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2883/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2883/comments | https://api.github.com/repos/huggingface/datasets/issues/2883/events | https://github.com/huggingface/datasets/pull/2883 | 991,969,875 | MDExOlB1bGxSZXF1ZXN0NzMwMzYzNTQz | 2,883 | Fix data URLs and metadata in DocRED dataset | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,631,177,734,000 | 1,631,532,271,000 | 1,631,532,271,000 | MEMBER | null | The host of `docred` dataset has updated the `dev` data file. This PR:
- Updates the dev URL
- Updates dataset metadata
This PR also fixes the URL of the `train_distant` split, which was wrong.
Fix #2882. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2883/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2883/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2883",
"html_url": "https://github.com/huggingface/datasets/pull/2883",
"diff_url": "https://github.com/huggingface/datasets/pull/2883.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2883.patch",
"merged_at": 1631532270000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2882 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2882/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2882/comments | https://api.github.com/repos/huggingface/datasets/issues/2882/events | https://github.com/huggingface/datasets/issues/2882 | 991,800,141 | MDU6SXNzdWU5OTE4MDAxNDE= | 2,882 | `load_dataset('docred')` results in a `NonMatchingChecksumError` | {
"login": "tmpr",
"id": 51313597,
"node_id": "MDQ6VXNlcjUxMzEzNTk3",
"avatar_url": "https://avatars.githubusercontent.com/u/51313597?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tmpr",
"html_url": "https://github.com/tmpr",
"followers_url": "https://api.github.com/users/tmpr/followers"... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Hi @tmpr, thanks for reporting.\r\n\r\nTwo weeks ago (23th Aug), the host of the source `docred` dataset updated one of the files (`dev.json`): you can see it [here](https://drive.google.com/drive/folders/1c5-0YwnoJx8NS6CV2f-NoTHR__BdkNqw).\r\n\r\nTherefore, the checksum needs to be updated.\r\n\r\nNormally, in th... | 1,631,166,902,000 | 1,631,532,270,000 | 1,631,532,270,000 | NONE | null | ## Describe the bug
I get consistent `NonMatchingChecksumError: Checksums didn't match for dataset source files` errors when trying to execute `datasets.load_dataset('docred')`.
## Steps to reproduce the bug
It is quasi only this code:
```python
import datasets
data = datasets.load_dataset('docred')
```
## ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2882/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2882/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2881 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2881/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2881/comments | https://api.github.com/repos/huggingface/datasets/issues/2881/events | https://github.com/huggingface/datasets/pull/2881 | 991,639,142 | MDExOlB1bGxSZXF1ZXN0NzMwMDc1OTAy | 2,881 | Add BIOSSES dataset | {
"login": "bwang482",
"id": 6764450,
"node_id": "MDQ6VXNlcjY3NjQ0NTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/6764450?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bwang482",
"html_url": "https://github.com/bwang482",
"followers_url": "https://api.github.com/users/bwang... | [] | closed | false | null | [] | null | [] | 1,631,147,736,000 | 1,631,542,840,000 | 1,631,542,840,000 | CONTRIBUTOR | null | Adding the biomedical semantic sentence similarity dataset, BIOSSES, listed in "Biomedical Datasets - BigScience Workshop 2021" | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2881/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2881/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2881",
"html_url": "https://github.com/huggingface/datasets/pull/2881",
"diff_url": "https://github.com/huggingface/datasets/pull/2881.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2881.patch",
"merged_at": 1631542840000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2880 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2880/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2880/comments | https://api.github.com/repos/huggingface/datasets/issues/2880/events | https://github.com/huggingface/datasets/pull/2880 | 990,877,940 | MDExOlB1bGxSZXF1ZXN0NzI5NDIzMDMy | 2,880 | Extend support for streaming datasets that use pathlib.Path stem/suffix | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,631,090,563,000 | 1,631,193,209,000 | 1,631,193,209,000 | MEMBER | null | This PR extends the support in streaming mode for datasets that use `pathlib`, by patching the properties `pathlib.Path.stem` and `pathlib.Path.suffix`.
Related to #2876, #2874, #2866.
CC: @severo | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2880/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2880/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2880",
"html_url": "https://github.com/huggingface/datasets/pull/2880",
"diff_url": "https://github.com/huggingface/datasets/pull/2880.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2880.patch",
"merged_at": 1631193209000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2879 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2879/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2879/comments | https://api.github.com/repos/huggingface/datasets/issues/2879/events | https://github.com/huggingface/datasets/issues/2879 | 990,257,404 | MDU6SXNzdWU5OTAyNTc0MDQ= | 2,879 | In v1.4.1, all TIMIT train transcripts are "Would such an act of refusal be useful?" | {
"login": "rcgale",
"id": 2279700,
"node_id": "MDQ6VXNlcjIyNzk3MDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2279700?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rcgale",
"html_url": "https://github.com/rcgale",
"followers_url": "https://api.github.com/users/rcgale/foll... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Hi @rcgale, thanks for reporting.\r\n\r\nPlease note that this bug was fixed on `datasets` version 1.5.0: https://github.com/huggingface/datasets/commit/a23c73e526e1c30263834164f16f1fdf76722c8c#diff-f12a7a42d4673bb6c2ca5a40c92c29eb4fe3475908c84fd4ce4fad5dc2514878\r\n\r\nIf you update `datasets` version, that shoul... | 1,631,040,825,000 | 1,631,120,119,000 | 1,631,092,348,000 | NONE | null | ## Describe the bug
Using version 1.4.1 of `datasets`, TIMIT transcripts are all the same.
## Steps to reproduce the bug
I was following this tutorial
- https://huggingface.co/blog/fine-tune-wav2vec2-english
But here's a distilled repro:
```python
!pip install datasets==1.4.1
from datasets import load_datas... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2879/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2879/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2877 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2877/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2877/comments | https://api.github.com/repos/huggingface/datasets/issues/2877/events | https://github.com/huggingface/datasets/issues/2877 | 990,027,249 | MDU6SXNzdWU5OTAwMjcyNDk= | 2,877 | Don't keep the dummy data folder or dataset_infos.json when resolving data files | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | null | [] | null | [
"Hi @lhoestq I am new to huggingface datasets, I would like to work on this issue!\r\n",
"Thanks for the help :) \r\n\r\nAs mentioned in the PR, excluding files named \"dummy_data.zip\" is actually more general than excluding the files inside a \"dummy\" folder. I just did the change in the PR, I think we can mer... | 1,631,023,744,000 | 1,632,906,338,000 | 1,632,906,338,000 | MEMBER | null | When there's no dataset script, all the data files of a folder or a repository on the Hub are loaded as data files.
There are already a few exceptions:
- files starting with "." are ignored
- the dataset card "README.md" is ignored
- any file named "config.json" is ignored (currently it isn't used anywhere, but i... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2877/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2877/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2876 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2876/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2876/comments | https://api.github.com/repos/huggingface/datasets/issues/2876/events | https://github.com/huggingface/datasets/pull/2876 | 990,001,079 | MDExOlB1bGxSZXF1ZXN0NzI4NjU3MDc2 | 2,876 | Extend support for streaming datasets that use pathlib.Path.glob | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"I am thinking that ideally we should call `fs.glob()` instead...",
"Thanks, @lhoestq: the idea of adding the mock filesystem is to avoid network calls and reduce testing time ;) \r\n\r\nI have added `rglob` as well and fixed some bugs."
] | 1,631,022,225,000 | 1,631,267,449,000 | 1,631,267,448,000 | MEMBER | null | This PR extends the support in streaming mode for datasets that use `pathlib`, by patching the method `pathlib.Path.glob`.
Related to #2874, #2866.
CC: @severo | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2876/reactions",
"total_count": 2,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2876/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2876",
"html_url": "https://github.com/huggingface/datasets/pull/2876",
"diff_url": "https://github.com/huggingface/datasets/pull/2876.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2876.patch",
"merged_at": 1631267448000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2874 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2874/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2874/comments | https://api.github.com/repos/huggingface/datasets/issues/2874/events | https://github.com/huggingface/datasets/pull/2874 | 989,685,328 | MDExOlB1bGxSZXF1ZXN0NzI4Mzg2Mjg4 | 2,874 | Support streaming datasets that use pathlib | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"I've tried https://github.com/huggingface/datasets/issues/2866 again, and I get the same error.\r\n\r\n```python\r\nimport datasets as ds\r\nds.load_dataset('counter', split=\"train\", streaming=False)\r\n```",
"@severo Issue #2866 is not fully fixed yet: multiple patches need to be implemented for `pathlib`, as... | 1,631,000,149,000 | 1,631,039,122,000 | 1,631,014,875,000 | MEMBER | null | This PR extends the support in streaming mode for datasets that use `pathlib.Path`.
Related to: #2866.
CC: @severo | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2874/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2874/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2874",
"html_url": "https://github.com/huggingface/datasets/pull/2874",
"diff_url": "https://github.com/huggingface/datasets/pull/2874.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2874.patch",
"merged_at": 1631014875000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2873 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2873/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2873/comments | https://api.github.com/repos/huggingface/datasets/issues/2873/events | https://github.com/huggingface/datasets/pull/2873 | 989,587,695 | MDExOlB1bGxSZXF1ZXN0NzI4MzA0MTMw | 2,873 | adding swedish_medical_ner | {
"login": "bwang482",
"id": 6764450,
"node_id": "MDQ6VXNlcjY3NjQ0NTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/6764450?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bwang482",
"html_url": "https://github.com/bwang482",
"followers_url": "https://api.github.com/users/bwang... | [] | closed | false | null | [] | null | [
"Hi, what's the current status of this request? It says Changes requested, but I can't see what changes?",
"Hi, it looks like this PR includes changes to other files that `swedish_medical_ner`.\r\n\r\nFeel free to remove these changes, or simply create a new PR that only contains the addition of the dataset"
] | 1,630,989,893,000 | 1,631,911,657,000 | 1,631,911,657,000 | CONTRIBUTOR | null | Adding the Swedish Medical NER dataset, listed in "Biomedical Datasets - BigScience Workshop 2021"
Code refactored | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2873/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2873/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2873",
"html_url": "https://github.com/huggingface/datasets/pull/2873",
"diff_url": "https://github.com/huggingface/datasets/pull/2873.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2873.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2872 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2872/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2872/comments | https://api.github.com/repos/huggingface/datasets/issues/2872/events | https://github.com/huggingface/datasets/pull/2872 | 989,453,069 | MDExOlB1bGxSZXF1ZXN0NzI4MTkzMjkz | 2,872 | adding swedish_medical_ner | {
"login": "bwang482",
"id": 6764450,
"node_id": "MDQ6VXNlcjY3NjQ0NTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/6764450?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bwang482",
"html_url": "https://github.com/bwang482",
"followers_url": "https://api.github.com/users/bwang... | [] | closed | false | null | [] | null | [] | 1,630,965,652,000 | 1,630,989,392,000 | 1,630,989,392,000 | CONTRIBUTOR | null | Adding the Swedish Medical NER dataset, listed in "Biomedical Datasets - BigScience Workshop 2021" | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2872/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2872/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2872",
"html_url": "https://github.com/huggingface/datasets/pull/2872",
"diff_url": "https://github.com/huggingface/datasets/pull/2872.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2872.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2871 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2871/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2871/comments | https://api.github.com/repos/huggingface/datasets/issues/2871/events | https://github.com/huggingface/datasets/issues/2871 | 989,436,088 | MDU6SXNzdWU5ODk0MzYwODg= | 2,871 | datasets.config.PYARROW_VERSION has no attribute 'major' | {
"login": "bwang482",
"id": 6764450,
"node_id": "MDQ6VXNlcjY3NjQ0NTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/6764450?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bwang482",
"html_url": "https://github.com/bwang482",
"followers_url": "https://api.github.com/users/bwang... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"I have changed line 288 to `if int(datasets.config.PYARROW_VERSION.split(\".\")[0]) < 3:` just to get around it.",
"Hi @bwang482,\r\n\r\nI'm sorry but I'm not able to reproduce your bug.\r\n\r\nPlease note that in our current master branch, we made a commit (d03223d4d64b89e76b48b00602aba5aa2f817f1e) that simulta... | 1,630,962,417,000 | 1,631,091,112,000 | 1,631,091,112,000 | CONTRIBUTOR | null | In the test_dataset_common.py script, line 288-289
```
if datasets.config.PYARROW_VERSION.major < 3:
packaged_datasets = [pd for pd in packaged_datasets if pd["dataset_name"] != "parquet"]
```
which throws the error below. `datasets.config.PYARROW_VERSION` itself return the string '4.0.1'. I have tested thi... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2871/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2871/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2870 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2870/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2870/comments | https://api.github.com/repos/huggingface/datasets/issues/2870/events | https://github.com/huggingface/datasets/pull/2870 | 988,276,859 | MDExOlB1bGxSZXF1ZXN0NzI3MjI4Njk5 | 2,870 | Fix three typos in two files for documentation | {
"login": "leny-mi",
"id": 25124853,
"node_id": "MDQ6VXNlcjI1MTI0ODUz",
"avatar_url": "https://avatars.githubusercontent.com/u/25124853?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/leny-mi",
"html_url": "https://github.com/leny-mi",
"followers_url": "https://api.github.com/users/leny-m... | [] | closed | false | null | [] | null | [] | 1,630,756,183,000 | 1,630,916,481,000 | 1,630,916,375,000 | CONTRIBUTOR | null | Changed "bacth_size" to "batch_size" (2x)
Changed "intsructions" to "instructions" | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2870/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2870/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2870",
"html_url": "https://github.com/huggingface/datasets/pull/2870",
"diff_url": "https://github.com/huggingface/datasets/pull/2870.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2870.patch",
"merged_at": 1630916375000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2869 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2869/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2869/comments | https://api.github.com/repos/huggingface/datasets/issues/2869/events | https://github.com/huggingface/datasets/issues/2869 | 987,676,420 | MDU6SXNzdWU5ODc2NzY0MjA= | 2,869 | TypeError: 'NoneType' object is not callable | {
"login": "Chenfei-Kang",
"id": 40911446,
"node_id": "MDQ6VXNlcjQwOTExNDQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/40911446?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Chenfei-Kang",
"html_url": "https://github.com/Chenfei-Kang",
"followers_url": "https://api.github.c... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Hi, @Chenfei-Kang.\r\n\r\nI'm sorry, but I'm not able to reproduce your bug:\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nds = load_dataset(\"glue\", 'cola')\r\nds\r\n```\r\n```\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['sentence', 'label', 'idx'],\r\n num_rows: 8551\r\n ... | 1,630,668,459,000 | 1,631,102,998,000 | 1,631,093,095,000 | NONE | null | ## Describe the bug
TypeError: 'NoneType' object is not callable
## Steps to reproduce the bug
```python
from datasets import load_dataset, load_metric
dataset = datasets.load_dataset("glue", 'cola')
```
## Expected results
A clear and concise description of the expected results.
## Actual results
Speci... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2869/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2869/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2867 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2867/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2867/comments | https://api.github.com/repos/huggingface/datasets/issues/2867/events | https://github.com/huggingface/datasets/pull/2867 | 986,971,224 | MDExOlB1bGxSZXF1ZXN0NzI2MTE3NzAw | 2,867 | Add CaSiNo dataset | {
"login": "kushalchawla",
"id": 8416863,
"node_id": "MDQ6VXNlcjg0MTY4NjM=",
"avatar_url": "https://avatars.githubusercontent.com/u/8416863?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kushalchawla",
"html_url": "https://github.com/kushalchawla",
"followers_url": "https://api.github.com... | [] | closed | false | null | [] | null | [
"Hi @lhoestq \r\n\r\nJust a request to look at the dataset. Please let me know if any changes are necessary before merging it into the repo. Thank you.",
"Hey @lhoestq \r\n\r\nThanks for merging it. One question: I still cannot find the dataset on https://huggingface.co/datasets. Does it take some time or did I ... | 1,630,602,383,000 | 1,631,805,174,000 | 1,631,784,224,000 | CONTRIBUTOR | null | Hi. I request you to add our dataset to the repository.
This data was recently published at NAACL 2021: https://aclanthology.org/2021.naacl-main.254.pdf | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2867/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2867/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2867",
"html_url": "https://github.com/huggingface/datasets/pull/2867",
"diff_url": "https://github.com/huggingface/datasets/pull/2867.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2867.patch",
"merged_at": 1631784224000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2866 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2866/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2866/comments | https://api.github.com/repos/huggingface/datasets/issues/2866/events | https://github.com/huggingface/datasets/issues/2866 | 986,706,676 | MDU6SXNzdWU5ODY3MDY2NzY= | 2,866 | "counter" dataset raises an error in normal mode, but not in streaming mode | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"Hi @severo, thanks for reporting.\r\n\r\nJust note that currently not all canonical datasets support streaming mode: this is one case!\r\n\r\nAll datasets that use `pathlib` joins (using `/`) instead of `os.path.join` (as in this dataset) do not support streaming mode yet.",
"OK. Do you think it's possible to de... | 1,630,588,253,000 | 1,634,203,449,000 | 1,634,203,449,000 | CONTRIBUTOR | null | ## Describe the bug
`counter` dataset raises an error on `load_dataset()`, but simply returns an empty iterator in streaming mode.
## Steps to reproduce the bug
```python
>>> import datasets as ds
>>> a = ds.load_dataset('counter', split="train", streaming=False)
Using custom data configuration default
Dow... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2866/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2866/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2865 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2865/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2865/comments | https://api.github.com/repos/huggingface/datasets/issues/2865/events | https://github.com/huggingface/datasets/pull/2865 | 986,460,698 | MDExOlB1bGxSZXF1ZXN0NzI1NjY1ODgx | 2,865 | Add MultiEURLEX dataset | {
"login": "iliaschalkidis",
"id": 1626984,
"node_id": "MDQ6VXNlcjE2MjY5ODQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/1626984?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/iliaschalkidis",
"html_url": "https://github.com/iliaschalkidis",
"followers_url": "https://api.gith... | [] | closed | false | null | [] | null | [
"Hi @lhoestq, we have this new cool multilingual dataset coming at EMNLP 2021. It would be really nice if we could have it in Hugging Face asap. Thanks! ",
"Hi @lhoestq, I adopted most of your suggestions:\r\n\r\n- Dummy data files reduced, including the 2 smallest documents per subset JSONL.\r\n- README was upda... | 1,630,575,744,000 | 1,631,274,606,000 | 1,631,274,606,000 | CONTRIBUTOR | null | **Add new MultiEURLEX Dataset**
MultiEURLEX comprises 65k EU laws in 23 official EU languages (some low-ish resource). Each EU law has been annotated with EUROVOC concepts (labels) by the Publication Office of EU. As with the English EURLEX, the goal is to predict the relevant EUROVOC concepts (labels); this is mult... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2865/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2865/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2865",
"html_url": "https://github.com/huggingface/datasets/pull/2865",
"diff_url": "https://github.com/huggingface/datasets/pull/2865.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2865.patch",
"merged_at": 1631274606000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2864 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2864/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2864/comments | https://api.github.com/repos/huggingface/datasets/issues/2864/events | https://github.com/huggingface/datasets/pull/2864 | 986,159,438 | MDExOlB1bGxSZXF1ZXN0NzI1MzkyNjcw | 2,864 | Fix data URL in ToTTo dataset | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | {
"url": "https://api.github.com/repos/huggingface/datasets/milestones/8",
"html_url": "https://github.com/huggingface/datasets/milestone/8",
"labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/8/labels",
"id": 6968069,
"node_id": "MI_kwDODunzps4AalMF",
"number": 8,
"title": "1.12"... | [] | 1,630,560,308,000 | 1,630,565,260,000 | 1,630,565,260,000 | MEMBER | null | Data source host changed their data URL: google-research-datasets/ToTTo@cebeb43.
Fix #2860. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2864/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2864/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2864",
"html_url": "https://github.com/huggingface/datasets/pull/2864",
"diff_url": "https://github.com/huggingface/datasets/pull/2864.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2864.patch",
"merged_at": 1630565260000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2863 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2863/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2863/comments | https://api.github.com/repos/huggingface/datasets/issues/2863/events | https://github.com/huggingface/datasets/pull/2863 | 986,156,755 | MDExOlB1bGxSZXF1ZXN0NzI1MzkwMTkx | 2,863 | Update dataset URL | {
"login": "mrm8488",
"id": 3653789,
"node_id": "MDQ6VXNlcjM2NTM3ODk=",
"avatar_url": "https://avatars.githubusercontent.com/u/3653789?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mrm8488",
"html_url": "https://github.com/mrm8488",
"followers_url": "https://api.github.com/users/mrm8488/... | [] | closed | false | null | [] | null | [
"Superseded by PR #2864.\r\n\r\n@mrm8488 next time you would like to work on an issue, you can first self-assign it to you (by writing `#self-assign` in a comment on the issue). That way, other people can see you are already working on it and there are not multiple people working on the same issue. 😉 "
] | 1,630,560,138,000 | 1,630,570,250,000 | 1,630,570,250,000 | NONE | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2863/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2863/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2863",
"html_url": "https://github.com/huggingface/datasets/pull/2863",
"diff_url": "https://github.com/huggingface/datasets/pull/2863.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2863.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2861 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2861/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2861/comments | https://api.github.com/repos/huggingface/datasets/issues/2861/events | https://github.com/huggingface/datasets/pull/2861 | 985,081,871 | MDExOlB1bGxSZXF1ZXN0NzI0NDM2OTcw | 2,861 | fix: 🐛 be more specific when catching exceptions | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [] | closed | false | null | [] | null | [
"To give more context: after our discussion, if I understood properly, you are trying to fix a call to `datasets` that takes 15 minutes: https://github.com/huggingface/datasets-preview-backend/issues/17 Is this right?\r\n\r\n",
"Yes, that's it. And to do that I'm trying to use https://pypi.org/project/stopit/, wh... | 1,630,498,692,000 | 1,630,576,416,000 | 1,630,576,323,000 | CONTRIBUTOR | null | The same specific exception is catched in other parts of the same
function. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2861/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2861/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2861",
"html_url": "https://github.com/huggingface/datasets/pull/2861",
"diff_url": "https://github.com/huggingface/datasets/pull/2861.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2861.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2860 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2860/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2860/comments | https://api.github.com/repos/huggingface/datasets/issues/2860/events | https://github.com/huggingface/datasets/issues/2860 | 985,013,339 | MDU6SXNzdWU5ODUwMTMzMzk= | 2,860 | Cannot download TOTTO dataset | {
"login": "mrm8488",
"id": 3653789,
"node_id": "MDQ6VXNlcjM2NTM3ODk=",
"avatar_url": "https://avatars.githubusercontent.com/u/3653789?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mrm8488",
"html_url": "https://github.com/mrm8488",
"followers_url": "https://api.github.com/users/mrm8488/... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Hola @mrm8488, thanks for reporting.\r\n\r\nApparently, the data source host changed their URL one week ago: https://github.com/google-research-datasets/ToTTo/commit/cebeb430ec2a97747e704d16a9354f7d9073ff8f\r\n\r\nI'm fixing it."
] | 1,630,494,250,000 | 1,630,565,260,000 | 1,630,565,260,000 | NONE | null | Error: Couldn't find file at https://storage.googleapis.com/totto/totto_data.zip
`datasets version: 1.11.0`
# How to reproduce:
```py
from datasets import load_dataset
dataset = load_dataset('totto')
```
| {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2860/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2860/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2859 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2859/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2859/comments | https://api.github.com/repos/huggingface/datasets/issues/2859/events | https://github.com/huggingface/datasets/issues/2859 | 984,324,500 | MDU6SXNzdWU5ODQzMjQ1MDA= | 2,859 | Loading allenai/c4 in streaming mode does too many HEAD requests | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
},
{
"id": 3287858981,
"node_id": "MDU6... | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"https://github.com/huggingface/datasets/blob/6c766f9115d686182d76b1b937cb27e099c45d68/src/datasets/builder.py#L179-L186",
"Thanks a lot!!!"
] | 1,630,444,264,000 | 1,634,024,152,000 | 1,633,950,351,000 | MEMBER | null | This does 60,000+ HEAD requests to get all the ETags of all the data files:
```python
from datasets import load_dataset
load_dataset("allenai/c4", streaming=True)
```
It makes loading the dataset completely impractical.
The ETags are used to compute the config id (it must depend on the data files being used).
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2859/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2859/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2858 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2858/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2858/comments | https://api.github.com/repos/huggingface/datasets/issues/2858/events | https://github.com/huggingface/datasets/pull/2858 | 984,145,568 | MDExOlB1bGxSZXF1ZXN0NzIzNjEzNzQ0 | 2,858 | Fix s3fs version in CI | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [] | 1,630,433,143,000 | 1,630,935,215,000 | 1,630,445,391,000 | MEMBER | null | The latest s3fs version has new constrains on aiobotocore, and therefore on boto3 and botocore
This PR changes the constrains to avoid the new conflicts
In particular it pins the version of s3fs. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2858/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2858/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2858",
"html_url": "https://github.com/huggingface/datasets/pull/2858",
"diff_url": "https://github.com/huggingface/datasets/pull/2858.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2858.patch",
"merged_at": 1630445391000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2857 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2857/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2857/comments | https://api.github.com/repos/huggingface/datasets/issues/2857/events | https://github.com/huggingface/datasets/pull/2857 | 984,093,938 | MDExOlB1bGxSZXF1ZXN0NzIzNTY5OTE4 | 2,857 | Update: Openwebtext - update size | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"merging since the CI error in unrelated to this PR and fixed on master"
] | 1,630,429,863,000 | 1,631,007,872,000 | 1,631,007,872,000 | MEMBER | null | Update the size of the Openwebtext dataset
I also regenerated the dataset_infos.json but the data file checksum didn't change, and the number of examples either (8013769 examples)
related to #2839 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2857/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2857/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2857",
"html_url": "https://github.com/huggingface/datasets/pull/2857",
"diff_url": "https://github.com/huggingface/datasets/pull/2857.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2857.patch",
"merged_at": 1631007872000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2856 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2856/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2856/comments | https://api.github.com/repos/huggingface/datasets/issues/2856/events | https://github.com/huggingface/datasets/pull/2856 | 983,876,734 | MDExOlB1bGxSZXF1ZXN0NzIzMzg2NzIw | 2,856 | fix: 🐛 remove URL's query string only if it's ?dl=1 | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [] | closed | false | null | [] | null | [] | 1,630,417,207,000 | 1,630,419,732,000 | 1,630,419,732,000 | CONTRIBUTOR | null | A lot of URL use the query strings, for example
http://opus.nlpl.eu/download.php?f=Bianet/v1/moses/en-ku.txt.zip, we
must not remove it when trying to detect the protocol. We thus remove it
only in the case of the query string being ?dl=1 which occurs on dropbox
and dl.orangedox.com. Also: add unit tests.
See ht... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2856/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2856/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2856",
"html_url": "https://github.com/huggingface/datasets/pull/2856",
"diff_url": "https://github.com/huggingface/datasets/pull/2856.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2856.patch",
"merged_at": 1630419732000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2855 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2855/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2855/comments | https://api.github.com/repos/huggingface/datasets/issues/2855/events | https://github.com/huggingface/datasets/pull/2855 | 983,858,229 | MDExOlB1bGxSZXF1ZXN0NzIzMzcxMTIy | 2,855 | Fix windows CI CondaError | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [] | 1,630,416,122,000 | 1,630,416,934,000 | 1,630,416,933,000 | MEMBER | null | From this thread: https://github.com/conda/conda/issues/6057
We can fix the conda error
```
CondaError: Cannot link a source that does not exist.
C:\Users\...\Anaconda3\Scripts\conda.exe
```
by doing
```bash
conda update conda
```
before doing any install in the windows CI | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2855/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2855/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2855",
"html_url": "https://github.com/huggingface/datasets/pull/2855",
"diff_url": "https://github.com/huggingface/datasets/pull/2855.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2855.patch",
"merged_at": 1630416933000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2854 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2854/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2854/comments | https://api.github.com/repos/huggingface/datasets/issues/2854/events | https://github.com/huggingface/datasets/pull/2854 | 983,726,084 | MDExOlB1bGxSZXF1ZXN0NzIzMjU3NDg5 | 2,854 | Fix caching when moving script | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"Merging since the CI failure is unrelated to this PR"
] | 1,630,407,515,000 | 1,630,415,616,000 | 1,630,415,616,000 | MEMBER | null | When caching the result of a `map` function, the hash that is computed depends on many properties of this function, such as all the python objects it uses, its code and also the location of this code.
Using the full path of the python script for the location of the code makes the hash change if a script like `run_ml... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2854/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2854/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2854",
"html_url": "https://github.com/huggingface/datasets/pull/2854",
"diff_url": "https://github.com/huggingface/datasets/pull/2854.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2854.patch",
"merged_at": 1630415616000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2853 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2853/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2853/comments | https://api.github.com/repos/huggingface/datasets/issues/2853/events | https://github.com/huggingface/datasets/pull/2853 | 983,692,026 | MDExOlB1bGxSZXF1ZXN0NzIzMjI4NDY3 | 2,853 | Add AMI dataset | {
"login": "cahya-wirawan",
"id": 7669893,
"node_id": "MDQ6VXNlcjc2Njk4OTM=",
"avatar_url": "https://avatars.githubusercontent.com/u/7669893?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/cahya-wirawan",
"html_url": "https://github.com/cahya-wirawan",
"followers_url": "https://api.github.... | [] | closed | false | null | [] | null | [
"Hey @cahya-wirawan, \r\n\r\nI played around with the dataset a bit and it looks already very good to me! That's exactly how it should be constructed :-) I can help you a bit with defining the config, etc... on Monday!",
"@lhoestq - I think the dataset is ready to be merged :-) \r\n\r\nAt the moment, I don't real... | 1,630,405,141,000 | 1,632,907,159,000 | 1,632,907,159,000 | CONTRIBUTOR | null | This is an initial commit for AMI dataset | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2853/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 2,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2853/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2853",
"html_url": "https://github.com/huggingface/datasets/pull/2853",
"diff_url": "https://github.com/huggingface/datasets/pull/2853.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2853.patch",
"merged_at": 1632907158000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2852 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2852/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2852/comments | https://api.github.com/repos/huggingface/datasets/issues/2852/events | https://github.com/huggingface/datasets/pull/2852 | 983,609,352 | MDExOlB1bGxSZXF1ZXN0NzIzMTU4Mzc4 | 2,852 | Fix: linnaeus - fix url | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"Merging since the CI error is unrelated this this PR"
] | 1,630,399,873,000 | 1,630,415,530,000 | 1,630,415,529,000 | MEMBER | null | The url was causing a `ConnectionError` because of the "/" at the end
Close https://github.com/huggingface/datasets/issues/2821 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2852/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2852/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2852",
"html_url": "https://github.com/huggingface/datasets/pull/2852",
"diff_url": "https://github.com/huggingface/datasets/pull/2852.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2852.patch",
"merged_at": 1630415529000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2851 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2851/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2851/comments | https://api.github.com/repos/huggingface/datasets/issues/2851/events | https://github.com/huggingface/datasets/pull/2851 | 982,789,593 | MDExOlB1bGxSZXF1ZXN0NzIyNDg4MDY2 | 2,851 | Update `column_names` showed as `:func:` in exploring.st | {
"login": "ClementRomac",
"id": 8899812,
"node_id": "MDQ6VXNlcjg4OTk4MTI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8899812?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ClementRomac",
"html_url": "https://github.com/ClementRomac",
"followers_url": "https://api.github.com... | [] | closed | false | null | [] | null | [] | 1,630,329,706,000 | 1,630,485,731,000 | 1,630,421,146,000 | CONTRIBUTOR | null | Hi,
One mention of `column_names` in exploring.st was showing it as `:func:` instead of `:attr:`. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2851/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2851/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2851",
"html_url": "https://github.com/huggingface/datasets/pull/2851",
"diff_url": "https://github.com/huggingface/datasets/pull/2851.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2851.patch",
"merged_at": 1630421146000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2848 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2848/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2848/comments | https://api.github.com/repos/huggingface/datasets/issues/2848/events | https://github.com/huggingface/datasets/pull/2848 | 981,953,908 | MDExOlB1bGxSZXF1ZXN0NzIxODYyMDQx | 2,848 | Update README.md | {
"login": "odellus",
"id": 4686956,
"node_id": "MDQ6VXNlcjQ2ODY5NTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4686956?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/odellus",
"html_url": "https://github.com/odellus",
"followers_url": "https://api.github.com/users/odellus/... | [] | closed | false | null | [] | null | [
"Merging since the CI error is unrelated to this PR and fixed on master"
] | 1,630,195,106,000 | 1,631,007,632,000 | 1,631,007,632,000 | CONTRIBUTOR | null | Changed 'Tain' to 'Train'. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2848/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2848/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2848",
"html_url": "https://github.com/huggingface/datasets/pull/2848",
"diff_url": "https://github.com/huggingface/datasets/pull/2848.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2848.patch",
"merged_at": 1631007632000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2847 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2847/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2847/comments | https://api.github.com/repos/huggingface/datasets/issues/2847/events | https://github.com/huggingface/datasets/pull/2847 | 981,589,693 | MDExOlB1bGxSZXF1ZXN0NzIxNjA3OTA0 | 2,847 | fix regex to accept negative timezone | {
"login": "jadermcs",
"id": 7156771,
"node_id": "MDQ6VXNlcjcxNTY3NzE=",
"avatar_url": "https://avatars.githubusercontent.com/u/7156771?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jadermcs",
"html_url": "https://github.com/jadermcs",
"followers_url": "https://api.github.com/users/jader... | [] | closed | false | null | [] | null | [] | 1,630,097,645,000 | 1,631,565,590,000 | 1,631,007,263,000 | CONTRIBUTOR | null | fix #2846 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2847/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2847/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2847",
"html_url": "https://github.com/huggingface/datasets/pull/2847",
"diff_url": "https://github.com/huggingface/datasets/pull/2847.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2847.patch",
"merged_at": 1631007263000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2846 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2846/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2846/comments | https://api.github.com/repos/huggingface/datasets/issues/2846/events | https://github.com/huggingface/datasets/issues/2846 | 981,587,590 | MDU6SXNzdWU5ODE1ODc1OTA= | 2,846 | Negative timezone | {
"login": "jadermcs",
"id": 7156771,
"node_id": "MDQ6VXNlcjcxNTY3NzE=",
"avatar_url": "https://avatars.githubusercontent.com/u/7156771?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jadermcs",
"html_url": "https://github.com/jadermcs",
"followers_url": "https://api.github.com/users/jader... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Fixed by #2847."
] | 1,630,097,433,000 | 1,631,274,667,000 | 1,631,274,667,000 | CONTRIBUTOR | null | ## Describe the bug
The load_dataset method do not accept a parquet file with a negative timezone, as it has the following regex:
```
"^(s|ms|us|ns),\s*tz=([a-zA-Z0-9/_+:]*)$"
```
So a valid timestap ```timestamp[us, tz=-03:00]``` returns an error when loading parquet files.
## Steps to reproduce the bug
```py... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2846/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2846/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2844 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2844/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2844/comments | https://api.github.com/repos/huggingface/datasets/issues/2844/events | https://github.com/huggingface/datasets/pull/2844 | 981,382,806 | MDExOlB1bGxSZXF1ZXN0NzIxNDQzMjY2 | 2,844 | Fix: wikicorpus - fix keys | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"The CI error is unrelated to this PR\r\n\r\n... merging !"
] | 1,630,079,766,000 | 1,630,937,248,000 | 1,630,937,247,000 | MEMBER | null | As mentioned in https://github.com/huggingface/datasets/issues/2552, there is a duplicate keys error in `wikicorpus`.
I fixed that by taking into account the file index in the keys | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2844/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2844/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2844",
"html_url": "https://github.com/huggingface/datasets/pull/2844",
"diff_url": "https://github.com/huggingface/datasets/pull/2844.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2844.patch",
"merged_at": 1630937247000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2843 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2843/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2843/comments | https://api.github.com/repos/huggingface/datasets/issues/2843/events | https://github.com/huggingface/datasets/pull/2843 | 981,317,775 | MDExOlB1bGxSZXF1ZXN0NzIxMzkwODA5 | 2,843 | Fix extraction protocol inference from urls with params | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"merging since the windows error is just a CircleCI issue",
"It works, eg https://observablehq.com/@huggingface/datasets-preview-backend-client#{%22datasetId%22%3A%22discovery%22} and https://datasets-preview.huggingface.tech/rows?dataset=discovery&config=discovery&split=train",
"Nice !"
] | 1,630,075,257,000 | 1,630,343,509,000 | 1,630,329,121,000 | MEMBER | null | Previously it was unable to infer the compression protocol for files at URLs like
```
https://foo.bar/train.json.gz?dl=1
```
because of the query parameters.
I fixed that, this should allow 10+ datasets to work in streaming mode:
```
"discovery",
"emotion",
"grail_qa",
"guardian_authorship",
"pra... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2843/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2843/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2843",
"html_url": "https://github.com/huggingface/datasets/pull/2843",
"diff_url": "https://github.com/huggingface/datasets/pull/2843.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2843.patch",
"merged_at": 1630329121000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2842 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2842/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2842/comments | https://api.github.com/repos/huggingface/datasets/issues/2842/events | https://github.com/huggingface/datasets/issues/2842 | 980,725,899 | MDU6SXNzdWU5ODA3MjU4OTk= | 2,842 | always requiring the username in the dataset name when there is one | {
"login": "stas00",
"id": 10676103,
"node_id": "MDQ6VXNlcjEwNjc2MTAz",
"avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/stas00",
"html_url": "https://github.com/stas00",
"followers_url": "https://api.github.com/users/stas00/fo... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"From what I can understand, you want the saved arrow file directory to have username as well instead of just dataset name if it was downloaded with the user prefix?",
"I don't think the user cares of how this is done, but the 2nd command should fail, IMHO, as its dataset name is invalid:\r\n```\r\n# first run\r\... | 1,630,020,713,000 | 1,634,895,815,000 | 1,634,895,815,000 | CONTRIBUTOR | null | Me and now another person have been bitten by the `datasets`'s non-strictness on requiring a dataset creator's username when it's due.
So both of us started with `stas/openwebtext-10k`, somewhere along the lines lost `stas/` and continued using `openwebtext-10k` and it all was good until we published the software an... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2842/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2842/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2840 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2840/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2840/comments | https://api.github.com/repos/huggingface/datasets/issues/2840/events | https://github.com/huggingface/datasets/issues/2840 | 980,489,074 | MDU6SXNzdWU5ODA0ODkwNzQ= | 2,840 | How can I compute BLEU-4 score use `load_metric` ? | {
"login": "Doragd",
"id": 26213546,
"node_id": "MDQ6VXNlcjI2MjEzNTQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/26213546?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Doragd",
"html_url": "https://github.com/Doragd",
"followers_url": "https://api.github.com/users/Doragd/fo... | [] | closed | false | null | [] | null | [] | 1,629,999,397,000 | 1,630,052,004,000 | 1,630,052,004,000 | NONE | null | I have found the sacrebleu metric. But, I do not know the difference between it and BLEU-4.
If I want to compute BLEU-4 score, what can i do? | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2840/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2840/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2839 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2839/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2839/comments | https://api.github.com/repos/huggingface/datasets/issues/2839/events | https://github.com/huggingface/datasets/issues/2839 | 980,271,715 | MDU6SXNzdWU5ODAyNzE3MTU= | 2,839 | OpenWebText: NonMatchingSplitsSizesError | {
"login": "thomasw21",
"id": 24695242,
"node_id": "MDQ6VXNlcjI0Njk1MjQy",
"avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/thomasw21",
"html_url": "https://github.com/thomasw21",
"followers_url": "https://api.github.com/users/... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"Thanks for reporting, I'm updating the verifications metadata",
"I just regenerated the verifications metadata and noticed that nothing changed: the data file is fine (the checksum didn't change), and the number of examples is still 8013769. Not sure how you managed to get 7982430 examples.\r\n\r\nCan you try to... | 1,629,985,826,000 | 1,632,233,560,000 | 1,632,233,383,000 | CONTRIBUTOR | null | ## Describe the bug
When downloading `openwebtext`, I'm getting:
```
datasets.utils.info_utils.NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=39769494896, num_examples=8013769, dataset_name='openwebtext'), 'recorded': SplitInfo(name='train', num_bytes=39611023912, num_examples=7982430... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2839/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2839/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2837 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2837/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2837/comments | https://api.github.com/repos/huggingface/datasets/issues/2837/events | https://github.com/huggingface/datasets/issues/2837 | 979,298,297 | MDU6SXNzdWU5NzkyOTgyOTc= | 2,837 | prepare_module issue when loading from read-only fs | {
"login": "Dref360",
"id": 8976546,
"node_id": "MDQ6VXNlcjg5NzY1NDY=",
"avatar_url": "https://avatars.githubusercontent.com/u/8976546?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Dref360",
"html_url": "https://github.com/Dref360",
"followers_url": "https://api.github.com/users/Dref360/... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Hello, I opened #2887 to fix this."
] | 1,629,904,886,000 | 1,633,456,702,000 | 1,633,456,702,000 | CONTRIBUTOR | null | ## Describe the bug
When we use prepare_module from a readonly file system, we create a FileLock using the `local_path`.
This path is not necessarily writable.
`lock_path = local_path + ".lock"`
## Steps to reproduce the bug
Run `load_dataset` on a readonly python loader file.
```python
ds = load_datas... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2837/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2837/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2836 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2836/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2836/comments | https://api.github.com/repos/huggingface/datasets/issues/2836/events | https://github.com/huggingface/datasets/pull/2836 | 979,230,142 | MDExOlB1bGxSZXF1ZXN0NzE5NjY5MDUy | 2,836 | Optimize Dataset.filter to only compute the indices to keep | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"Maybe worth updating the docs here as well?",
"Yup, will do !"
] | 1,629,902,482,000 | 1,631,631,113,000 | 1,631,548,221,000 | MEMBER | null | Optimize `Dataset.filter` to only compute the indices of the rows to keep, instead of creating a new Arrow table with the rows to keep. Creating a new table was an issue because it could take a lot of disk space.
This will be useful to process audio datasets for example cc @patrickvonplaten | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2836/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 1,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2836/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2836",
"html_url": "https://github.com/huggingface/datasets/pull/2836",
"diff_url": "https://github.com/huggingface/datasets/pull/2836.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2836.patch",
"merged_at": 1631548221000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2835 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2835/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2835/comments | https://api.github.com/repos/huggingface/datasets/issues/2835/events | https://github.com/huggingface/datasets/pull/2835 | 979,209,394 | MDExOlB1bGxSZXF1ZXN0NzE5NjUxOTE4 | 2,835 | Update: timit_asr - make the dataset streamable | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [] | 1,629,901,369,000 | 1,631,020,547,000 | 1,631,020,546,000 | MEMBER | null | The TIMIT ASR dataset had two issues that was preventing it from being streamable:
1. it was missing a call to `open` before `pd.read_csv`
2. it was using `os.path.dirname` which is not supported for streaming
I made the dataset streamable by using `open` to load the CSV, and by adding the support for `os.path.d... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2835/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2835/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2835",
"html_url": "https://github.com/huggingface/datasets/pull/2835",
"diff_url": "https://github.com/huggingface/datasets/pull/2835.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2835.patch",
"merged_at": 1631020546000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2834 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2834/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2834/comments | https://api.github.com/repos/huggingface/datasets/issues/2834/events | https://github.com/huggingface/datasets/pull/2834 | 978,309,749 | MDExOlB1bGxSZXF1ZXN0NzE4OTE5NjQ0 | 2,834 | Fix IndexError by ignoring empty RecordBatch | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [] | 1,629,824,773,000 | 1,629,825,678,000 | 1,629,825,678,000 | MEMBER | null | We need to ignore the empty record batches for the interpolation search to work correctly when querying arrow tables
Close #2833
cc @SaulLu | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2834/reactions",
"total_count": 2,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 1,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2834/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2834",
"html_url": "https://github.com/huggingface/datasets/pull/2834",
"diff_url": "https://github.com/huggingface/datasets/pull/2834.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2834.patch",
"merged_at": 1629825677000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2833 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2833/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2833/comments | https://api.github.com/repos/huggingface/datasets/issues/2833/events | https://github.com/huggingface/datasets/issues/2833 | 978,296,140 | MDU6SXNzdWU5NzgyOTYxNDA= | 2,833 | IndexError when accessing first element of a Dataset if first RecordBatch is empty | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [] | 1,629,823,760,000 | 1,629,825,677,000 | 1,629,825,677,000 | MEMBER | null | The computation of the offsets of the underlying Table of a Dataset has some issues if the first RecordBatch is empty.
```python
from datasets import Dataset
import pyarrow as pa
pa_table = pa.Table.from_pydict({"a": [1]})
pa_table2 = pa.Table.from_pydict({"a": []}, schema=pa_table.schema)
ds_table = pa.conca... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2833/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 1,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2833/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2828 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2828/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2828/comments | https://api.github.com/repos/huggingface/datasets/issues/2828/events | https://github.com/huggingface/datasets/pull/2828 | 977,181,517 | MDExOlB1bGxSZXF1ZXN0NzE3OTYwODg3 | 2,828 | Add code-mixed Kannada Hope speech dataset | {
"login": "adeepH",
"id": 46108405,
"node_id": "MDQ6VXNlcjQ2MTA4NDA1",
"avatar_url": "https://avatars.githubusercontent.com/u/46108405?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/adeepH",
"html_url": "https://github.com/adeepH",
"followers_url": "https://api.github.com/users/adeepH/fo... | [] | closed | false | null | [] | null | [] | 1,629,734,109,000 | 1,633,108,863,000 | 1,633,108,863,000 | CONTRIBUTOR | null | ## Adding a Dataset
- **Name:** *KanHope*
- **Description:** *A code-mixed English-Kannada dataset for Hope speech detection*
- **Paper:** *https://arxiv.org/abs/2108.04616*
- **Data:** *https://github.com/adeepH/KanHope/tree/main/dataset*
- **Motivation:** *The dataset is amongst the very few resources available... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2828/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2828/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2828",
"html_url": "https://github.com/huggingface/datasets/pull/2828",
"diff_url": "https://github.com/huggingface/datasets/pull/2828.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2828.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2827 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2827/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2827/comments | https://api.github.com/repos/huggingface/datasets/issues/2827/events | https://github.com/huggingface/datasets/pull/2827 | 976,976,552 | MDExOlB1bGxSZXF1ZXN0NzE3Nzg3MjEw | 2,827 | add a text classification dataset | {
"login": "adeepH",
"id": 46108405,
"node_id": "MDQ6VXNlcjQ2MTA4NDA1",
"avatar_url": "https://avatars.githubusercontent.com/u/46108405?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/adeepH",
"html_url": "https://github.com/adeepH",
"followers_url": "https://api.github.com/users/adeepH/fo... | [] | closed | false | null | [] | null | [] | 1,629,721,481,000 | 1,629,733,878,000 | 1,629,733,878,000 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2827/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2827/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2827",
"html_url": "https://github.com/huggingface/datasets/pull/2827",
"diff_url": "https://github.com/huggingface/datasets/pull/2827.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2827.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2826 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2826/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2826/comments | https://api.github.com/repos/huggingface/datasets/issues/2826/events | https://github.com/huggingface/datasets/issues/2826 | 976,974,254 | MDU6SXNzdWU5NzY5NzQyNTQ= | 2,826 | Add a Text Classification dataset: KanHope | {
"login": "adeepH",
"id": 46108405,
"node_id": "MDQ6VXNlcjQ2MTA4NDA1",
"avatar_url": "https://avatars.githubusercontent.com/u/46108405?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/adeepH",
"html_url": "https://github.com/adeepH",
"followers_url": "https://api.github.com/users/adeepH/fo... | [
{
"id": 2067376369,
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request",
"name": "dataset request",
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset"
}
] | closed | false | null | [] | null | [
"Hi ! In your script it looks like you're trying to load the dataset `bn_hate_speech,`, not KanHope.\r\n\r\nMoreover the error `KeyError: ' '` means that you have a feature of type ClassLabel, but for a certain example of the dataset, it looks like the label is empty (it's just a string with a space). Can you make ... | 1,629,721,318,000 | 1,633,111,619,000 | 1,633,111,619,000 | CONTRIBUTOR | null | ## Adding a Dataset
- **Name:** *KanHope*
- **Description:** *A code-mixed English-Kannada dataset for Hope speech detection*
- **Paper:** *https://arxiv.org/abs/2108.04616* (I am the author of the paper}
- **Author:** *[AdeepH](https://github.com/adeepH)*
- **Data:** *https://github.com/adeepH/KanHope/tree/main/d... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2826/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2826/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2825 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2825/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2825/comments | https://api.github.com/repos/huggingface/datasets/issues/2825/events | https://github.com/huggingface/datasets/issues/2825 | 976,584,926 | MDU6SXNzdWU5NzY1ODQ5MjY= | 2,825 | The datasets.map function does not load cached dataset after moving python script | {
"login": "hobbitlzy",
"id": 35392624,
"node_id": "MDQ6VXNlcjM1MzkyNjI0",
"avatar_url": "https://avatars.githubusercontent.com/u/35392624?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hobbitlzy",
"html_url": "https://github.com/hobbitlzy",
"followers_url": "https://api.github.com/users/... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"This also happened to me on COLAB.\r\nDetails:\r\nI ran the `run_mlm.py` in two different notebooks. \r\nIn the first notebook, I do tokenization since I can get 4 CPU cores without any GPUs, and save the cache into a folder which I copy to drive.\r\nIn the second notebook, I copy the cache folder from drive and r... | 1,629,689,017,000 | 1,630,415,681,000 | 1,630,415,616,000 | NONE | null | ## Describe the bug
The datasets.map function caches the processed data to a certain directory. When the map function is called another time with totally the same parameters, the cached data are supposed to be reloaded instead of re-processing. However, it doesn't reuse cached data sometimes. I use the common data pro... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2825/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2825/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2824 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2824/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2824/comments | https://api.github.com/repos/huggingface/datasets/issues/2824/events | https://github.com/huggingface/datasets/pull/2824 | 976,394,721 | MDExOlB1bGxSZXF1ZXN0NzE3MzIyMzY5 | 2,824 | Fix defaults in cache_dir docstring in load.py | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [] | 1,629,643,717,000 | 1,629,984,212,000 | 1,629,978,916,000 | CONTRIBUTOR | null | Fix defaults in the `cache_dir` docstring. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2824/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2824/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2824",
"html_url": "https://github.com/huggingface/datasets/pull/2824",
"diff_url": "https://github.com/huggingface/datasets/pull/2824.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2824.patch",
"merged_at": 1629978916000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2823 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2823/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2823/comments | https://api.github.com/repos/huggingface/datasets/issues/2823/events | https://github.com/huggingface/datasets/issues/2823 | 976,135,355 | MDU6SXNzdWU5NzYxMzUzNTU= | 2,823 | HF_DATASETS_CACHE variable in Windows | {
"login": "rp2839",
"id": 8453798,
"node_id": "MDQ6VXNlcjg0NTM3OTg=",
"avatar_url": "https://avatars.githubusercontent.com/u/8453798?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rp2839",
"html_url": "https://github.com/rp2839",
"followers_url": "https://api.github.com/users/rp2839/foll... | [] | closed | false | null | [] | null | [
"Agh - I'm a muppet. No quote marks are needed.\r\nset HF_DATASETS_CACHE = C:\\Datasets\r\nworks as intended."
] | 1,629,551,864,000 | 1,629,552,011,000 | 1,629,552,011,000 | NONE | null | I can't seem to use a custom Cache directory in Windows. I have tried:
set HF_DATASETS_CACHE = "C:\Datasets"
set HF_DATASETS_CACHE = "C:/Datasets"
set HF_DATASETS_CACHE = "C:\\Datasets"
set HF_DATASETS_CACHE = "r'C:\Datasets'"
set HF_DATASETS_CACHE = "\Datasets"
set HF_DATASETS_CACHE = "/Datasets"
In each in... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2823/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2823/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2822 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2822/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2822/comments | https://api.github.com/repos/huggingface/datasets/issues/2822/events | https://github.com/huggingface/datasets/pull/2822 | 975,744,463 | MDExOlB1bGxSZXF1ZXN0NzE2ODUxMTAy | 2,822 | Add url prefix convention for many compression formats | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"Thanks for the feedback :) I will also complete the documentation to explain this convention",
"I just added some documentation about how streaming works with chained URLs.\r\n\r\nI will also add some docs about how to use chained URLs directly in `load_dataset` in #2662, since #2662 does change the documentatio... | 1,629,475,883,000 | 1,629,734,356,000 | 1,629,734,354,000 | MEMBER | null | ## Intro
When doing dataset streaming, the uncompression of compressed files is done on the fly using `fsspec`.
In particular, the download manager method `download_and_extract` doesn't return a path to the local download and extracted file, but instead a chained URL so that the uncompression can be done when the... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2822/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2822/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2822",
"html_url": "https://github.com/huggingface/datasets/pull/2822",
"diff_url": "https://github.com/huggingface/datasets/pull/2822.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2822.patch",
"merged_at": 1629734354000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2821 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2821/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2821/comments | https://api.github.com/repos/huggingface/datasets/issues/2821/events | https://github.com/huggingface/datasets/issues/2821 | 975,556,032 | MDU6SXNzdWU5NzU1NTYwMzI= | 2,821 | Cannot load linnaeus dataset | {
"login": "NielsRogge",
"id": 48327001,
"node_id": "MDQ6VXNlcjQ4MzI3MDAx",
"avatar_url": "https://avatars.githubusercontent.com/u/48327001?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/NielsRogge",
"html_url": "https://github.com/NielsRogge",
"followers_url": "https://api.github.com/use... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Thanks for reporting ! #2852 fixed this error\r\n\r\nWe'll do a new release of `datasets` soon :)"
] | 1,629,461,715,000 | 1,630,415,582,000 | 1,630,415,529,000 | NONE | null | ## Describe the bug
The [linnaeus](https://huggingface.co/datasets/linnaeus) dataset cannot be loaded. To reproduce:
```
from datasets import load_dataset
datasets = load_dataset("linnaeus")
```
This results in:
```
Downloading and preparing dataset linnaeus/linnaeus (download: 17.36 MiB, generated: 8.74 MiB,... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2821/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2821/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2820 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2820/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2820/comments | https://api.github.com/repos/huggingface/datasets/issues/2820/events | https://github.com/huggingface/datasets/issues/2820 | 975,210,712 | MDU6SXNzdWU5NzUyMTA3MTI= | 2,820 | Downloading “reddit” dataset keeps timing out. | {
"login": "smeyerhot",
"id": 43877130,
"node_id": "MDQ6VXNlcjQzODc3MTMw",
"avatar_url": "https://avatars.githubusercontent.com/u/43877130?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/smeyerhot",
"html_url": "https://github.com/smeyerhot",
"followers_url": "https://api.github.com/users/... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"```\r\nUsing custom data configuration default\r\nDownloading and preparing dataset reddit/default (download: 2.93 GiB, generated: 17.64 GiB, post-processed: Unknown size, total: 20.57 GiB) to /Volumes/My Passport for Mac/og-chat-data/reddit/default/1.0.0/98ba5abea674d3178f7588aa6518a5510dc0c6fa8176d9653a3546d5afc... | 1,629,427,956,000 | 1,631,112,722,000 | 1,631,112,722,000 | NONE | null | ## Describe the bug
A clear and concise description of what the bug is.
Everytime I try and download the reddit dataset it times out before finishing and I have to try again.
There is some timeout error that I will post once it happens again.
## Steps to reproduce the bug
```python
from datasets import load_d... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2820/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2820/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2819 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2819/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2819/comments | https://api.github.com/repos/huggingface/datasets/issues/2819/events | https://github.com/huggingface/datasets/pull/2819 | 974,683,155 | MDExOlB1bGxSZXF1ZXN0NzE1OTUyMjE1 | 2,819 | Added XL-Sum dataset | {
"login": "abhik1505040",
"id": 49608995,
"node_id": "MDQ6VXNlcjQ5NjA4OTk1",
"avatar_url": "https://avatars.githubusercontent.com/u/49608995?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/abhik1505040",
"html_url": "https://github.com/abhik1505040",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | [
"Thanks for adding this one ! I just did some minor changes and set the timeout back to 100sec instead of 1000",
"The CI failure is unrelated to this PR - let me take a look",
"> Thanks for adding this one! I just did some minor changes and set the timeout back to 100sec instead of 1000\r\n\r\nThank you for upd... | 1,629,380,865,000 | 1,632,903,224,000 | 1,632,419,345,000 | NONE | null | Added XL-Sum dataset published in ACL-IJCNLP 2021. (https://aclanthology.org/2021.findings-acl.413/). The default timeout values in `src/datasets/utils/file_utls.py` were increased to enable downloading from the original google drive links. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2819/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2819/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2819",
"html_url": "https://github.com/huggingface/datasets/pull/2819",
"diff_url": "https://github.com/huggingface/datasets/pull/2819.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2819.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2817 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2817/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2817/comments | https://api.github.com/repos/huggingface/datasets/issues/2817/events | https://github.com/huggingface/datasets/pull/2817 | 974,486,051 | MDExOlB1bGxSZXF1ZXN0NzE1NzgzMDQ3 | 2,817 | Rename The Pile subsets | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"Sounds good. Should we also have a “the_pile” dataset with the subsets as configuration?",
"I think the main `the_pile` datasets will be the one that is the mix of all the subsets: https://the-eye.eu/public/AI/pile/\r\n\r\nWe can also add configurations for each subset, and even allow users to specify the subset... | 1,629,366,982,000 | 1,629,735,850,000 | 1,629,735,849,000 | MEMBER | null | After discussing with @yjernite we think it's better to have the subsets of The Pile explicitly have "the_pile" in their names.
I'm doing the changes for the subsets that @richarddwang added:
- [x] books3 -> the_pile_books3 https://github.com/huggingface/datasets/pull/2801
- [x] stack_exchange -> the_pile_stack_ex... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2817/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2817/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2817",
"html_url": "https://github.com/huggingface/datasets/pull/2817",
"diff_url": "https://github.com/huggingface/datasets/pull/2817.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2817.patch",
"merged_at": 1629735849000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2815 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2815/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2815/comments | https://api.github.com/repos/huggingface/datasets/issues/2815/events | https://github.com/huggingface/datasets/pull/2815 | 973,862,024 | MDExOlB1bGxSZXF1ZXN0NzE1MjUxNDQ5 | 2,815 | Tiny typo fixes of "fo" -> "of" | {
"login": "aronszanto",
"id": 9934829,
"node_id": "MDQ6VXNlcjk5MzQ4Mjk=",
"avatar_url": "https://avatars.githubusercontent.com/u/9934829?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/aronszanto",
"html_url": "https://github.com/aronszanto",
"followers_url": "https://api.github.com/users... | [] | closed | false | null | [] | null | [] | 1,629,304,571,000 | 1,629,360,182,000 | 1,629,360,182,000 | CONTRIBUTOR | null | Noticed a few of these when reading docs- feel free to ignore the PR and just fix on some main contributor branch if more helpful. Thanks for the great library! :) | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2815/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2815/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2815",
"html_url": "https://github.com/huggingface/datasets/pull/2815",
"diff_url": "https://github.com/huggingface/datasets/pull/2815.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2815.patch",
"merged_at": 1629360182000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2814 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2814/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2814/comments | https://api.github.com/repos/huggingface/datasets/issues/2814/events | https://github.com/huggingface/datasets/pull/2814 | 973,632,645 | MDExOlB1bGxSZXF1ZXN0NzE1MDUwODc4 | 2,814 | Bump tqdm version | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [] | 1,629,291,089,000 | 1,629,294,251,000 | 1,629,293,990,000 | CONTRIBUTOR | null | The recently released tqdm 4.62.1 includes a fix for PermissionError on Windows (submitted by me in https://github.com/tqdm/tqdm/pull/1207), which means we can remove expensive `gc.collect` calls by bumping tqdm to that version. This PR does exactly that and, additionally, fixes a `disable_tqdm` definition that would p... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2814/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2814/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2814",
"html_url": "https://github.com/huggingface/datasets/pull/2814",
"diff_url": "https://github.com/huggingface/datasets/pull/2814.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2814.patch",
"merged_at": 1629293989000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2813 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2813/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2813/comments | https://api.github.com/repos/huggingface/datasets/issues/2813/events | https://github.com/huggingface/datasets/issues/2813 | 973,470,580 | MDU6SXNzdWU5NzM0NzA1ODA= | 2,813 | Remove compression from xopen | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 2067400324,
"node_id": "MDU6TGFiZWwyMDY3NDAwMzI0",
"url": "https://api.github.com/repos/huggingface/datasets/labels/generic%20discussion",
"name": "generic discussion",
"color": "c5def5",
"default": false,
"description": "Generic discussion on the library"
}
] | closed | false | null | [] | null | [
"After discussing with @lhoestq, a reasonable alternative:\r\n- `download_manager.extract(urlpath)` adds prefixes to `urlpath` in the same way as `fsspec` does for protocols, but we implement custom prefixes for all compression formats: \r\n `bz2::http://domain.org/filename.bz2`\r\n- `xopen` parses the `urlpath` a... | 1,629,279,359,000 | 1,629,734,354,000 | 1,629,734,354,000 | MEMBER | null | We implemented support for streaming with 2 requirements:
- transparent use for the end user: just needs to pass the parameter `streaming=True`
- no additional work for the contributors: previous loading scripts should also work in streaming mode with no (or minor) changes; and new loading scripts should not involve ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2813/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2813/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2811 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2811/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2811/comments | https://api.github.com/repos/huggingface/datasets/issues/2811/events | https://github.com/huggingface/datasets/pull/2811 | 972,522,480 | MDExOlB1bGxSZXF1ZXN0NzE0MTAzNDIy | 2,811 | Fix stream oscar | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"One additional note: if we can try to not change the code of oscar.py too often, I'm sure users that have it in their cache directory will be happy to not have to redownload it every time they update the library ;)\r\n\r\n(since changing the code changes the cache directory of the dataset)",
"I don't think this ... | 1,629,195,059,000 | 1,629,973,575,000 | 1,629,973,574,000 | MEMBER | null | Previously, an additional `open` was added to oscar to make it stream-compatible: 587bbb94e891b22863b312b99696e32708c379f4.
This was argued that might be problematic: https://github.com/huggingface/datasets/pull/2786#discussion_r690045921
This PR:
- removes that additional `open`
- patches `gzip.open` with `xop... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2811/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2811/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2811",
"html_url": "https://github.com/huggingface/datasets/pull/2811",
"diff_url": "https://github.com/huggingface/datasets/pull/2811.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2811.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2809 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2809/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2809/comments | https://api.github.com/repos/huggingface/datasets/issues/2809/events | https://github.com/huggingface/datasets/pull/2809 | 971,902,613 | MDExOlB1bGxSZXF1ZXN0NzEzNTc2Njcz | 2,809 | Add Beans Dataset | {
"login": "nateraw",
"id": 32437151,
"node_id": "MDQ6VXNlcjMyNDM3MTUx",
"avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nateraw",
"html_url": "https://github.com/nateraw",
"followers_url": "https://api.github.com/users/natera... | [] | closed | false | null | [] | null | [] | 1,629,130,953,000 | 1,629,978,147,000 | 1,629,978,147,000 | CONTRIBUTOR | null | Adds the [beans](https://github.com/AI-Lab-Makerere/ibean/) image classification dataset. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2809/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2809/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2809",
"html_url": "https://github.com/huggingface/datasets/pull/2809",
"diff_url": "https://github.com/huggingface/datasets/pull/2809.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2809.patch",
"merged_at": 1629978147000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2807 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2807/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2807/comments | https://api.github.com/repos/huggingface/datasets/issues/2807/events | https://github.com/huggingface/datasets/pull/2807 | 971,849,863 | MDExOlB1bGxSZXF1ZXN0NzEzNTMxNjIw | 2,807 | Add cats_vs_dogs dataset | {
"login": "nateraw",
"id": 32437151,
"node_id": "MDQ6VXNlcjMyNDM3MTUx",
"avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nateraw",
"html_url": "https://github.com/nateraw",
"followers_url": "https://api.github.com/users/natera... | [] | closed | false | null | [] | null | [] | 1,629,127,271,000 | 1,630,341,325,000 | 1,630,341,324,000 | CONTRIBUTOR | null | Adds Microsoft's [Cats vs. Dogs](https://www.microsoft.com/en-us/download/details.aspx?id=54765) dataset. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2807/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2807/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2807",
"html_url": "https://github.com/huggingface/datasets/pull/2807",
"diff_url": "https://github.com/huggingface/datasets/pull/2807.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2807.patch",
"merged_at": 1630341324000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2806 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2806/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2806/comments | https://api.github.com/repos/huggingface/datasets/issues/2806/events | https://github.com/huggingface/datasets/pull/2806 | 971,625,449 | MDExOlB1bGxSZXF1ZXN0NzEzMzM5NDUw | 2,806 | Fix streaming tar files from canonical datasets | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"In case it's relevant for this PR, I'm finding that I cannot stream the `bookcorpus` dataset (using the `master` branch of `datasets`), which is a `.tar.bz2` file:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nbooks_dataset_streamed = load_dataset(\"bookcorpus\", split=\"train\", streaming=True)\r\n... | 1,629,112,228,000 | 1,634,115,843,000 | 1,634,115,842,000 | MEMBER | null | Previous PR #2800 implemented support to stream remote tar files when passing the parameter `data_files`: they required a glob string `"*"`.
However, this glob string creates an error when streaming canonical datasets (with a `join` after the `open`).
This PR fixes this issue and allows streaming tar files both f... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2806/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 1,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2806/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2806",
"html_url": "https://github.com/huggingface/datasets/pull/2806",
"diff_url": "https://github.com/huggingface/datasets/pull/2806.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2806.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2805 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2805/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2805/comments | https://api.github.com/repos/huggingface/datasets/issues/2805/events | https://github.com/huggingface/datasets/pull/2805 | 971,436,456 | MDExOlB1bGxSZXF1ZXN0NzEzMTc3MTI4 | 2,805 | Fix streaming zip files from canonical datasets | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,629,097,900,000 | 1,629,110,040,000 | 1,629,110,040,000 | MEMBER | null | Previous PR #2798 fixed streaming remote zip files when passing the parameter `data_files`.
However, that broke streaming zip files used in canonical `datasets` scripts, which normally have a subsequent `join()` (patched with `xjoin()`) after the `StreamingDownloadManager.download_and_extract()` is called.
This P... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2805/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2805/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2805",
"html_url": "https://github.com/huggingface/datasets/pull/2805",
"diff_url": "https://github.com/huggingface/datasets/pull/2805.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2805.patch",
"merged_at": 1629110040000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2804 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2804/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2804/comments | https://api.github.com/repos/huggingface/datasets/issues/2804/events | https://github.com/huggingface/datasets/pull/2804 | 971,353,437 | MDExOlB1bGxSZXF1ZXN0NzEzMTA2NTMw | 2,804 | Add Food-101 | {
"login": "nateraw",
"id": 32437151,
"node_id": "MDQ6VXNlcjMyNDM3MTUx",
"avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nateraw",
"html_url": "https://github.com/nateraw",
"followers_url": "https://api.github.com/users/natera... | [] | closed | false | null | [] | null | [] | 1,629,087,975,000 | 1,629,469,893,000 | 1,629,377,286,000 | CONTRIBUTOR | null | Adds image classification dataset [Food-101](https://data.vision.ee.ethz.ch/cvl/datasets_extra/food-101/). | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2804/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2804/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2804",
"html_url": "https://github.com/huggingface/datasets/pull/2804",
"diff_url": "https://github.com/huggingface/datasets/pull/2804.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2804.patch",
"merged_at": 1629377286000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2803 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2803/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2803/comments | https://api.github.com/repos/huggingface/datasets/issues/2803/events | https://github.com/huggingface/datasets/pull/2803 | 970,858,928 | MDExOlB1bGxSZXF1ZXN0NzEyNzQxODMz | 2,803 | add stack exchange | {
"login": "richarddwang",
"id": 17963619,
"node_id": "MDQ6VXNlcjE3OTYzNjE5",
"avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/richarddwang",
"html_url": "https://github.com/richarddwang",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | [
"Hi ! Merging this one since it's all good :)\r\n\r\nHowever I think it would also be better to actually rename it `the_pile_stack_exchange` to make things clearer and to avoid name collisions in the future. I would like to do the same for `books3` as well.\r\n\r\nIf you don't mind I'll open a PR to do the renaming... | 1,628,928,662,000 | 1,629,367,653,000 | 1,629,360,458,000 | CONTRIBUTOR | null | stack exchange is part of EleutherAI/The Pile, but AFAIK, The Pile dataset blend all sub datasets together thus we are not able to use just one of its sub dataset from The Pile data. So I create an independent dataset using The Pile preliminary components.
I also change default `timeout` to 100 seconds instead of 10... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2803/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2803/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2803",
"html_url": "https://github.com/huggingface/datasets/pull/2803",
"diff_url": "https://github.com/huggingface/datasets/pull/2803.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2803.patch",
"merged_at": 1629360458000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2802 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2802/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2802/comments | https://api.github.com/repos/huggingface/datasets/issues/2802/events | https://github.com/huggingface/datasets/pull/2802 | 970,848,302 | MDExOlB1bGxSZXF1ZXN0NzEyNzM0MTc3 | 2,802 | add openwebtext2 | {
"login": "richarddwang",
"id": 17963619,
"node_id": "MDQ6VXNlcjE3OTYzNjE5",
"avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/richarddwang",
"html_url": "https://github.com/richarddwang",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | [
"It seems we need to `pip install jsonlines` to pass the checks ?",
"Hi ! Do you really need `jsonlines` ? I think it simply uses `json.loads` under the hood.\r\n\r\nCurrently the test are failing because `jsonlines` is not part of the extra requirements `TESTS_REQUIRE` in setup.py\r\n\r\nSo either you can replac... | 1,628,924,943,000 | 1,629,727,574,000 | 1,629,727,574,000 | CONTRIBUTOR | null | openwebtext2 is part of EleutherAI/The Pile, but AFAIK, The Pile dataset blend all sub datasets together thus we are not able to use just one of its sub dataset from The Pile data. So I create an independent dataset using The Pile preliminary components.
When I was creating dataset card. I found there is room for cr... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2802/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2802/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2802",
"html_url": "https://github.com/huggingface/datasets/pull/2802",
"diff_url": "https://github.com/huggingface/datasets/pull/2802.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2802.patch",
"merged_at": 1629727574000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2801 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2801/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2801/comments | https://api.github.com/repos/huggingface/datasets/issues/2801/events | https://github.com/huggingface/datasets/pull/2801 | 970,844,617 | MDExOlB1bGxSZXF1ZXN0NzEyNzMwODEz | 2,801 | add books3 | {
"login": "richarddwang",
"id": 17963619,
"node_id": "MDQ6VXNlcjE3OTYzNjE5",
"avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/richarddwang",
"html_url": "https://github.com/richarddwang",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | [
"> When I was creating dataset card. I found there is room for creating / editing dataset card. I've made it an issue. #2797\r\n\r\nThanks for the message, we'll definitely improve this\r\n\r\n> Also I am wondering whether the import of The Pile dataset is actively undertaken (because I may need it recently)? #1675... | 1,628,924,665,000 | 1,629,391,389,000 | 1,629,301,019,000 | CONTRIBUTOR | null | books3 is part of EleutherAI/The Pile, but AFAIK, The Pile dataset blend all sub datasets together thus we are not able to use just one of its sub dataset from The Pile data. So I create an independent dataset using The Pile preliminary components.
When I was creating dataset card. I found there is room for creating... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2801/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2801/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2801",
"html_url": "https://github.com/huggingface/datasets/pull/2801",
"diff_url": "https://github.com/huggingface/datasets/pull/2801.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2801.patch",
"merged_at": 1629301019000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2800 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2800/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2800/comments | https://api.github.com/repos/huggingface/datasets/issues/2800/events | https://github.com/huggingface/datasets/pull/2800 | 970,819,988 | MDExOlB1bGxSZXF1ZXN0NzEyNzExNTcx | 2,800 | Support streaming tar files | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"Hi ! Why do we need the custom `readline` for exactly ? feel free to add a comment to say why it's needed"
] | 1,628,916,017,000 | 1,629,972,150,000 | 1,628,916,957,000 | MEMBER | null | This PR adds support to stream tar files by using the `fsspec` tar protocol.
It also uses the custom `readline` implemented in PR #2786.
The corresponding test is implemented in PR #2786. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2800/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2800/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2800",
"html_url": "https://github.com/huggingface/datasets/pull/2800",
"diff_url": "https://github.com/huggingface/datasets/pull/2800.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2800.patch",
"merged_at": 1628916957000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2798 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2798/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2798/comments | https://api.github.com/repos/huggingface/datasets/issues/2798/events | https://github.com/huggingface/datasets/pull/2798 | 970,493,126 | MDExOlB1bGxSZXF1ZXN0NzEyNDM3ODc2 | 2,798 | Fix streaming zip files | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"Hi ! I don't fully understand this change @albertvillanova \r\nThe `_extract` method used to return the compound URL that points to the root of the inside of the archive.\r\nThis way users can use the usual os.path.join or other functions to point to the relevant files. I don't see why you're using a glob pattern ... | 1,628,867,821,000 | 1,629,123,410,000 | 1,628,869,108,000 | MEMBER | null | Currently, streaming remote zip data files gives `FileNotFoundError` message:
```python
data_files = f"https://huggingface.co/datasets/albertvillanova/datasets-tests-compression/resolve/main/sample.zip"
ds = load_dataset("json", split="train", data_files=data_files, streaming=True)
next(iter(ds))
```
This PR fi... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2798/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2798/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2798",
"html_url": "https://github.com/huggingface/datasets/pull/2798",
"diff_url": "https://github.com/huggingface/datasets/pull/2798.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2798.patch",
"merged_at": 1628869108000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2796 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2796/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2796/comments | https://api.github.com/repos/huggingface/datasets/issues/2796/events | https://github.com/huggingface/datasets/pull/2796 | 970,235,846 | MDExOlB1bGxSZXF1ZXN0NzEyMjE1ODM2 | 2,796 | add cedr dataset | {
"login": "naumov-al",
"id": 22640075,
"node_id": "MDQ6VXNlcjIyNjQwMDc1",
"avatar_url": "https://avatars.githubusercontent.com/u/22640075?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/naumov-al",
"html_url": "https://github.com/naumov-al",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | [
"> Hi ! Thanks a lot for adding this one :)\r\n> \r\n> Good job with the dataset card and the dataset script !\r\n> \r\n> I left a few suggestions\r\n\r\nThank you very much for your helpful suggestions. I have tried to carry them all out."
] | 1,628,847,455,000 | 1,630,080,096,000 | 1,630,080,096,000 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2796/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2796/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2796",
"html_url": "https://github.com/huggingface/datasets/pull/2796",
"diff_url": "https://github.com/huggingface/datasets/pull/2796.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2796.patch",
"merged_at": 1630080095000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2793 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2793/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2793/comments | https://api.github.com/repos/huggingface/datasets/issues/2793/events | https://github.com/huggingface/datasets/pull/2793 | 968,967,773 | MDExOlB1bGxSZXF1ZXN0NzExMDQ4NDY2 | 2,793 | Fix type hint for data_files | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,628,779,357,000 | 1,628,782,529,000 | 1,628,782,529,000 | MEMBER | null | Fix type hint for `data_files` in signatures and docstrings. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2793/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2793/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2793",
"html_url": "https://github.com/huggingface/datasets/pull/2793",
"diff_url": "https://github.com/huggingface/datasets/pull/2793.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2793.patch",
"merged_at": 1628782529000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/2792 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2792/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2792/comments | https://api.github.com/repos/huggingface/datasets/issues/2792/events | https://github.com/huggingface/datasets/pull/2792 | 968,650,274 | MDExOlB1bGxSZXF1ZXN0NzEwNzUyMjc0 | 2,792 | Update: GooAQ - add train/val/test splits | {
"login": "bhavitvyamalik",
"id": 19718818,
"node_id": "MDQ6VXNlcjE5NzE4ODE4",
"avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bhavitvyamalik",
"html_url": "https://github.com/bhavitvyamalik",
"followers_url": "https://api.gi... | [] | closed | false | null | [] | null | [
"@albertvillanova my tests are failing here:\r\n```\r\ndataset_name = 'gooaq'\r\n\r\n def test_load_dataset(self, dataset_name):\r\n configs = self.dataset_tester.load_all_configs(dataset_name, is_local=True)[:1]\r\n> self.dataset_tester.check_load_dataset(dataset_name, configs, is_local=True, use_l... | 1,628,768,418,000 | 1,630,079,925,000 | 1,630,079,894,000 | CONTRIBUTOR | null | [GooAQ](https://github.com/allenai/gooaq) dataset was recently updated after splits were added for the same. This PR contains new updated GooAQ with train/val/test splits and updated README as well. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/2792/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/2792/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2792",
"html_url": "https://github.com/huggingface/datasets/pull/2792",
"diff_url": "https://github.com/huggingface/datasets/pull/2792.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/2792.patch",
"merged_at": 1630079894000
} | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.