url stringlengths 58 61 | repository_url stringclasses 1
value | labels_url stringlengths 72 75 | comments_url stringlengths 67 70 | events_url stringlengths 65 68 | html_url stringlengths 46 51 | id int64 600M 2.05B | node_id stringlengths 18 32 | number int64 2 6.51k | title stringlengths 1 290 | user dict | labels listlengths 0 4 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 4 | milestone dict | comments listlengths 0 30 | created_at timestamp[ns, tz=UTC] | updated_at timestamp[ns, tz=UTC] | closed_at timestamp[ns, tz=UTC] | author_association stringclasses 3
values | active_lock_reason float64 | draft float64 0 1 ⌀ | pull_request dict | body stringlengths 0 228k ⌀ | reactions dict | timeline_url stringlengths 67 70 | performed_via_github_app float64 | state_reason stringclasses 3
values | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/2827 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2827/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2827/comments | https://api.github.com/repos/huggingface/datasets/issues/2827/events | https://github.com/huggingface/datasets/pull/2827 | 976,976,552 | MDExOlB1bGxSZXF1ZXN0NzE3Nzg3MjEw | 2,827 | add a text classification dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/46108405?v=4",
"events_url": "https://api.github.com/users/adeepH/events{/privacy}",
"followers_url": "https://api.github.com/users/adeepH/followers",
"following_url": "https://api.github.com/users/adeepH/following{/other_user}",
"gists_url": "https://a... | [] | closed | false | null | [] | null | [] | 2021-08-23T12:24:41Z | 2021-08-23T15:51:18Z | 2021-08-23T15:51:18Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2827.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2827",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/2827.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2827"
} | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2827/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2827/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3136 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3136/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3136/comments | https://api.github.com/repos/huggingface/datasets/issues/3136/events | https://github.com/huggingface/datasets/pull/3136 | 1,033,360,396 | PR_kwDODunzps4tieFi | 3,136 | Fix script of Arabic Billion Words dataset to return all data | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [] | 2021-10-22T09:14:24Z | 2021-10-22T13:28:41Z | 2021-10-22T13:28:40Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3136.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3136",
"merged_at": "2021-10-22T13:28:39Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3136.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | The script has a bug and only parses and generates a portion of the entire dataset.
This PR fixes the loading script so that is properly parses the entire dataset.
Current implementation generates the same number of examples as reported in the [original paper](https://arxiv.org/abs/1611.04033) for all configurat... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3136/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3136/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4034 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4034/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4034/comments | https://api.github.com/repos/huggingface/datasets/issues/4034/events | https://github.com/huggingface/datasets/pull/4034 | 1,183,033,285 | PR_kwDODunzps41IpN1 | 4,034 | Fix null checksum in xcopa dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [] | 2022-03-28T07:48:14Z | 2022-03-28T08:06:14Z | 2022-03-28T08:06:14Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4034.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4034",
"merged_at": "2022-03-28T08:06:14Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4034.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4034/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4034/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5827 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5827/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5827/comments | https://api.github.com/repos/huggingface/datasets/issues/5827/events | https://github.com/huggingface/datasets/issues/5827 | 1,698,891,246 | I_kwDODunzps5lQwXu | 5,827 | load json dataset interrupt when dtype cast problem occured | {
"avatar_url": "https://avatars.githubusercontent.com/u/46060451?v=4",
"events_url": "https://api.github.com/users/1014661165/events{/privacy}",
"followers_url": "https://api.github.com/users/1014661165/followers",
"following_url": "https://api.github.com/users/1014661165/following{/other_user}",
"gists_url"... | [] | open | false | null | [] | null | [
"Indeed the JSON dataset builder raises an error when it encounters an unexpected type.\r\n\r\nThere's an old PR open to add away to ignore such elements though, if it can help: https://github.com/huggingface/datasets/pull/2838"
] | 2023-05-07T04:52:09Z | 2023-05-10T12:32:28Z | null | NONE | null | null | null | ### Describe the bug
i have a json like this:
[
{"id": 1, "name": 1},
{"id": 2, "name": "Nan"},
{"id": 3, "name": 3},
....
]
,which have several problematic rows data like row 2, then i load it with datasets.load_dataset('json', data_files=['xx.json'], split='train'), it will report like this:
... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5827/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5827/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/5178 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5178/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5178/comments | https://api.github.com/repos/huggingface/datasets/issues/5178/events | https://github.com/huggingface/datasets/issues/5178 | 1,430,800,810 | I_kwDODunzps5VSEmq | 5,178 | Unable to download the Chinese `wikipedia`, the dumpstatus.json not found! | {
"avatar_url": "https://avatars.githubusercontent.com/u/37113676?v=4",
"events_url": "https://api.github.com/users/beyondguo/events{/privacy}",
"followers_url": "https://api.github.com/users/beyondguo/followers",
"following_url": "https://api.github.com/users/beyondguo/following{/other_user}",
"gists_url": "... | [] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/o... | null | [
"In the dumps page of the wiki (https://dumps.wikimedia.org/zhwiki/), I found the following dumps:\r\n```\r\nIndex of /zhwiki/\r\n[../](https://dumps.wikimedia.org/)\r\n[20220701/](https://dumps.wikimedia.org/zhwiki/20220701/) 21-Aug-2022 01:48 -\r\n[202207... | 2022-11-01T03:17:55Z | 2022-11-02T08:27:15Z | 2022-11-02T08:24:29Z | NONE | null | null | null | ### Describe the bug
I tried:
`data = load_dataset('wikipedia', '20220301.zh', beam_runner='DirectRunner')`
and
`data = load_dataset("wikipedia", language="zh", date="20220301", beam_runner='DirectRunner')`
but both got:
`FileNotFoundError: Couldn't find file at https://dumps.wikimedia.org/zhwiki/20220301/dumpsta... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5178/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5178/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/924 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/924/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/924/comments | https://api.github.com/repos/huggingface/datasets/issues/924/events | https://github.com/huggingface/datasets/pull/924 | 753,631,951 | MDExOlB1bGxSZXF1ZXN0NTI5NjcyMzgw | 924 | Add DART | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"LGTM!"
] | 2020-11-30T16:42:37Z | 2020-12-02T03:13:42Z | 2020-12-02T03:13:41Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/924.diff",
"html_url": "https://github.com/huggingface/datasets/pull/924",
"merged_at": "2020-12-02T03:13:41Z",
"patch_url": "https://github.com/huggingface/datasets/pull/924.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/924... | - **Name:** *DART*
- **Description:** *DART is a large dataset for open-domain structured data record to text generation.*
- **Paper:** *https://arxiv.org/abs/2007.02871*
- **Data:** *https://github.com/Yale-LILY/dart#leaderboard*
### Checkbox
- [x] Create the dataset script `/datasets/my_dataset/my_dataset.py... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/924/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/924/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1874 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1874/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1874/comments | https://api.github.com/repos/huggingface/datasets/issues/1874/events | https://github.com/huggingface/datasets/pull/1874 | 807,786,094 | MDExOlB1bGxSZXF1ZXN0NTcyOTYzMjAy | 1,874 | Adding Europarl Bilingual dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/23355969?v=4",
"events_url": "https://api.github.com/users/lucadiliello/events{/privacy}",
"followers_url": "https://api.github.com/users/lucadiliello/followers",
"following_url": "https://api.github.com/users/lucadiliello/following{/other_user}",
"gist... | [] | closed | false | null | [] | null | [
"is there a way to check errors without subscribing to CircleCI? Because they want access to private repositories when logging.",
"I think you need to be logged in to check the errors unfortunately. Feel free to create an account with bitbucket maybe if you don't want it to access your private github repos",
"I... | 2021-02-13T17:02:04Z | 2021-03-04T10:38:22Z | 2021-03-04T10:38:22Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1874.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1874",
"merged_at": "2021-03-04T10:38:22Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1874.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Implementation of Europarl bilingual dataset from described [here](https://opus.nlpl.eu/Europarl.php).
This dataset allows to use every language pair detailed in the original dataset. The loading script manages also the small errors contained in the original dataset (in very rare cases (1 over 10M) there are some ke... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1874/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1874/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2128 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2128/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2128/comments | https://api.github.com/repos/huggingface/datasets/issues/2128/events | https://github.com/huggingface/datasets/issues/2128 | 843,023,910 | MDU6SXNzdWU4NDMwMjM5MTA= | 2,128 | Dialogue action slot name and value are reversed in MultiWoZ 2.2 | {
"avatar_url": "https://avatars.githubusercontent.com/u/31605305?v=4",
"events_url": "https://api.github.com/users/adamlin120/events{/privacy}",
"followers_url": "https://api.github.com/users/adamlin120/followers",
"following_url": "https://api.github.com/users/adamlin120/following{/other_user}",
"gists_url"... | [
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] | closed | false | null | [] | null | [
"Hi\r\nGood catch ! Thanks for reporting\r\n\r\nIf you are interested in contributing, feel free to open a PR to fix this :) "
] | 2021-03-29T06:34:02Z | 2021-03-31T12:48:01Z | 2021-03-31T12:48:01Z | CONTRIBUTOR | null | null | null | Hi @yjernite, thank you for adding MultiWoZ 2.2 in the huggingface datasets platform. It is beneficial!
I spot an error that the order of Dialogue action slot names and values are reversed.
https://github.com/huggingface/datasets/blob/649b2c469779bc4221e1b6969aa2496d63eb5953/datasets/multi_woz_v22/multi_woz_v22.p... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 1,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2128/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2128/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1243 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1243/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1243/comments | https://api.github.com/repos/huggingface/datasets/issues/1243/events | https://github.com/huggingface/datasets/pull/1243 | 758,378,904 | MDExOlB1bGxSZXF1ZXN0NTMzNTYxNDAx | 1,243 | Add Google Noun Verb Dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/1183441?v=4",
"events_url": "https://api.github.com/users/abhishekkrthakur/events{/privacy}",
"followers_url": "https://api.github.com/users/abhishekkrthakur/followers",
"following_url": "https://api.github.com/users/abhishekkrthakur/following{/other_user... | [
{
"color": "0e8a16",
"default": false,
"description": "Contribution to a dataset script",
"id": 4564477500,
"name": "dataset contribution",
"node_id": "LA_kwDODunzps8AAAABEBBmPA",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution"
}
] | closed | false | null | [] | null | [
"Thanks for your contribution, @abhishekkrthakur. Are you still interested in adding this dataset?\r\n\r\nWe are removing the dataset scripts from this GitHub repo and moving them to the Hugging Face Hub: https://huggingface.co/datasets\r\n\r\nWe would suggest you create this dataset there. Please, feel free to tel... | 2020-12-07T10:26:05Z | 2023-09-24T09:40:54Z | 2022-10-03T09:39:37Z | MEMBER | null | 1 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1243.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1243",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1243.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1243"
} | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1243/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1243/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1877 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1877/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1877/comments | https://api.github.com/repos/huggingface/datasets/issues/1877/events | https://github.com/huggingface/datasets/issues/1877 | 808,462,272 | MDU6SXNzdWU4MDg0NjIyNzI= | 1,877 | Allow concatenation of both in-memory and on-disk datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists... | null | [
"I started working on this. My idea is to first add the pyarrow Table wrappers InMemoryTable and MemoryMappedTable that both implement what's necessary regarding copy/pickle. Then have another wrapper that takes the concatenation of InMemoryTable/MemoryMappedTable objects.\r\n\r\nWhat's important here is that conca... | 2021-02-15T11:39:46Z | 2021-03-26T16:51:58Z | 2021-03-26T16:51:58Z | MEMBER | null | null | null | This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).
This assumption is used for pickl... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 1,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1877/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1877/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1552 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1552/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1552/comments | https://api.github.com/repos/huggingface/datasets/issues/1552/events | https://github.com/huggingface/datasets/pull/1552 | 765,664,411 | MDExOlB1bGxSZXF1ZXN0NTM5MDI2MzAx | 1,552 | Added OPUS ParaCrawl | {
"avatar_url": "https://avatars.githubusercontent.com/u/22396042?v=4",
"events_url": "https://api.github.com/users/rkc007/events{/privacy}",
"followers_url": "https://api.github.com/users/rkc007/followers",
"following_url": "https://api.github.com/users/rkc007/following{/other_user}",
"gists_url": "https://a... | [] | closed | false | null | [] | null | [
"@lhoestq I saw some common changes you made on the other PR's (Similar Opus Datasets). I fixed those changes here. Can you please review it once ? \r\nThanks.",
"@rkc007 @lhoestq just noticed a dataset named para_crawl has been added a long time ago: #91.",
"They're not exactly the same so it's ok to have both... | 2020-12-13T21:44:29Z | 2020-12-21T09:50:26Z | 2020-12-21T09:50:25Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1552.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1552",
"merged_at": "2020-12-21T09:50:25Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1552.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Dataset : http://opus.nlpl.eu/ParaCrawl.php | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1552/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1552/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4487 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4487/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4487/comments | https://api.github.com/repos/huggingface/datasets/issues/4487/events | https://github.com/huggingface/datasets/pull/4487 | 1,270,525,163 | PR_kwDODunzps45nm5J | 4,487 | Support streaming UDHR dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-06-14T09:33:33Z | 2022-06-15T05:09:22Z | 2022-06-15T04:59:49Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4487.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4487",
"merged_at": "2022-06-15T04:59:49Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4487.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This PR:
- Adds support for streaming UDHR dataset
- Adds the BCP 47 language code as feature | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4487/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4487/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1089 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1089/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1089/comments | https://api.github.com/repos/huggingface/datasets/issues/1089/events | https://github.com/huggingface/datasets/pull/1089 | 756,823,690 | MDExOlB1bGxSZXF1ZXN0NTMyMzA0MDM2 | 1,089 | add sharc_modified | {
"avatar_url": "https://avatars.githubusercontent.com/u/27137566?v=4",
"events_url": "https://api.github.com/users/patil-suraj/events{/privacy}",
"followers_url": "https://api.github.com/users/patil-suraj/followers",
"following_url": "https://api.github.com/users/patil-suraj/following{/other_user}",
"gists_u... | [] | closed | false | null | [] | null | [] | 2020-12-04T05:49:49Z | 2020-12-04T10:41:30Z | 2020-12-04T10:31:44Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1089.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1089",
"merged_at": "2020-12-04T10:31:44Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1089.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Adding modified ShARC dataset https://github.com/nikhilweee/neural-conv-qa | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1089/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1089/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/365 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/365/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/365/comments | https://api.github.com/repos/huggingface/datasets/issues/365/events | https://github.com/huggingface/datasets/issues/365 | 653,845,964 | MDU6SXNzdWU2NTM4NDU5NjQ= | 365 | How to augment data ? | {
"avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4",
"events_url": "https://api.github.com/users/astariul/events{/privacy}",
"followers_url": "https://api.github.com/users/astariul/followers",
"following_url": "https://api.github.com/users/astariul/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [
"Using batched map is probably the easiest way at the moment.\r\nWhat kind of augmentation would you like to do ?",
"Some samples in the dataset are too long, I want to divide them in several samples.",
"Using batched map is the way to go then.\r\nWe'll make it clearer in the docs that map could be used for aug... | 2020-07-09T07:52:37Z | 2020-07-10T09:12:07Z | 2020-07-10T08:22:15Z | NONE | null | null | null | Is there any clean way to augment data ?
For now my work-around is to use batched map, like this :
```python
def aug(samples):
# Simply copy the existing data to have x2 amount of data
for k, v in samples.items():
samples[k].extend(v)
return samples
dataset = dataset.map(aug, batched=T... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/365/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/365/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/3002 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3002/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3002/comments | https://api.github.com/repos/huggingface/datasets/issues/3002/events | https://github.com/huggingface/datasets/pull/3002 | 1,014,120,524 | PR_kwDODunzps4smCNO | 3,002 | Remove a reference to the open Arrow file when deleting a TF dataset created with to_tf_dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [
"@lhoestq The test passes even without the try/except block!",
"Hey, I'm a little late because I was caught up in the course work, but I double-checked this and it looks great. Thanks for fixing!"
] | 2021-10-02T17:44:09Z | 2021-10-13T11:48:00Z | 2021-10-13T09:03:23Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3002.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3002",
"merged_at": "2021-10-13T09:03:23Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3002.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This [comment](https://github.com/huggingface/datasets/issues/2934#issuecomment-922970919) explains the issue. This PR fixes that with a `weakref` callback, and additionally:
* renames `TensorflowDatasetMixIn` to `TensorflowDatasetMixin` for consistency
* correctly indents `TensorflowDatasetMixin`'s docstring
* repl... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3002/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3002/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/6491 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6491/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6491/comments | https://api.github.com/repos/huggingface/datasets/issues/6491/events | https://github.com/huggingface/datasets/pull/6491 | 2,037,690,643 | PR_kwDODunzps5hyiTY | 6,491 | Fix metrics dead link | {
"avatar_url": "https://avatars.githubusercontent.com/u/45557362?v=4",
"events_url": "https://api.github.com/users/qgallouedec/events{/privacy}",
"followers_url": "https://api.github.com/users/qgallouedec/followers",
"following_url": "https://api.github.com/users/qgallouedec/following{/other_user}",
"gists_u... | [] | open | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6491). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update."
] | 2023-12-12T12:51:49Z | 2023-12-12T12:58:25Z | null | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/6491.diff",
"html_url": "https://github.com/huggingface/datasets/pull/6491",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/6491.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/6491"
} | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6491/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6491/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/174 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/174/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/174/comments | https://api.github.com/repos/huggingface/datasets/issues/174/events | https://github.com/huggingface/datasets/issues/174 | 621,928,403 | MDU6SXNzdWU2MjE5Mjg0MDM= | 174 | nlp.load_dataset('xsum') -> TypeError | {
"avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4",
"events_url": "https://api.github.com/users/sshleifer/events{/privacy}",
"followers_url": "https://api.github.com/users/sshleifer/followers",
"following_url": "https://api.github.com/users/sshleifer/following{/other_user}",
"gists_url": "h... | [] | closed | false | null | [] | null | [] | 2020-05-20T16:59:09Z | 2020-05-20T17:43:46Z | 2020-05-20T17:43:46Z | CONTRIBUTOR | null | null | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/174/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/174/timeline | null | completed | false | |
https://api.github.com/repos/huggingface/datasets/issues/6095 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6095/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6095/comments | https://api.github.com/repos/huggingface/datasets/issues/6095/events | https://github.com/huggingface/datasets/pull/6095 | 1,826,496,967 | PR_kwDODunzps5WqJtr | 6,095 | Fix deprecation of errors in TextConfig | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... | 2023-07-28T14:08:37Z | 2023-07-31T05:26:32Z | 2023-07-31T05:17:38Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/6095.diff",
"html_url": "https://github.com/huggingface/datasets/pull/6095",
"merged_at": "2023-07-31T05:17:38Z",
"patch_url": "https://github.com/huggingface/datasets/pull/6095.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This PR fixes an issue with the deprecation of `errors` in `TextConfig` introduced by:
- #5974
```python
In [1]: ds = load_dataset("text", data_files="test.txt", errors="strict")
---------------------------------------------------------------------------
TypeError Traceback (most ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6095/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6095/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3617 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3617/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3617/comments | https://api.github.com/repos/huggingface/datasets/issues/3617/events | https://github.com/huggingface/datasets/pull/3617 | 1,111,938,691 | PR_kwDODunzps4xdb8K | 3,617 | PR for the CFPB Consumer Complaints dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/42403093?v=4",
"events_url": "https://api.github.com/users/kayvane1/events{/privacy}",
"followers_url": "https://api.github.com/users/kayvane1/followers",
"following_url": "https://api.github.com/users/kayvane1/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [
"> Nice ! Thanks for adding this dataset :)\n> \n> \n> \n> I left a few comments:\n\nThanks!\n\nI'd be interested in contributing to the core codebase - I had to go down the custom loading approach because I couldn't pull this dataset in using the load_dataset() method. Using either the json or csv files available ... | 2022-01-23T17:47:12Z | 2022-02-07T21:08:31Z | 2022-02-07T21:08:31Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3617.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3617",
"merged_at": "2022-02-07T21:08:31Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3617.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Think I followed all the steps but please let me know if anything needs changing or any improvements I can make to the code quality | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 1,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3617/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3617/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4563 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4563/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4563/comments | https://api.github.com/repos/huggingface/datasets/issues/4563/events | https://github.com/huggingface/datasets/pull/4563 | 1,283,914,383 | PR_kwDODunzps46UmZQ | 4,563 | Support streaming allocine dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-06-24T15:55:03Z | 2022-06-24T16:54:57Z | 2022-06-24T16:44:41Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4563.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4563",
"merged_at": "2022-06-24T16:44:41Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4563.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Support streaming allocine dataset.
Fix #4562. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4563/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4563/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/155 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/155/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/155/comments | https://api.github.com/repos/huggingface/datasets/issues/155/events | https://github.com/huggingface/datasets/pull/155 | 620,067,946 | MDExOlB1bGxSZXF1ZXN0NDE5Mzg1ODM0 | 155 | Include more links in README, fix typos | {
"avatar_url": "https://avatars.githubusercontent.com/u/13381361?v=4",
"events_url": "https://api.github.com/users/bharatr21/events{/privacy}",
"followers_url": "https://api.github.com/users/bharatr21/followers",
"following_url": "https://api.github.com/users/bharatr21/following{/other_user}",
"gists_url": "... | [] | closed | false | null | [] | null | [
"I fixed a conflict :) thanks !"
] | 2020-05-18T09:47:08Z | 2020-05-28T08:31:57Z | 2020-05-28T08:31:57Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/155.diff",
"html_url": "https://github.com/huggingface/datasets/pull/155",
"merged_at": "2020-05-28T08:31:57Z",
"patch_url": "https://github.com/huggingface/datasets/pull/155.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/155... | Include more links and fix typos in README | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/155/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/155/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2407 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2407/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2407/comments | https://api.github.com/repos/huggingface/datasets/issues/2407/events | https://github.com/huggingface/datasets/issues/2407 | 903,111,755 | MDU6SXNzdWU5MDMxMTE3NTU= | 2,407 | .map() function got an unexpected keyword argument 'cache_file_name' | {
"avatar_url": "https://avatars.githubusercontent.com/u/7390482?v=4",
"events_url": "https://api.github.com/users/cindyxinyiwang/events{/privacy}",
"followers_url": "https://api.github.com/users/cindyxinyiwang/followers",
"following_url": "https://api.github.com/users/cindyxinyiwang/following{/other_user}",
... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"Hi @cindyxinyiwang,\r\nDid you try adding `.arrow` after `cache_file_name` argument? Here I think they're expecting something like that only for a cache file:\r\nhttps://github.com/huggingface/datasets/blob/e08362256fb157c0b3038437fc0d7a0bbb50de5c/src/datasets/arrow_dataset.py#L1556-L1558",
"Hi ! `cache_file_nam... | 2021-05-27T01:54:26Z | 2021-05-27T13:46:40Z | 2021-05-27T13:46:40Z | NONE | null | null | null | ## Describe the bug
I'm trying to save the result of datasets.map() to a specific file, so that I can easily share it among multiple computers without reprocessing the dataset. However, when I try to pass an argument 'cache_file_name' to the .map() function, it throws an error that ".map() function got an unexpected... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2407/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2407/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/6140 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6140/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6140/comments | https://api.github.com/repos/huggingface/datasets/issues/6140/events | https://github.com/huggingface/datasets/issues/6140 | 1,845,384,712 | I_kwDODunzps5t_lYI | 6,140 | Misalignment between file format specified in configs metadata YAML and the inferred builder | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [] | 2023-08-10T15:07:34Z | 2023-08-17T20:37:20Z | 2023-08-17T20:37:20Z | MEMBER | null | null | null | There is a misalignment between the format of the `data_files` specified in the configs metadata YAML (CSV):
```yaml
configs:
- config_name: default
data_files:
- split: train
path: data.csv
```
and the inferred builder (JSON). Note there are multiple JSON files in the repo, but they do not... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6140/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6140/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1954 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1954/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1954/comments | https://api.github.com/repos/huggingface/datasets/issues/1954/events | https://github.com/huggingface/datasets/issues/1954 | 817,565,563 | MDU6SXNzdWU4MTc1NjU1NjM= | 1,954 | add a new column | {
"avatar_url": "https://avatars.githubusercontent.com/u/79165106?v=4",
"events_url": "https://api.github.com/users/dorost1234/events{/privacy}",
"followers_url": "https://api.github.com/users/dorost1234/followers",
"following_url": "https://api.github.com/users/dorost1234/following{/other_user}",
"gists_url"... | [] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/o... | null | [
"Hi\r\nnot sure how change the lable after creation, but this is an issue not dataset request. thanks ",
"Hi ! Currently you have to use `map` . You can see an example of how to do it in this comment: https://github.com/huggingface/datasets/issues/853#issuecomment-727872188\r\n\r\nIn the future we'll add support ... | 2021-02-26T18:17:27Z | 2021-04-29T14:50:43Z | 2021-04-29T14:50:43Z | NONE | null | null | null | Hi
I'd need to add a new column to the dataset, I was wondering how this can be done? thanks
@lhoestq | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1954/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1954/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/3871 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3871/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3871/comments | https://api.github.com/repos/huggingface/datasets/issues/3871/events | https://github.com/huggingface/datasets/pull/3871 | 1,163,714,113 | PR_kwDODunzps40KRcM | 3,871 | add pandas to env command | {
"avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4",
"events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}",
"followers_url": "https://api.github.com/users/patrickvonplaten/followers",
"following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use... | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_3871). All of your documentation changes will be reflected on that endpoint.",
"Think failures are unrelated - feel free to merge whenever you want :-)"
] | 2022-03-09T09:48:51Z | 2022-03-09T11:21:38Z | 2022-03-09T11:21:37Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3871.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3871",
"merged_at": "2022-03-09T11:21:37Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3871.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Pandas is a required packages and used quite a bit. I don't see any downside with adding its version to the `datasets-cli env` command. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3871/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3871/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3354 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3354/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3354/comments | https://api.github.com/repos/huggingface/datasets/issues/3354/events | https://github.com/huggingface/datasets/pull/3354 | 1,068,307,271 | PR_kwDODunzps4vPl9d | 3,354 | Remove duplicate name from dataset cards | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [] | 2021-12-01T11:45:40Z | 2021-12-01T13:14:30Z | 2021-12-01T13:14:29Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3354.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3354",
"merged_at": "2021-12-01T13:14:29Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3354.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Remove duplicate name from dataset card for:
- ajgt_twitter_ar
- emotone_ar | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3354/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3354/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1611 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1611/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1611/comments | https://api.github.com/repos/huggingface/datasets/issues/1611/events | https://github.com/huggingface/datasets/issues/1611 | 771,486,456 | MDU6SXNzdWU3NzE0ODY0NTY= | 1,611 | shuffle with torch generator | {
"avatar_url": "https://avatars.githubusercontent.com/u/73364383?v=4",
"events_url": "https://api.github.com/users/rabeehkarimimahabadi/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehkarimimahabadi/followers",
"following_url": "https://api.github.com/users/rabeehkarimimahabadi/followin... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | [] | null | [
"Is there a way one can convert the two generator? not sure overall what alternatives I could have to shuffle the datasets with a torch generator, thanks ",
"@lhoestq let me please expalin in more details, maybe you could help me suggesting an alternative to solve the issue for now, I have multiple large dataset... | 2020-12-20T00:57:14Z | 2022-06-01T15:30:13Z | 2022-06-01T15:30:13Z | NONE | null | null | null | Hi
I need to shuffle mutliple large datasets with `generator = torch.Generator()` for a distributed sampler which needs to make sure datasets are consistent across different cores, for this, this is really necessary for me to use torch generator, based on documentation this generator is not supported with datasets, I... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1611/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1611/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/369 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/369/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/369/comments | https://api.github.com/repos/huggingface/datasets/issues/369/events | https://github.com/huggingface/datasets/issues/369 | 654,186,890 | MDU6SXNzdWU2NTQxODY4OTA= | 369 | can't load local dataset: pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries | {
"avatar_url": "https://avatars.githubusercontent.com/u/24683907?v=4",
"events_url": "https://api.github.com/users/vegarab/events{/privacy}",
"followers_url": "https://api.github.com/users/vegarab/followers",
"following_url": "https://api.github.com/users/vegarab/following{/other_user}",
"gists_url": "https:... | [
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] | closed | false | null | [] | null | [
"I am able to reproduce this with the official SQuAD `train-v2.0.json` file downloaded directly from https://rajpurkar.github.io/SQuAD-explorer/",
"I am facing this issue in transformers library 3.0.2 while reading a csv using datasets.\r\nIs this fixed in latest version? \r\nI updated the latest version 4.0.1 bu... | 2020-07-09T16:16:53Z | 2020-12-15T23:07:22Z | 2020-07-10T14:52:06Z | CONTRIBUTOR | null | null | null | Trying to load a local SQuAD-formatted dataset (from a JSON file, about 60MB):
```
dataset = nlp.load_dataset(path='json', data_files={nlp.Split.TRAIN: ["./path/to/file.json"]})
```
causes
```
Traceback (most recent call last):
File "dataloader.py", line 9, in <module>
["./path/to/file.json"]})
File "/... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/369/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/369/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/2448 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2448/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2448/comments | https://api.github.com/repos/huggingface/datasets/issues/2448/events | https://github.com/huggingface/datasets/pull/2448 | 912,360,109 | MDExOlB1bGxSZXF1ZXN0NjYyNTI2NjA3 | 2,448 | Fix flores download link | {
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [] | 2021-06-05T17:30:24Z | 2021-06-08T20:02:58Z | 2021-06-07T08:18:25Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2448.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2448",
"merged_at": "2021-06-07T08:18:25Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2448.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2448/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2448/timeline | null | null | true | |
https://api.github.com/repos/huggingface/datasets/issues/3141 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3141/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3141/comments | https://api.github.com/repos/huggingface/datasets/issues/3141/events | https://github.com/huggingface/datasets/pull/3141 | 1,033,555,910 | PR_kwDODunzps4tjGYz | 3,141 | Fix caching bugs | {
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [] | 2021-10-22T12:59:25Z | 2021-10-22T20:52:08Z | 2021-10-22T13:47:05Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3141.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3141",
"merged_at": "2021-10-22T13:47:04Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3141.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This PR fixes some caching bugs (most likely introduced in the latest refactor):
* remove ")" added by accident in the dataset dir name
* correctly pass the namespace kwargs in `CachedDatasetModuleFactory`
* improve the warning message if `HF_DATASETS_OFFLINE is `True`
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 1,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3141/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3141/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/515 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/515/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/515/comments | https://api.github.com/repos/huggingface/datasets/issues/515/events | https://github.com/huggingface/datasets/pull/515 | 681,845,619 | MDExOlB1bGxSZXF1ZXN0NDcwMTY5MTQ0 | 515 | Fix batched map for formatted dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [] | 2020-08-19T13:34:50Z | 2020-08-20T20:30:43Z | 2020-08-20T20:30:42Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/515.diff",
"html_url": "https://github.com/huggingface/datasets/pull/515",
"merged_at": "2020-08-20T20:30:42Z",
"patch_url": "https://github.com/huggingface/datasets/pull/515.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/515... | If you had a dataset formatted as numpy for example, and tried to do a batched map, then it would crash because one of the elements from the inputs was missing for unchanged columns (ex: batch of length 999 instead of 1000).
The happened during the creation of the `pa.Table`, since columns had different lengths. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/515/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/515/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1462 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1462/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1462/comments | https://api.github.com/repos/huggingface/datasets/issues/1462/events | https://github.com/huggingface/datasets/pull/1462 | 761,489,274 | MDExOlB1bGxSZXF1ZXN0NTM2MTQ4Njc1 | 1,462 | Added conv ai 2 (Again) | {
"avatar_url": "https://avatars.githubusercontent.com/u/22396042?v=4",
"events_url": "https://api.github.com/users/rkc007/events{/privacy}",
"followers_url": "https://api.github.com/users/rkc007/followers",
"following_url": "https://api.github.com/users/rkc007/following{/other_user}",
"gists_url": "https://a... | [] | closed | false | null | [] | null | [
"Looking perfect to me, need to rerun the tests\r\n",
"Thanks, @tanmoyio. \r\nHow do I rerun the tests? Should I change something or push a new commit?",
"@rkc007 you don't need to rerun it, @lhoestq @yjernite will rerun it, as there are huge number of PRs in the queue it might take lil bit of time. ",
"ive j... | 2020-12-10T18:21:55Z | 2020-12-13T00:21:32Z | 2020-12-13T00:21:31Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1462.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1462",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1462.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1462"
} | The original PR -> https://github.com/huggingface/datasets/pull/1383
Reason for creating again -
The reason I had to create the PR again was due to the master rebasing issue. After rebasing the changes, all the previous commits got added to the branch. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1462/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1462/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3954 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3954/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3954/comments | https://api.github.com/repos/huggingface/datasets/issues/3954/events | https://github.com/huggingface/datasets/issues/3954 | 1,172,141,664 | I_kwDODunzps5F3XZg | 3,954 | The dataset preview is not available for tdklab/Hebrew_Squad_v1.1 dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/49593805?v=4",
"events_url": "https://api.github.com/users/MatanBenChorin/events{/privacy}",
"followers_url": "https://api.github.com/users/MatanBenChorin/followers",
"following_url": "https://api.github.com/users/MatanBenChorin/following{/other_user}",
... | [] | closed | false | null | [] | null | [
"Hi @MatanBenChorin, thanks for reporting.\r\n\r\nPlease, take into account that the preview may take some time until it properly renders (we are working to reduce this time).\r\n\r\nMaybe @severo can give more details on this.",
"Hi, \r\nThank you",
"Thanks for reporting. We are looking at it and will give upd... | 2022-03-17T09:38:11Z | 2022-04-20T12:39:07Z | 2022-04-20T12:39:07Z | NONE | null | null | null | ## Dataset viewer issue for 'tdklab/Hebrew_Squad_v1.1'
**Link:** https://huggingface.co/api/datasets/tdklab/Hebrew_Squad_v1.1?full=true
The dataset preview is not available for this dataset.
Am I the one who added this dataset ? Yes | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3954/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3954/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/2864 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2864/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2864/comments | https://api.github.com/repos/huggingface/datasets/issues/2864/events | https://github.com/huggingface/datasets/pull/2864 | 986,159,438 | MDExOlB1bGxSZXF1ZXN0NzI1MzkyNjcw | 2,864 | Fix data URL in ToTTo dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | {
"closed_at": null,
"closed_issues": 2,
"created_at": "2021-07-21T15:34:56Z",
"creator": {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillano... | [] | 2021-09-02T05:25:08Z | 2021-09-02T06:47:40Z | 2021-09-02T06:47:40Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2864.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2864",
"merged_at": "2021-09-02T06:47:40Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2864.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Data source host changed their data URL: google-research-datasets/ToTTo@cebeb43.
Fix #2860. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2864/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2864/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/6443 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6443/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6443/comments | https://api.github.com/repos/huggingface/datasets/issues/6443/events | https://github.com/huggingface/datasets/issues/6443 | 2,006,568,368 | I_kwDODunzps53mc2w | 6,443 | Trouble loading files defined in YAML explicitly | {
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"events_url": "https://api.github.com/users/severo/events{/privacy}",
"followers_url": "https://api.github.com/users/severo/followers",
"following_url": "https://api.github.com/users/severo/following{/other_user}",
"gists_url": "https://ap... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | open | false | null | [] | null | [
"There is a typo in one of the file names - `data/edf.csv` should be renamed to `data/def.csv` 🙂. ",
"wow, I reviewed it twice to avoid being ashamed like that, but... I didn't notice the typo.\r\n\r\n---\r\n\r\nBesides this: do you think we would be able to improve the error message to make this clearer?"
] | 2023-11-22T15:18:10Z | 2023-11-23T09:06:20Z | null | CONTRIBUTOR | null | null | null | Look at https://huggingface.co/datasets/severo/doc-yaml-2
It's a reproduction of the example given in the docs at https://huggingface.co/docs/hub/datasets-manual-configuration
```
You can select multiple files per split using a list of paths:
my_dataset_repository/
├── README.md
├── data/
│ ├── abc.csv
... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6443/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6443/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2001 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2001/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2001/comments | https://api.github.com/repos/huggingface/datasets/issues/2001/events | https://github.com/huggingface/datasets/issues/2001 | 823,946,706 | MDU6SXNzdWU4MjM5NDY3MDY= | 2,001 | Empty evidence document ("provenance") in KILT ELI5 dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/16605764?v=4",
"events_url": "https://api.github.com/users/donggyukimc/events{/privacy}",
"followers_url": "https://api.github.com/users/donggyukimc/followers",
"following_url": "https://api.github.com/users/donggyukimc/following{/other_user}",
"gists_u... | [] | closed | false | null | [] | null | [
"Why did you close this issue? How did you end up finding the evidence documents? I'm running into a similar issue with other KILT tasks."
] | 2021-03-07T15:41:35Z | 2022-12-19T19:25:14Z | 2021-03-17T05:51:01Z | NONE | null | null | null | In the original KILT benchmark(https://github.com/facebookresearch/KILT),
all samples has its evidence document (i.e. wikipedia page id) for prediction.
For example, a sample in ELI5 dataset has the format including provenance (=evidence document) like this
`{"id": "1kiwfx", "input": "In Trading Places (1983... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2001/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2001/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/3166 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3166/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3166/comments | https://api.github.com/repos/huggingface/datasets/issues/3166/events | https://github.com/huggingface/datasets/pull/3166 | 1,036,450,283 | PR_kwDODunzps4tsVQJ | 3,166 | Deprecate prepare_module | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"Sounds good, thanks !"
] | 2021-10-26T15:28:24Z | 2021-11-05T09:27:37Z | 2021-11-05T09:27:36Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3166.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3166",
"merged_at": "2021-11-05T09:27:36Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3166.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | In version 1.13, `prepare_module` was deprecated.
This PR adds a deprecation warning and removes it from all the library, using `dataset_module_factory` or `metric_module_factory` instead.
Fix #3165. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3166/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3166/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5203 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5203/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5203/comments | https://api.github.com/repos/huggingface/datasets/issues/5203/events | https://github.com/huggingface/datasets/pull/5203 | 1,436,710,518 | PR_kwDODunzps5CPlnW | 5,203 | Update canonical links to Hub links | {
"avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4",
"events_url": "https://api.github.com/users/stevhliu/events{/privacy}",
"followers_url": "https://api.github.com/users/stevhliu/followers",
"following_url": "https://api.github.com/users/stevhliu/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-11-04T22:50:50Z | 2022-11-07T18:43:05Z | 2022-11-07T18:40:19Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/5203.diff",
"html_url": "https://github.com/huggingface/datasets/pull/5203",
"merged_at": "2022-11-07T18:40:19Z",
"patch_url": "https://github.com/huggingface/datasets/pull/5203.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This PR updates some of the canonical dataset links to their corresponding links on the Hub; closes #5200. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5203/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5203/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/6463 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6463/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6463/comments | https://api.github.com/repos/huggingface/datasets/issues/6463/events | https://github.com/huggingface/datasets/pull/6463 | 2,020,702,967 | PR_kwDODunzps5g46_4 | 6,463 | Disable benchmarks in PRs | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"It's a way to detect regressions in performance sensitive methods like map, and find the commit that lead to the regression",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_bat... | 2023-12-01T11:35:30Z | 2023-12-01T12:09:09Z | 2023-12-01T12:03:04Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/6463.diff",
"html_url": "https://github.com/huggingface/datasets/pull/6463",
"merged_at": "2023-12-01T12:03:04Z",
"patch_url": "https://github.com/huggingface/datasets/pull/6463.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | In order to keep PR pages less spammy / more readable.
Having the benchmarks on commits on `main` is enough imo | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6463/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6463/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2599 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2599/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2599/comments | https://api.github.com/repos/huggingface/datasets/issues/2599/events | https://github.com/huggingface/datasets/pull/2599 | 937,980,229 | MDExOlB1bGxSZXF1ZXN0Njg0NDQ2MTYx | 2,599 | Update processing.rst with other export formats | {
"avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4",
"events_url": "https://api.github.com/users/TevenLeScao/events{/privacy}",
"followers_url": "https://api.github.com/users/TevenLeScao/followers",
"following_url": "https://api.github.com/users/TevenLeScao/following{/other_user}",
"gists_u... | [] | closed | false | null | [] | {
"closed_at": "2021-07-21T15:36:49Z",
"closed_issues": 29,
"created_at": "2021-06-08T18:48:33Z",
"creator": {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/... | [] | 2021-07-06T14:50:38Z | 2021-07-12T14:10:16Z | 2021-07-07T08:05:48Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2599.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2599",
"merged_at": "2021-07-07T08:05:48Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2599.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Add other supported export formats than CSV in the docs. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2599/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2599/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1819 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1819/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1819/comments | https://api.github.com/repos/huggingface/datasets/issues/1819/events | https://github.com/huggingface/datasets/pull/1819 | 801,448,670 | MDExOlB1bGxSZXF1ZXN0NTY3NzYyMzI2 | 1,819 | Fixed spelling `S3Fileystem` to `S3FileSystem` | {
"avatar_url": "https://avatars.githubusercontent.com/u/32632186?v=4",
"events_url": "https://api.github.com/users/philschmid/events{/privacy}",
"followers_url": "https://api.github.com/users/philschmid/followers",
"following_url": "https://api.github.com/users/philschmid/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [] | 2021-02-04T16:36:46Z | 2021-02-04T16:52:27Z | 2021-02-04T16:52:26Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1819.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1819",
"merged_at": "2021-02-04T16:52:26Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1819.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Fixed documentation spelling errors.
Wrong `S3Fileystem`
Right `S3FileSystem` | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1819/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1819/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3112 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3112/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3112/comments | https://api.github.com/repos/huggingface/datasets/issues/3112/events | https://github.com/huggingface/datasets/issues/3112 | 1,030,613,083 | I_kwDODunzps49behb | 3,112 | OverflowError: There was an overflow in the <class 'pyarrow.lib.ListArray'>. Try to reduce writer_batch_size to have batches smaller than 2GB | {
"avatar_url": "https://avatars.githubusercontent.com/u/69694610?v=4",
"events_url": "https://api.github.com/users/BenoitDalFerro/events{/privacy}",
"followers_url": "https://api.github.com/users/BenoitDalFerro/followers",
"following_url": "https://api.github.com/users/BenoitDalFerro/following{/other_user}",
... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | open | false | null | [] | null | [
"I am very unsure on why you tagged me here. I am not a maintainer of the Datasets library and have no idea how to help you.",
"fixed",
"Ok got it, tensor full of NaNs, cf.\r\n\r\n~\\anaconda3\\envs\\xxx\\lib\\site-packages\\datasets\\arrow_writer.py in write_examples_on_file(self)\r\n315 # This check fails wit... | 2021-10-19T18:21:41Z | 2021-10-19T18:52:29Z | null | NONE | null | null | null | ## Describe the bug
Despite having batches way under 2Gb when running `datasets.map()`, after processing correctly the data of the first batch without fuss and irrespective of writer_batch_size (say 2,4,8,16,32,64 and 128 in my case), it returns the following error :
> OverflowError: There was an overflow in the <c... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3112/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3112/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2789 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2789/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2789/comments | https://api.github.com/repos/huggingface/datasets/issues/2789/events | https://github.com/huggingface/datasets/pull/2789 | 967,361,934 | MDExOlB1bGxSZXF1ZXN0NzA5NTQwMzY5 | 2,789 | Updated dataset description of DaNE | {
"avatar_url": "https://avatars.githubusercontent.com/u/23721977?v=4",
"events_url": "https://api.github.com/users/KennethEnevoldsen/events{/privacy}",
"followers_url": "https://api.github.com/users/KennethEnevoldsen/followers",
"following_url": "https://api.github.com/users/KennethEnevoldsen/following{/other_... | [] | closed | false | null | [] | null | [
"Thanks for finishing it @albertvillanova "
] | 2021-08-11T19:58:48Z | 2021-08-12T16:10:59Z | 2021-08-12T16:06:01Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2789.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2789",
"merged_at": "2021-08-12T16:06:01Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2789.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2789/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2789/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2787 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2787/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2787/comments | https://api.github.com/repos/huggingface/datasets/issues/2787/events | https://github.com/huggingface/datasets/issues/2787 | 967,018,406 | MDU6SXNzdWU5NjcwMTg0MDY= | 2,787 | ConnectionError: Couldn't reach https://raw.githubusercontent.com | {
"avatar_url": "https://avatars.githubusercontent.com/u/39627475?v=4",
"events_url": "https://api.github.com/users/jinec/events{/privacy}",
"followers_url": "https://api.github.com/users/jinec/followers",
"following_url": "https://api.github.com/users/jinec/following{/other_user}",
"gists_url": "https://api.... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"the bug code locate in :\r\n if data_args.task_name is not None:\r\n # Downloading and loading a dataset from the hub.\r\n datasets = load_dataset(\"glue\", data_args.task_name, cache_dir=model_args.cache_dir)",
"Hi @jinec,\r\n\r\nFrom time to time we get this kind of `ConnectionError` coming fr... | 2021-08-11T16:19:01Z | 2023-10-03T12:39:25Z | 2021-08-18T15:09:18Z | NONE | null | null | null | Hello,
I am trying to run run_glue.py and it gives me this error -
Traceback (most recent call last):
File "E:/BERT/pytorch_hugging/transformers/examples/pytorch/text-classification/run_glue.py", line 546, in <module>
main()
File "E:/BERT/pytorch_hugging/transformers/examples/pytorch/text-classification/... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2787/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2787/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/5299 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5299/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5299/comments | https://api.github.com/repos/huggingface/datasets/issues/5299/events | https://github.com/huggingface/datasets/pull/5299 | 1,464,695,091 | PR_kwDODunzps5Dt3Sk | 5,299 | Fix xopen for Windows pathnames | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-11-25T15:35:28Z | 2022-11-29T08:23:58Z | 2022-11-29T08:21:24Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/5299.diff",
"html_url": "https://github.com/huggingface/datasets/pull/5299",
"merged_at": "2022-11-29T08:21:24Z",
"patch_url": "https://github.com/huggingface/datasets/pull/5299.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This PR fixes a bug in `xopen` function for Windows pathnames.
Fix #5298. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5299/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5299/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3465 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3465/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3465/comments | https://api.github.com/repos/huggingface/datasets/issues/3465/events | https://github.com/huggingface/datasets/issues/3465 | 1,085,400,432 | I_kwDODunzps5AseVw | 3,465 | Unable to load 'cnn_dailymail' dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/42352729?v=4",
"events_url": "https://api.github.com/users/talha1503/events{/privacy}",
"followers_url": "https://api.github.com/users/talha1503/followers",
"following_url": "https://api.github.com/users/talha1503/following{/other_user}",
"gists_url": "... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
},
{
"color": "cfd3d7",
"default": true,
"descript... | closed | false | null | [] | null | [
"Hi @talha1503, thanks for reporting.\r\n\r\nIt seems there is an issue with one of the data files hosted at Google Drive:\r\n```\r\nGoogle Drive - Quota exceeded\r\n\r\nSorry, you can't view or download this file at this time.\r\n\r\nToo many users have viewed or downloaded this file recently. Please try accessing... | 2021-12-21T03:32:21Z | 2022-02-17T14:13:57Z | 2022-02-17T14:13:57Z | NONE | null | null | null | ## Describe the bug
I wanted to load cnn_dailymail dataset from huggingface datasets on Google Colab, but I am getting an error while loading it.
## Steps to reproduce the bug
```python
from datasets import load_dataset
dataset = load_dataset('cnn_dailymail', '3.0.0', ignore_verifications = True)
```
## Expe... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3465/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3465/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1322 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1322/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1322/comments | https://api.github.com/repos/huggingface/datasets/issues/1322/events | https://github.com/huggingface/datasets/pull/1322 | 759,576,003 | MDExOlB1bGxSZXF1ZXN0NTM0NTU3Njg3 | 1,322 | add indonlu benchmark datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/6518504?v=4",
"events_url": "https://api.github.com/users/yasirabd/events{/privacy}",
"followers_url": "https://api.github.com/users/yasirabd/followers",
"following_url": "https://api.github.com/users/yasirabd/following{/other_user}",
"gists_url": "http... | [] | closed | false | null | [] | null | [] | 2020-12-08T16:10:58Z | 2020-12-13T02:11:27Z | 2020-12-13T01:54:28Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1322.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1322",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1322.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1322"
} | The IndoNLU benchmark is a collection of resources for training, evaluating, and analyzing natural language understanding systems for the Indonesian language. There are 12 datasets in IndoNLU. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1322/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1322/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5251 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5251/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5251/comments | https://api.github.com/repos/huggingface/datasets/issues/5251/events | https://github.com/huggingface/datasets/issues/5251 | 1,451,761,321 | I_kwDODunzps5WiB6p | 5,251 | Docs are not generated after latest release | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"color": "d4c5f9",
"default": false,
"description": "Maintenance tasks",
"id": 4296013012,
"name": "maintenance",
"node_id": "LA_kwDODunzps8AAAABAA_01A",
"url": "https://api.github.com/repos/huggingface/datasets/labels/maintenance"
}
] | closed | false | null | [] | null | [
"After a discussion with @mishig25:\r\n- He said that this action should be triggered if we call our release branch according to the regex `v*-release`, as transformers does\r\n- I said that our procedure is different: our release branch is *temporary* and it is deleted just after the release PR is merged to main\r... | 2022-11-16T14:59:31Z | 2022-11-22T16:27:50Z | 2022-11-22T16:27:50Z | MEMBER | null | null | null | After the latest `datasets` release version 0.7.0, the docs were not generated.
As we have changed the release procedure (so that now we do not push directly to main branch), maybe we should also change the corresponding GitHub action:
https://github.com/huggingface/datasets/blob/edf1902f954c5568daadebcd8754bdad4... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5251/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5251/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/2988 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2988/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2988/comments | https://api.github.com/repos/huggingface/datasets/issues/2988/events | https://github.com/huggingface/datasets/issues/2988 | 1,011,148,017 | I_kwDODunzps48ROTx | 2,988 | IndexError: Invalid key: 14 is out of bounds for size 0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/79165106?v=4",
"events_url": "https://api.github.com/users/dorost1234/events{/privacy}",
"followers_url": "https://api.github.com/users/dorost1234/followers",
"following_url": "https://api.github.com/users/dorost1234/following{/other_user}",
"gists_url"... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"Hi ! Could you check the length of the `self.dataset` object (i.e. the Dataset object passed to the data loader) ? It looks like the dataset is empty.\r\nNot sure why the SWA optimizer would cause this though.",
"Any updates on this? \r\nThe same error occurred to me too when running `cardiffnlp/twitter-roberta-... | 2021-09-29T16:04:24Z | 2022-04-10T14:49:49Z | 2022-04-10T14:49:49Z | NONE | null | null | null | ## Describe the bug
A clear and concise description of what the bug is.
Hi. I am trying to implement stochastic weighted averaging optimizer with transformer library as described here https://pytorch.org/blog/pytorch-1.6-now-includes-stochastic-weight-averaging/ , for this I am using a run_clm.py codes which is wor... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2988/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2988/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/5458 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5458/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5458/comments | https://api.github.com/repos/huggingface/datasets/issues/5458/events | https://github.com/huggingface/datasets/issues/5458 | 1,555,054,737 | I_kwDODunzps5csECR | 5,458 | slice split while streaming | {
"avatar_url": "https://avatars.githubusercontent.com/u/122370631?v=4",
"events_url": "https://api.github.com/users/SvenDS9/events{/privacy}",
"followers_url": "https://api.github.com/users/SvenDS9/followers",
"following_url": "https://api.github.com/users/SvenDS9/following{/other_user}",
"gists_url": "https... | [] | closed | false | null | [] | null | [
"Hi! Yes, that's correct. When `streaming` is `True`, only split names can be specified as `split`, and for slicing, you have to use `.skip`/`.take` instead.\r\n\r\nE.g. \r\n`load_dataset(\"lhoestq/demo1\",revision=None, streaming=True, split=\"train[:3]\")`\r\n\r\nrewritten with `.skip`/`.take`:\r\n`load_dataset(\... | 2023-01-24T14:08:17Z | 2023-01-24T15:11:47Z | 2023-01-24T15:11:47Z | NONE | null | null | null | ### Describe the bug
When using the `load_dataset` function with streaming set to True, slicing splits is apparently not supported.
Did I miss this in the documentation?
### Steps to reproduce the bug
`load_dataset("lhoestq/demo1",revision=None, streaming=True, split="train[:3]")`
causes ValueError: Bad split:... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5458/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5458/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1569 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1569/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1569/comments | https://api.github.com/repos/huggingface/datasets/issues/1569/events | https://github.com/huggingface/datasets/pull/1569 | 766,758,895 | MDExOlB1bGxSZXF1ZXN0NTM5NjkwMjc2 | 1,569 | added un_ga dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/26374564?v=4",
"events_url": "https://api.github.com/users/param087/events{/privacy}",
"followers_url": "https://api.github.com/users/param087/followers",
"following_url": "https://api.github.com/users/param087/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [] | 2020-12-14T17:42:04Z | 2020-12-15T15:28:58Z | 2020-12-15T15:28:58Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1569.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1569",
"merged_at": "2020-12-15T15:28:58Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1569.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Hi :hugs:, This is a PR for [United nations general assembly resolutions: A six-language parallel corpus](http://opus.nlpl.eu/UN.php) dataset.
With suggested changes in #1330 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1569/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1569/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1481 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1481/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1481/comments | https://api.github.com/repos/huggingface/datasets/issues/1481/events | https://github.com/huggingface/datasets/pull/1481 | 762,579,658 | MDExOlB1bGxSZXF1ZXN0NTM3MTEwOTM1 | 1,481 | Fix ADD_NEW_DATASET to avoid rebasing once pushed | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [] | 2020-12-11T16:27:49Z | 2021-01-07T10:10:20Z | 2021-01-07T10:10:20Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1481.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1481",
"merged_at": "2021-01-07T10:10:20Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1481.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 1,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1481/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1481/timeline | null | null | true | |
https://api.github.com/repos/huggingface/datasets/issues/144 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/144/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/144/comments | https://api.github.com/repos/huggingface/datasets/issues/144/events | https://github.com/huggingface/datasets/pull/144 | 619,477,367 | MDExOlB1bGxSZXF1ZXN0NDE4OTY5NjA1 | 144 | [AWS tests] AWS test should not run for canonical datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4",
"events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}",
"followers_url": "https://api.github.com/users/patrickvonplaten/followers",
"following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use... | [] | closed | false | null | [] | null | [] | 2020-05-16T13:39:30Z | 2020-05-16T13:44:34Z | 2020-05-16T13:44:33Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/144.diff",
"html_url": "https://github.com/huggingface/datasets/pull/144",
"merged_at": "2020-05-16T13:44:33Z",
"patch_url": "https://github.com/huggingface/datasets/pull/144.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/144... | AWS tests should in general not run for canonical datasets. Only local tests will run in this case. This way a PR is able to pass when adding a new dataset.
This PR changes to logic to the following:
1) All datasets that are present in `nlp/datasets` are tested only locally. This way when one adds a canonical da... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/144/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/144/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2563 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2563/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2563/comments | https://api.github.com/repos/huggingface/datasets/issues/2563/events | https://github.com/huggingface/datasets/issues/2563 | 932,387,639 | MDU6SXNzdWU5MzIzODc2Mzk= | 2,563 | interleave_datasets for map-style datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists... | null | [] | 2021-06-29T08:57:24Z | 2021-07-01T09:33:33Z | 2021-07-01T09:33:33Z | MEMBER | null | null | null | Currently the `interleave_datasets` functions only works for `IterableDataset`.
Let's make it work for map-style `Dataset` objects as well.
It would work the same way: either alternate between the datasets in order or randomly given probabilities specified by the user. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2563/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2563/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/4383 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4383/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4383/comments | https://api.github.com/repos/huggingface/datasets/issues/4383/events | https://github.com/huggingface/datasets/issues/4383 | 1,243,856,981 | I_kwDODunzps5KI8BV | 4,383 | L | {
"avatar_url": "https://avatars.githubusercontent.com/u/99847861?v=4",
"events_url": "https://api.github.com/users/AronCodes21/events{/privacy}",
"followers_url": "https://api.github.com/users/AronCodes21/followers",
"following_url": "https://api.github.com/users/AronCodes21/following{/other_user}",
"gists_u... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [] | 2022-05-21T03:47:58Z | 2022-05-21T19:20:13Z | 2022-05-21T19:20:13Z | NONE | null | null | null | ## Describe the L
L
## Expected L
A clear and concise lmll
Specify the actual results or traceback.
## Environment info
<!-- You can run the command `datasets-cli env` and copy-and-paste its output below. -->
- `datasets` version:
- Platform:
- Python version:
- PyArrow version: | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4383/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4383/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/803 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/803/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/803/comments | https://api.github.com/repos/huggingface/datasets/issues/803/events | https://github.com/huggingface/datasets/pull/803 | 736,818,917 | MDExOlB1bGxSZXF1ZXN0NTE1OTY1ODE2 | 803 | fix: typos in tutorial to map KILT and TriviaQA | {
"avatar_url": "https://avatars.githubusercontent.com/u/25532159?v=4",
"events_url": "https://api.github.com/users/PaulLerner/events{/privacy}",
"followers_url": "https://api.github.com/users/PaulLerner/followers",
"following_url": "https://api.github.com/users/PaulLerner/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [] | 2020-11-05T10:42:00Z | 2020-11-10T09:08:07Z | 2020-11-10T09:08:07Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/803.diff",
"html_url": "https://github.com/huggingface/datasets/pull/803",
"merged_at": "2020-11-10T09:08:07Z",
"patch_url": "https://github.com/huggingface/datasets/pull/803.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/803... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/803/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/803/timeline | null | null | true | |
https://api.github.com/repos/huggingface/datasets/issues/999 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/999/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/999/comments | https://api.github.com/repos/huggingface/datasets/issues/999/events | https://github.com/huggingface/datasets/pull/999 | 755,246,786 | MDExOlB1bGxSZXF1ZXN0NTMwOTk1MTY3 | 999 | add generated_reviews_enth | {
"avatar_url": "https://avatars.githubusercontent.com/u/15519308?v=4",
"events_url": "https://api.github.com/users/cstorm125/events{/privacy}",
"followers_url": "https://api.github.com/users/cstorm125/followers",
"following_url": "https://api.github.com/users/cstorm125/following{/other_user}",
"gists_url": "... | [] | closed | false | null | [] | null | [] | 2020-12-02T12:50:43Z | 2020-12-03T11:17:28Z | 2020-12-03T11:17:28Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/999.diff",
"html_url": "https://github.com/huggingface/datasets/pull/999",
"merged_at": "2020-12-03T11:17:28Z",
"patch_url": "https://github.com/huggingface/datasets/pull/999.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/999... | `generated_reviews_enth` is created as part of [scb-mt-en-th-2020](https://arxiv.org/pdf/2007.03541.pdf) for machine translation task. This dataset (referred to as `generated_reviews_yn` in [scb-mt-en-th-2020](https://arxiv.org/pdf/2007.03541.pdf)) are English product reviews generated by [CTRL](https://arxiv.org/abs/1... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/999/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/999/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4301 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4301/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4301/comments | https://api.github.com/repos/huggingface/datasets/issues/4301/events | https://github.com/huggingface/datasets/pull/4301 | 1,230,401,256 | PR_kwDODunzps43idlE | 4,301 | Add ImageNet-Sketch dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4",
"events_url": "https://api.github.com/users/nateraw/events{/privacy}",
"followers_url": "https://api.github.com/users/nateraw/followers",
"following_url": "https://api.github.com/users/nateraw/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I think you can go ahead with uploading the data, and also ping the author in parallel. I think the images may subject to copyright anyway (scrapped from google image) so the dataset author is not allowed to set a license to the data... | 2022-05-09T23:38:45Z | 2022-05-23T18:14:14Z | 2022-05-23T18:05:29Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4301.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4301",
"merged_at": "2022-05-23T18:05:29Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4301.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This PR adds the ImageNet-Sketch dataset and resolves #3953 . | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4301/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4301/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1246 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1246/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1246/comments | https://api.github.com/repos/huggingface/datasets/issues/1246/events | https://github.com/huggingface/datasets/pull/1246 | 758,418,652 | MDExOlB1bGxSZXF1ZXN0NTMzNTk0NjIz | 1,246 | arXiv dataset added | {
"avatar_url": "https://avatars.githubusercontent.com/u/33005287?v=4",
"events_url": "https://api.github.com/users/tanmoyio/events{/privacy}",
"followers_url": "https://api.github.com/users/tanmoyio/followers",
"following_url": "https://api.github.com/users/tanmoyio/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [] | 2020-12-07T11:20:23Z | 2020-12-07T14:22:58Z | 2020-12-07T14:22:58Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1246.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1246",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1246.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1246"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1246/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1246/timeline | null | null | true | |
https://api.github.com/repos/huggingface/datasets/issues/553 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/553/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/553/comments | https://api.github.com/repos/huggingface/datasets/issues/553/events | https://github.com/huggingface/datasets/pull/553 | 690,143,182 | MDExOlB1bGxSZXF1ZXN0NDc3MDgxNTg2 | 553 | [Fix GitHub Actions] test adding tmate | {
"avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4",
"events_url": "https://api.github.com/users/thomwolf/events{/privacy}",
"followers_url": "https://api.github.com/users/thomwolf/followers",
"following_url": "https://api.github.com/users/thomwolf/following{/other_user}",
"gists_url": "http... | [] | closed | false | null | [] | null | [] | 2020-09-01T13:28:03Z | 2021-05-05T18:24:38Z | 2020-09-03T09:01:13Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/553.diff",
"html_url": "https://github.com/huggingface/datasets/pull/553",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/553.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/553"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/553/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/553/timeline | null | null | true | |
https://api.github.com/repos/huggingface/datasets/issues/1460 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1460/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1460/comments | https://api.github.com/repos/huggingface/datasets/issues/1460/events | https://github.com/huggingface/datasets/pull/1460 | 761,349,149 | MDExOlB1bGxSZXF1ZXN0NTM2MDI3NzYy | 1,460 | add Bengali Hate Speech dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4",
"events_url": "https://api.github.com/users/stevhliu/events{/privacy}",
"followers_url": "https://api.github.com/users/stevhliu/followers",
"following_url": "https://api.github.com/users/stevhliu/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [
"@lhoestq I think you might want to look at the dataset, and the first data instances mentioned in the README.md is very much offensive. Though this dataset is based on hate speech but I found the dataset heavily disturbing as Bengali is my native language.",
"Hi @tanmoyio indeed you're right.\r\nWe should *at le... | 2020-12-10T15:40:55Z | 2021-09-17T16:54:53Z | 2021-01-04T14:08:29Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1460.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1460",
"merged_at": "2021-01-04T14:08:29Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1460.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1460/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1460/timeline | null | null | true | |
https://api.github.com/repos/huggingface/datasets/issues/1092 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1092/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1092/comments | https://api.github.com/repos/huggingface/datasets/issues/1092/events | https://github.com/huggingface/datasets/pull/1092 | 756,913,134 | MDExOlB1bGxSZXF1ZXN0NTMyMzc5MDY0 | 1,092 | Add Coached Conversation Preference Dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/50873201?v=4",
"events_url": "https://api.github.com/users/vineeths96/events{/privacy}",
"followers_url": "https://api.github.com/users/vineeths96/followers",
"following_url": "https://api.github.com/users/vineeths96/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [] | 2020-12-04T08:36:49Z | 2020-12-20T13:34:00Z | 2020-12-04T13:49:50Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1092.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1092",
"merged_at": "2020-12-04T13:49:50Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1092.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Adding [Coached Conversation Preference Dataset](https://research.google/tools/datasets/coached-conversational-preference-elicitation/)
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1092/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1092/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5504 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5504/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5504/comments | https://api.github.com/repos/huggingface/datasets/issues/5504/events | https://github.com/huggingface/datasets/pull/5504 | 1,570,621,242 | PR_kwDODunzps5JPoWy | 5,504 | don't zero copy timestamps | {
"avatar_url": "https://avatars.githubusercontent.com/u/2512762?v=4",
"events_url": "https://api.github.com/users/dwyatte/events{/privacy}",
"followers_url": "https://api.github.com/users/dwyatte/followers",
"following_url": "https://api.github.com/users/dwyatte/following{/other_user}",
"gists_url": "https:/... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... | 2023-02-03T23:39:04Z | 2023-02-08T17:28:50Z | 2023-02-08T14:33:17Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/5504.diff",
"html_url": "https://github.com/huggingface/datasets/pull/5504",
"merged_at": "2023-02-08T14:33:17Z",
"patch_url": "https://github.com/huggingface/datasets/pull/5504.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Fixes https://github.com/huggingface/datasets/issues/5495
I'm not sure whether we prefer a test here or if timestamps are known to be unsupported (like booleans). The current test at least covers the bug | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5504/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5504/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/217 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/217/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/217/comments | https://api.github.com/repos/huggingface/datasets/issues/217/events | https://github.com/huggingface/datasets/issues/217 | 627,128,403 | MDU6SXNzdWU2MjcxMjg0MDM= | 217 | Multi-task dataset mixing | {
"avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4",
"events_url": "https://api.github.com/users/ghomasHudson/events{/privacy}",
"followers_url": "https://api.github.com/users/ghomasHudson/followers",
"following_url": "https://api.github.com/users/ghomasHudson/following{/other_user}",
"gist... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "c5def5",
"default": fals... | open | false | null | [] | null | [
"I like this feature! I think the first question we should decide on is how to convert all datasets into the same format. In T5, the authors decided to format every dataset into a text-to-text format. If the dataset had \"multiple\" inputs like MNLI, the inputs were concatenated. So in MNLI the input:\r\n\r\n> - **... | 2020-05-29T09:22:26Z | 2022-10-22T00:45:50Z | null | CONTRIBUTOR | null | null | null | It seems like many of the best performing models on the GLUE benchmark make some use of multitask learning (simultaneous training on multiple tasks).
The [T5 paper](https://arxiv.org/pdf/1910.10683.pdf) highlights multiple ways of mixing the tasks together during finetuning:
- **Examples-proportional mixing** - sam... | {
"+1": 12,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 12,
"url": "https://api.github.com/repos/huggingface/datasets/issues/217/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/217/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2323 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2323/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2323/comments | https://api.github.com/repos/huggingface/datasets/issues/2323/events | https://github.com/huggingface/datasets/issues/2323 | 876,438,507 | MDU6SXNzdWU4NzY0Mzg1MDc= | 2,323 | load_dataset("timit_asr") gives back duplicates of just one sample text | {
"avatar_url": "https://avatars.githubusercontent.com/u/33647474?v=4",
"events_url": "https://api.github.com/users/ekeleshian/events{/privacy}",
"followers_url": "https://api.github.com/users/ekeleshian/followers",
"following_url": "https://api.github.com/users/ekeleshian/following{/other_user}",
"gists_url"... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"Upgrading datasets to version 1.6 fixes the issue",
"This bug was fixed in #1995. Upgrading the `datasets` should work! ",
"Thanks @ekeleshian for having reported.\r\n\r\nI am closing this issue once that you updated `datasets`. Feel free to reopen it if the problem persists."
] | 2021-05-05T13:14:48Z | 2021-05-07T10:32:30Z | 2021-05-07T10:32:30Z | NONE | null | null | null | ## Describe the bug
When you look up on key ["train"] and then ['text'], you get back a list with just one sentence duplicated 4620 times. Namely, the sentence "Would such an act of refusal be useful?". Similarly when you look up ['test'] and then ['text'], the list is one sentence repeated "The bungalow was pleasant... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2323/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2323/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/4463 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4463/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4463/comments | https://api.github.com/repos/huggingface/datasets/issues/4463/events | https://github.com/huggingface/datasets/pull/4463 | 1,265,093,211 | PR_kwDODunzps45Vnzu | 4,463 | Use config_id to check split sizes instead of config name | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._",
"closing in favor of https://github.com/huggingface/datasets/pull/4465"
] | 2022-06-08T17:45:24Z | 2023-09-24T10:03:00Z | 2022-06-09T08:06:37Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4463.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4463",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/4463.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4463"
} | Fix https://github.com/huggingface/datasets/issues/4462 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4463/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4463/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2237 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2237/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2237/comments | https://api.github.com/repos/huggingface/datasets/issues/2237/events | https://github.com/huggingface/datasets/issues/2237 | 861,427,439 | MDU6SXNzdWU4NjE0Mjc0Mzk= | 2,237 | Update Dataset.dataset_size after transformed with map | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | null | [] | null | [
"@albertvillanova I would like to take this up. It would be great if you could point me as to how the dataset size is calculated in HF. Thanks!"
] | 2021-04-19T15:19:38Z | 2021-04-20T14:22:05Z | null | MEMBER | null | null | null | After loading a dataset, if we transform it by using `.map` its `dataset_size` attirbute is not updated. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2237/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2237/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/2912 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2912/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2912/comments | https://api.github.com/repos/huggingface/datasets/issues/2912/events | https://github.com/huggingface/datasets/pull/2912 | 996,256,005 | PR_kwDODunzps4rvhgp | 2,912 | Update link to Blog in docs footer | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [] | 2021-09-14T17:23:14Z | 2021-09-15T07:59:23Z | 2021-09-15T07:59:23Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2912.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2912",
"merged_at": "2021-09-15T07:59:23Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2912.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Update link. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2912/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2912/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/6151 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6151/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6151/comments | https://api.github.com/repos/huggingface/datasets/issues/6151/events | https://github.com/huggingface/datasets/issues/6151 | 1,851,497,818 | I_kwDODunzps5uW51a | 6,151 | Faster sorting for single key items | {
"avatar_url": "https://avatars.githubusercontent.com/u/47942453?v=4",
"events_url": "https://api.github.com/users/jackapbutler/events{/privacy}",
"followers_url": "https://api.github.com/users/jackapbutler/followers",
"following_url": "https://api.github.com/users/jackapbutler/following{/other_user}",
"gist... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | [] | null | [
"`Dataset.sort` essentially does the same thing except it uses `pyarrow.compute.sort_indices` which doesn't involve copying the data into python objects (saving memory)\r\n\r\n```python\r\nsort_keys = [(col, \"ascending\") for col in column_names]\r\nindices = pc.sort_indices(self.data, sort_keys=sort_keys)\r\nretu... | 2023-08-15T14:02:31Z | 2023-08-21T14:38:26Z | 2023-08-21T14:38:25Z | NONE | null | null | null | ### Feature request
A faster way to sort a dataset which contains a large number of rows.
### Motivation
The current sorting implementations took significantly longer than expected when I was running on a dataset trying to sort by timestamps.
**Code snippet:**
```python
ds = datasets.load_dataset( "json"... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6151/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6151/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/5988 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5988/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5988/comments | https://api.github.com/repos/huggingface/datasets/issues/5988/events | https://github.com/huggingface/datasets/issues/5988 | 1,773,257,828 | I_kwDODunzps5pscRk | 5,988 | ConnectionError: Couldn't reach dataset_infos.json | {
"avatar_url": "https://avatars.githubusercontent.com/u/20674868?v=4",
"events_url": "https://api.github.com/users/yulingao/events{/privacy}",
"followers_url": "https://api.github.com/users/yulingao/followers",
"following_url": "https://api.github.com/users/yulingao/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [
"Unfortunately, I can't reproduce the error. What does the following code return for you?\r\n```python\r\nimport requests\r\nfrom huggingface_hub import hf_hub_url\r\nr = requests.get(hf_hub_url(\"codeparrot/codeparrot-clean-train\", \"dataset_infos.json\", repo_type=\"dataset\"))\r\n```\r\n\r\nAlso, can you provid... | 2023-06-25T12:39:31Z | 2023-07-07T13:20:57Z | 2023-07-07T13:20:57Z | NONE | null | null | null | ### Describe the bug
I'm trying to load codeparrot/codeparrot-clean-train, but get the following error:
ConnectionError: Couldn't reach https://huggingface.co/datasets/codeparrot/codeparrot-clean-train/resolve/main/dataset_infos.json (ConnectionError(ProtocolError('Connection aborted.', ConnectionResetError(104, 'C... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5988/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5988/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/5736 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5736/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5736/comments | https://api.github.com/repos/huggingface/datasets/issues/5736/events | https://github.com/huggingface/datasets/issues/5736 | 1,662,286,061 | I_kwDODunzps5jFHjt | 5,736 | FORCE_REDOWNLOAD raises "Directory not empty" exception on second run | {
"avatar_url": "https://avatars.githubusercontent.com/u/1219084?v=4",
"events_url": "https://api.github.com/users/rcasero/events{/privacy}",
"followers_url": "https://api.github.com/users/rcasero/followers",
"following_url": "https://api.github.com/users/rcasero/following{/other_user}",
"gists_url": "https:/... | [] | open | false | null | [] | null | [
"Hi ! I couldn't reproduce your issue :/\r\n\r\nIt seems that `shutil.rmtree` failed. It is supposed to work even if the directory is not empty, but you still end up with `OSError: [Errno 39] Directory not empty:`. Can you make sure another process is not using this directory at the same time ?",
"I have the same... | 2023-04-11T11:29:15Z | 2023-11-30T07:16:58Z | null | NONE | null | null | null | ### Describe the bug
Running `load_dataset(..., download_mode=datasets.DownloadMode.FORCE_REDOWNLOAD)` twice raises a `Directory not empty` exception on the second run.
### Steps to reproduce the bug
I cannot test this on datasets v2.11.0 due to #5711, but this happens in v2.10.1.
1. Set up a script `my_dataset.p... | {
"+1": 3,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 3,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5736/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5736/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/5579 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5579/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5579/comments | https://api.github.com/repos/huggingface/datasets/issues/5579/events | https://github.com/huggingface/datasets/pull/5579 | 1,599,732,211 | PR_kwDODunzps5Kwgo4 | 5,579 | Add instructions to create `DataLoader` from augmented dataset in object detection guide | {
"avatar_url": "https://avatars.githubusercontent.com/u/21087104?v=4",
"events_url": "https://api.github.com/users/Laurent2916/events{/privacy}",
"followers_url": "https://api.github.com/users/Laurent2916/followers",
"following_url": "https://api.github.com/users/Laurent2916/following{/other_user}",
"gists_u... | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5579). All of your documentation changes will be reflected on that endpoint.",
"I'm not sure we need this part as we provide a link to the notebook that shows how to train an object detection model, and this notebook instantiat... | 2023-02-25T14:53:17Z | 2023-03-23T19:24:59Z | 2023-03-23T19:24:50Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/5579.diff",
"html_url": "https://github.com/huggingface/datasets/pull/5579",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/5579.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/5579"
} | The following adds instructions on how to create a `DataLoader` from the guide on how to use object detection with augmentations (#4710). I am open to hearing any suggestions for improvement ! | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5579/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5579/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5176 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5176/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5176/comments | https://api.github.com/repos/huggingface/datasets/issues/5176/events | https://github.com/huggingface/datasets/issues/5176 | 1,430,214,539 | I_kwDODunzps5VP1eL | 5,176 | prepare dataset for cloud storage doesn't work | {
"avatar_url": "https://avatars.githubusercontent.com/u/27285078?v=4",
"events_url": "https://api.github.com/users/araonblake/events{/privacy}",
"followers_url": "https://api.github.com/users/araonblake/followers",
"following_url": "https://api.github.com/users/araonblake/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [
"It looks like an issue with `gcsfs`, are you able to instantiate a `GCSFileSystem` manually ?",
"closing since it was probably due to gcsfs"
] | 2022-10-31T17:28:57Z | 2023-03-28T09:11:46Z | 2023-03-28T09:11:45Z | NONE | null | null | null | ### Describe the bug
Following the [documentation](https://huggingface.co/docs/datasets/filesystems#load-and-save-your-datasets-using-your-cloud-storage-filesystem) and [this PR](https://github.com/huggingface/datasets/pull/4724), I was downloading and storing huggingface dataset to cloud storage.
```
from datasets ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5176/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5176/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/2206 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2206/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2206/comments | https://api.github.com/repos/huggingface/datasets/issues/2206/events | https://github.com/huggingface/datasets/issues/2206 | 855,252,415 | MDU6SXNzdWU4NTUyNTI0MTU= | 2,206 | Got pyarrow error when loading a dataset while adding special tokens into the tokenizer | {
"avatar_url": "https://avatars.githubusercontent.com/u/38536635?v=4",
"events_url": "https://api.github.com/users/yana-xuyan/events{/privacy}",
"followers_url": "https://api.github.com/users/yana-xuyan/followers",
"following_url": "https://api.github.com/users/yana-xuyan/following{/other_user}",
"gists_url"... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"Hi,\r\n\r\nthe output of the tokenizers is treated specially in the lib to optimize the dataset size (see the code [here](https://github.com/huggingface/datasets/blob/master/src/datasets/arrow_writer.py#L138-L141)). It looks like that one of the values in a dictionary returned by the tokenizer is out of the assume... | 2021-04-11T08:40:09Z | 2021-11-10T12:18:30Z | 2021-11-10T12:04:28Z | NONE | null | null | null | I added five more special tokens into the GPT2 tokenizer. But after that, when I try to pre-process the data using my previous code, I got an error shown below:
Traceback (most recent call last):
File "/home/xuyan/anaconda3/envs/convqa/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 1687, in _map_sin... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2206/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2206/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/3569 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3569/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3569/comments | https://api.github.com/repos/huggingface/datasets/issues/3569/events | https://github.com/huggingface/datasets/pull/3569 | 1,100,478,994 | PR_kwDODunzps4w3XGo | 3,569 | Add the DKTC dataset (Extension of #3564) | {
"avatar_url": "https://avatars.githubusercontent.com/u/42150335?v=4",
"events_url": "https://api.github.com/users/sooftware/events{/privacy}",
"followers_url": "https://api.github.com/users/sooftware/followers",
"following_url": "https://api.github.com/users/sooftware/following{/other_user}",
"gists_url": "... | [
{
"color": "0e8a16",
"default": false,
"description": "Contribution to a dataset script",
"id": 4564477500,
"name": "dataset contribution",
"node_id": "LA_kwDODunzps8AAAABEBBmPA",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution"
}
] | closed | false | null | [] | null | [
"I reflect your comment! @lhoestq ",
"Wait, the format of the data just changed, so I'll take it into consideration and commit it.",
"I update the code according to the dataset structure change.",
"Thanks ! I think the dummy data are not valid yet - the dummy train.csv file only contains a partial example (th... | 2022-01-12T15:31:29Z | 2022-10-01T06:43:05Z | 2022-10-01T06:43:04Z | NONE | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3569.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3569",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/3569.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3569"
} | New pull request of #3564. (for DKTC)
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3569/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3569/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4980 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4980/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4980/comments | https://api.github.com/repos/huggingface/datasets/issues/4980/events | https://github.com/huggingface/datasets/issues/4980 | 1,374,868,083 | I_kwDODunzps5R8tJz | 4,980 | Make `pyarrow` optional | {
"avatar_url": "https://avatars.githubusercontent.com/u/240344?v=4",
"events_url": "https://api.github.com/users/KOLANICH/events{/privacy}",
"followers_url": "https://api.github.com/users/KOLANICH/followers",
"following_url": "https://api.github.com/users/KOLANICH/following{/other_user}",
"gists_url": "https... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | [] | null | [
"The whole datasets library is pretty much a wrapper to pyarrow (just take a look at some of the source for a Dataset) https://github.com/huggingface/datasets/blob/51aef08ad7053c0bfe8f9a961207b26df15850d3/src/datasets/arrow_dataset.py#L639 \r\n\r\nI think removing the pyarrow dependency would involve a complete rew... | 2022-09-15T17:38:03Z | 2022-09-16T17:23:47Z | 2022-09-16T17:23:47Z | NONE | null | null | null | **Is your feature request related to a problem? Please describe.**
Is `pyarrow` really needed for every dataset?
**Describe the solution you'd like**
It is made optional.
**Describe alternatives you've considered**
Likely, no.
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4980/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4980/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/4533 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4533/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4533/comments | https://api.github.com/repos/huggingface/datasets/issues/4533/events | https://github.com/huggingface/datasets/issues/4533 | 1,277,211,490 | I_kwDODunzps5MILNi | 4,533 | Timestamp not returned as datetime objects in streaming mode | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [
{
"color": "fef2c0",
"default": false,
"description": "",
"id": 3287858981,
"name": "streaming",
"node_id": "MDU6TGFiZWwzMjg3ODU4OTgx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/streaming"
}
] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists... | null | [] | 2022-06-20T17:28:47Z | 2022-06-22T16:29:09Z | 2022-06-22T16:29:09Z | MEMBER | null | null | null | As reported in (internal) https://github.com/huggingface/datasets-server/issues/397
```python
>>> from datasets import load_dataset
>>> dataset = load_dataset("ett", name="h2", split="test", streaming=True)
>>> d = next(iter(dataset))
>>> d['start']
Timestamp('2016-07-01 00:00:00')
```
while loading in non-... | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4533/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4533/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1019 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1019/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1019/comments | https://api.github.com/repos/huggingface/datasets/issues/1019/events | https://github.com/huggingface/datasets/pull/1019 | 755,582,090 | MDExOlB1bGxSZXF1ZXN0NTMxMjY2NzAz | 1,019 | Add caWaC dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [] | 2020-12-02T20:18:55Z | 2020-12-03T14:47:09Z | 2020-12-03T14:47:09Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1019.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1019",
"merged_at": "2020-12-03T14:47:09Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1019.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Add dataset. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1019/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1019/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/6305 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6305/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6305/comments | https://api.github.com/repos/huggingface/datasets/issues/6305/events | https://github.com/huggingface/datasets/issues/6305 | 1,946,010,912 | I_kwDODunzps5z_cUg | 6,305 | Cannot load dataset with `2.14.5`: `FileNotFound` error | {
"avatar_url": "https://avatars.githubusercontent.com/u/167943?v=4",
"events_url": "https://api.github.com/users/finiteautomata/events{/privacy}",
"followers_url": "https://api.github.com/users/finiteautomata/followers",
"following_url": "https://api.github.com/users/finiteautomata/following{/other_user}",
"... | [] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/o... | null | [
"Thanks for reporting, @finiteautomata.\r\n\r\nWe are investigating it. ",
"There is a bug in `datasets`. You can see our proposed fix:\r\n- #6309 "
] | 2023-10-16T20:11:27Z | 2023-10-18T13:50:36Z | 2023-10-18T13:50:36Z | NONE | null | null | null | ### Describe the bug
I'm trying to load [piuba-bigdata/articles_and_comments] and I'm stumbling with this error on `2.14.5`. However, this works on `2.10.0`.
### Steps to reproduce the bug
[Colab link](https://colab.research.google.com/drive/1SAftFMQnFE708ikRnJJHIXZV7R5IBOCE#scrollTo=r2R2ipCCDmsg)
```python
D... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6305/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6305/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/4303 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4303/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4303/comments | https://api.github.com/repos/huggingface/datasets/issues/4303/events | https://github.com/huggingface/datasets/pull/4303 | 1,230,867,728 | PR_kwDODunzps43j8cH | 4,303 | Fix: Add missing comma | {
"avatar_url": "https://avatars.githubusercontent.com/u/3653789?v=4",
"events_url": "https://api.github.com/users/mrm8488/events{/privacy}",
"followers_url": "https://api.github.com/users/mrm8488/followers",
"following_url": "https://api.github.com/users/mrm8488/following{/other_user}",
"gists_url": "https:/... | [] | closed | false | null | [] | null | [
"The CI failure is unrelated to this PR and fixed on master, merging :)"
] | 2022-05-10T09:21:38Z | 2022-05-11T08:50:15Z | 2022-05-11T08:50:14Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4303.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4303",
"merged_at": "2022-05-11T08:50:14Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4303.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4303/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4303/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/714 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/714/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/714/comments | https://api.github.com/repos/huggingface/datasets/issues/714/events | https://github.com/huggingface/datasets/pull/714 | 714,487,881 | MDExOlB1bGxSZXF1ZXN0NDk3NTYzNjAx | 714 | Add the official dependabot implementation | {
"avatar_url": "https://avatars.githubusercontent.com/u/12804673?v=4",
"events_url": "https://api.github.com/users/ALazyMeme/events{/privacy}",
"followers_url": "https://api.github.com/users/ALazyMeme/followers",
"following_url": "https://api.github.com/users/ALazyMeme/following{/other_user}",
"gists_url": "... | [] | closed | false | null | [] | null | [] | 2020-10-05T03:49:45Z | 2020-10-12T11:49:21Z | 2020-10-12T11:49:21Z | NONE | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/714.diff",
"html_url": "https://github.com/huggingface/datasets/pull/714",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/714.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/714"
} | This will keep dependencies up to date. This will require a pr label `dependencies` being created in order to function correctly. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/714/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/714/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5169 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5169/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5169/comments | https://api.github.com/repos/huggingface/datasets/issues/5169/events | https://github.com/huggingface/datasets/pull/5169 | 1,425,075,254 | PR_kwDODunzps5Bow1Q | 5,169 | Add "ipykernel" to list of `co_filename`s to remove | {
"avatar_url": "https://avatars.githubusercontent.com/u/32967787?v=4",
"events_url": "https://api.github.com/users/gpucce/events{/privacy}",
"followers_url": "https://api.github.com/users/gpucce/followers",
"following_url": "https://api.github.com/users/gpucce/following{/other_user}",
"gists_url": "https://a... | [] | closed | false | null | [] | null | [
"I don't know how I could add some tests for this, although jupyter is not among the dependencies so at least that would need to be added. If someone can tell a recommended way I will try to do it!",
"So testing by myself and looking around the jupyter codebase it looks like the `co_filename` of objects created w... | 2022-10-27T05:56:17Z | 2022-11-02T15:46:00Z | 2022-11-02T15:43:20Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/5169.diff",
"html_url": "https://github.com/huggingface/datasets/pull/5169",
"merged_at": "2022-11-02T15:43:20Z",
"patch_url": "https://github.com/huggingface/datasets/pull/5169.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Should resolve #5157 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5169/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5169/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/779 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/779/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/779/comments | https://api.github.com/repos/huggingface/datasets/issues/779/events | https://github.com/huggingface/datasets/pull/779 | 732,514,887 | MDExOlB1bGxSZXF1ZXN0NTEyNDQzMjY0 | 779 | Feature/fidelity metrics from emnlp2020 evaluating and characterizing human rationales | {
"avatar_url": "https://avatars.githubusercontent.com/u/11327413?v=4",
"events_url": "https://api.github.com/users/rathoreanirudh/events{/privacy}",
"followers_url": "https://api.github.com/users/rathoreanirudh/followers",
"following_url": "https://api.github.com/users/rathoreanirudh/following{/other_user}",
... | [
{
"color": "E3165C",
"default": false,
"description": "",
"id": 4190228726,
"name": "transfer-to-evaluate",
"node_id": "LA_kwDODunzps75wdD2",
"url": "https://api.github.com/repos/huggingface/datasets/labels/transfer-to-evaluate"
}
] | closed | false | null | [] | null | [
"Hi ! This looks interesting, thanks for adding it :) \r\n\r\nFor metrics there should only be two features fields: references and predictions.\r\nBoth of them can be defined as you want using nested structures if you need to.\r\nAlso I'm not sure what goes into references and what goes into predictions, could you ... | 2020-10-29T17:31:14Z | 2023-07-11T09:36:30Z | 2023-07-11T09:36:30Z | NONE | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/779.diff",
"html_url": "https://github.com/huggingface/datasets/pull/779",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/779.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/779"
} | This metric computes fidelity (Yu et al. 2019, DeYoung et al. 2019) and normalized fidelity (Carton et al. 2020). | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/779/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/779/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2446 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2446/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2446/comments | https://api.github.com/repos/huggingface/datasets/issues/2446/events | https://github.com/huggingface/datasets/issues/2446 | 911,635,399 | MDU6SXNzdWU5MTE2MzUzOTk= | 2,446 | `yelp_polarity` is broken | {
"avatar_url": "https://avatars.githubusercontent.com/u/22514219?v=4",
"events_url": "https://api.github.com/users/JetRunner/events{/privacy}",
"followers_url": "https://api.github.com/users/JetRunner/followers",
"following_url": "https://api.github.com/users/JetRunner/following{/other_user}",
"gists_url": "... | [] | closed | false | null | [] | null | [
"```\r\nFile \"/home/sasha/.local/share/virtualenvs/lib-ogGKnCK_/lib/python3.7/site-packages/streamlit/script_runner.py\", line 332, in _run_script\r\n exec(code, module.__dict__)\r\nFile \"/home/sasha/nlp-viewer/run.py\", line 233, in <module>\r\n configs = get_confs(option)\r\nFile \"/home/sasha/.local/shar... | 2021-06-04T15:44:29Z | 2021-06-04T18:56:47Z | 2021-06-04T18:56:47Z | CONTRIBUTOR | null | null | null | 
| {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2446/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2446/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/4957 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4957/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4957/comments | https://api.github.com/repos/huggingface/datasets/issues/4957/events | https://github.com/huggingface/datasets/pull/4957 | 1,366,532,849 | PR_kwDODunzps4-nGIk | 4,957 | Add `Dataset.from_generator` | {
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [
"I restarted the builder PR job just in case",
"_The documentation is not available anymore as the PR was closed or merged._",
"CI is now green. https://github.com/huggingface/doc-builder/pull/296 explains why it failed."
] | 2022-09-08T15:08:25Z | 2022-09-16T14:46:35Z | 2022-09-16T14:44:18Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4957.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4957",
"merged_at": "2022-09-16T14:44:18Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4957.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Add `Dataset.from_generator` to the API to allow creating datasets from data larger than RAM. The implementation relies on a packaged module not exposed in `load_dataset` to tie this method with `datasets`' caching mechanism.
Closes https://github.com/huggingface/datasets/issues/4417 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 2,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 2,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4957/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4957/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/942 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/942/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/942/comments | https://api.github.com/repos/huggingface/datasets/issues/942/events | https://github.com/huggingface/datasets/issues/942 | 754,162,318 | MDU6SXNzdWU3NTQxNjIzMTg= | 942 | D | {
"avatar_url": "https://avatars.githubusercontent.com/u/74238514?v=4",
"events_url": "https://api.github.com/users/CryptoMiKKi/events{/privacy}",
"followers_url": "https://api.github.com/users/CryptoMiKKi/followers",
"following_url": "https://api.github.com/users/CryptoMiKKi/following{/other_user}",
"gists_u... | [] | closed | false | null | [] | null | [] | 2020-12-01T08:17:10Z | 2020-12-03T16:42:53Z | 2020-12-03T16:42:53Z | NONE | null | null | null | ## Adding a Dataset
- **Name:** *name of the dataset*
- **Description:** *short description of the dataset (or link to social media or blog post)*
- **Paper:** *link to the dataset paper if available*
- **Data:** *link to the Github repository or current dataset location*
- **Motivation:** *what are some good reasons t... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/942/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/942/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/5102 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5102/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5102/comments | https://api.github.com/repos/huggingface/datasets/issues/5102/events | https://github.com/huggingface/datasets/issues/5102 | 1,404,746,554 | I_kwDODunzps5Turs6 | 5,102 | Error in create a dataset from a Python generator | {
"avatar_url": "https://avatars.githubusercontent.com/u/9004682?v=4",
"events_url": "https://api.github.com/users/yangxuhui/events{/privacy}",
"followers_url": "https://api.github.com/users/yangxuhui/followers",
"following_url": "https://api.github.com/users/yangxuhui/following{/other_user}",
"gists_url": "h... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
},
{
"color": "7057ff",
"default": true,
"descript... | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/9295277?v=4",
"events_url": "https://api.github.com/users/riccardobucco/events{/privacy}",
"followers_url": "https://api.github.com/users/riccardobucco/followers",
"following_url": "https://api.github.com/users/riccardobucco/following{/other_user}",
"gi... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/9295277?v=4",
"events_url": "https://api.github.com/users/riccardobucco/events{/privacy}",
"followers_url": "https://api.github.com/users/riccardobucco/followers",
"following_url": "https://api.github.com/users/riccardobucco/following{/other_u... | null | [
"Hi, thanks for reporting! The last line should be `dataset = Dataset.from_generator(my_gen)`.",
"Can I work on this one?"
] | 2022-10-11T14:28:58Z | 2022-10-12T11:31:56Z | 2022-10-12T11:31:56Z | NONE | null | null | null | ## Describe the bug
In HOW-TO-GUIDES > Load > [Python generator](https://huggingface.co/docs/datasets/v2.5.2/en/loading#python-generator), the code example defines the `my_gen` function, but when creating the dataset, an undefined `my_dict` is passed in.
```Python
>>> from datasets import Dataset
>>> def my_gen... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5102/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5102/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/387 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/387/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/387/comments | https://api.github.com/repos/huggingface/datasets/issues/387/events | https://github.com/huggingface/datasets/issues/387 | 656,361,357 | MDU6SXNzdWU2NTYzNjEzNTc= | 387 | Conversion through to_pandas output numpy arrays for lists instead of python objects | {
"avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4",
"events_url": "https://api.github.com/users/thomwolf/events{/privacy}",
"followers_url": "https://api.github.com/users/thomwolf/followers",
"following_url": "https://api.github.com/users/thomwolf/following{/other_user}",
"gists_url": "http... | [] | closed | false | null | [] | null | [
"To convert from arrow type we have three options: to_numpy, to_pandas and to_pydict/to_pylist.\r\n\r\n- to_numpy and to_pandas return numpy arrays instead of lists but are very fast.\r\n- to_pydict/to_pylist can be 100x slower and become the bottleneck for reading data, but at least they return lists.\r\n\r\nMaybe... | 2020-07-14T06:24:01Z | 2020-07-17T11:37:00Z | 2020-07-17T11:37:00Z | MEMBER | null | null | null | In a related question, the conversion through to_pandas output numpy arrays for the lists instead of python objects.
Here is an example:
```python
>>> dataset._data.slice(key, 1).to_pandas().to_dict("list")
{'sentence1': ['Amrozi accused his brother , whom he called " the witness " , of deliberately distorting hi... | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/387/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/387/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/3742 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3742/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3742/comments | https://api.github.com/repos/huggingface/datasets/issues/3742/events | https://github.com/huggingface/datasets/pull/3742 | 1,141,174,549 | PR_kwDODunzps4y-1v5 | 3,742 | Fix ValueError message formatting in int2str | {
"avatar_url": "https://avatars.githubusercontent.com/u/41182803?v=4",
"events_url": "https://api.github.com/users/akulchik/events{/privacy}",
"followers_url": "https://api.github.com/users/akulchik/followers",
"following_url": "https://api.github.com/users/akulchik/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [] | 2022-02-17T10:50:08Z | 2022-02-17T15:32:02Z | 2022-02-17T15:32:02Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3742.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3742",
"merged_at": "2022-02-17T15:32:02Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3742.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Hi!
I bumped into this particular `ValueError` during my work (because an instance of `np.int64` was passed instead of regular Python `int`), and so I had to `print(type(values))` myself. Apparently, it's just the missing `f` to make message an f-string.
It ain't much for a contribution, but it's honest work. Hop... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3742/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3742/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2616 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2616/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2616/comments | https://api.github.com/repos/huggingface/datasets/issues/2616/events | https://github.com/huggingface/datasets/pull/2616 | 940,799,038 | MDExOlB1bGxSZXF1ZXN0Njg2ODE3NjYz | 2,616 | Support remote data files | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | [] | {
"closed_at": "2021-07-21T15:36:49Z",
"closed_issues": 29,
"created_at": "2021-06-08T18:48:33Z",
"creator": {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/... | [
"@lhoestq maybe we could also use (if available) the ETag of the remote file in `create_config_id`?",
"> @lhoestq maybe we could also use (if available) the ETag of the remote file in `create_config_id`?\r\n\r\nSure ! We can get the ETag with\r\n```python\r\nheaders = get_authentication_headers_for_url(url, use_a... | 2021-07-09T14:07:38Z | 2021-07-09T16:13:41Z | 2021-07-09T16:13:41Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2616.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2616",
"merged_at": "2021-07-09T16:13:41Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2616.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Add support for (streaming) remote data files:
```python
data_files = f"https://huggingface.co/datasets/{repo_id}/resolve/main/{relative_file_path}"
ds = load_dataset("json", split="train", data_files=data_files, streaming=True)
```
cc: @thomwolf | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2616/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2616/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3743 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3743/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3743/comments | https://api.github.com/repos/huggingface/datasets/issues/3743/events | https://github.com/huggingface/datasets/pull/3743 | 1,141,176,011 | PR_kwDODunzps4y-2Do | 3,743 | initial monash time series forecasting repository | {
"avatar_url": "https://avatars.githubusercontent.com/u/8100?v=4",
"events_url": "https://api.github.com/users/kashif/events{/privacy}",
"followers_url": "https://api.github.com/users/kashif/followers",
"following_url": "https://api.github.com/users/kashif/following{/other_user}",
"gists_url": "https://api.g... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._",
"The CI fails are unrelated to this PR, merging !",
"thanks 🙇🏽 "
] | 2022-02-17T10:51:31Z | 2022-03-21T09:54:41Z | 2022-03-21T09:50:16Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3743.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3743",
"merged_at": "2022-03-21T09:50:16Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3743.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | null | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3743/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3743/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3795 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3795/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3795/comments | https://api.github.com/repos/huggingface/datasets/issues/3795/events | https://github.com/huggingface/datasets/issues/3795 | 1,153,261,281 | I_kwDODunzps5EvV7h | 3,795 | can not flatten natural_questions dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/38466901?v=4",
"events_url": "https://api.github.com/users/Hannibal046/events{/privacy}",
"followers_url": "https://api.github.com/users/Hannibal046/followers",
"following_url": "https://api.github.com/users/Hannibal046/following{/other_user}",
"gists_u... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists... | null | [
"same issue. downgrade it to a lower version.",
"Thanks for reporting, I'll take a look tomorrow :)"
] | 2022-02-27T13:57:40Z | 2022-03-21T14:36:12Z | 2022-03-21T14:36:12Z | NONE | null | null | null | ## Describe the bug
after downloading the natural_questions dataset, can not flatten the dataset considering there are `long answer` and `short answer` in `annotations`.
## Steps to reproduce the bug
```python
from datasets import load_dataset
dataset = load_dataset('natural_questions',cache_dir = 'data/datase... | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3795/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3795/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/2 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2/comments | https://api.github.com/repos/huggingface/datasets/issues/2/events | https://github.com/huggingface/datasets/issues/2 | 599,767,671 | MDU6SXNzdWU1OTk3Njc2NzE= | 2 | Issue to read a local dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4",
"events_url": "https://api.github.com/users/jplu/events{/privacy}",
"followers_url": "https://api.github.com/users/jplu/followers",
"following_url": "https://api.github.com/users/jplu/following{/other_user}",
"gists_url": "https://api.githu... | [] | closed | false | null | [] | null | [
"My first bug report ❤️\r\nLooking into this right now!",
"Ok, there are some news, most good than bad :laughing: \r\n\r\nThe dataset script now became:\r\n```python\r\nimport csv\r\n\r\nimport nlp\r\n\r\n\r\nclass Bbc(nlp.GeneratorBasedBuilder):\r\n VERSION = nlp.Version(\"1.0.0\")\r\n\r\n def __init__(sel... | 2020-04-14T18:18:51Z | 2020-05-11T18:55:23Z | 2020-05-11T18:55:22Z | CONTRIBUTOR | null | null | null | Hello,
As proposed by @thomwolf, I open an issue to explain what I'm trying to do without success. What I want to do is to create and load a local dataset, the script I have done is the following:
```python
import os
import csv
import nlp
class BbcConfig(nlp.BuilderConfig):
def __init__(self, **kwarg... | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/6317 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6317/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6317/comments | https://api.github.com/repos/huggingface/datasets/issues/6317/events | https://github.com/huggingface/datasets/issues/6317 | 1,951,965,668 | I_kwDODunzps50WKHk | 6,317 | sentiment140 dataset unavailable | {
"avatar_url": "https://avatars.githubusercontent.com/u/52670382?v=4",
"events_url": "https://api.github.com/users/AndreasKarasenko/events{/privacy}",
"followers_url": "https://api.github.com/users/AndreasKarasenko/followers",
"following_url": "https://api.github.com/users/AndreasKarasenko/following{/other_use... | [] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/o... | null | [
"Thanks for reporting. We are investigating the issue.",
"We have opened an issue in the corresponding Hub dataset: https://huggingface.co/datasets/sentiment140/discussions/3\r\n\r\nLet's continue the discussion there."
] | 2023-10-19T11:25:21Z | 2023-10-19T13:04:56Z | 2023-10-19T13:04:56Z | NONE | null | null | null | ### Describe the bug
loading the dataset using load_dataset("sentiment140") returns the following error
ConnectionError: Couldn't reach http://cs.stanford.edu/people/alecmgo/trainingandtestdata.zip (error 403)
### Steps to reproduce the bug
Run the following code (version should not matter).
```
from data... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6317/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6317/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1796 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1796/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1796/comments | https://api.github.com/repos/huggingface/datasets/issues/1796/events | https://github.com/huggingface/datasets/issues/1796 | 797,329,905 | MDU6SXNzdWU3OTczMjk5MDU= | 1,796 | Filter on dataset too much slowww | {
"avatar_url": "https://avatars.githubusercontent.com/u/20911334?v=4",
"events_url": "https://api.github.com/users/ayubSubhaniya/events{/privacy}",
"followers_url": "https://api.github.com/users/ayubSubhaniya/followers",
"following_url": "https://api.github.com/users/ayubSubhaniya/following{/other_user}",
"g... | [] | open | false | null | [] | null | [
"When I use the filter on the arrow table directly, it works like butter. But I can't find a way to update the table in `Dataset` object.\r\n\r\n```\r\nds_table = dataset.data.filter(mask=dataset['flag'])\r\n```",
"@thomwolf @lhoestq can you guys please take a look and recommend some solution.",
"Hi ! Currently... | 2021-01-30T04:09:19Z | 2021-02-18T17:09:24Z | null | NONE | null | null | null | I have a dataset with 50M rows.
For pre-processing, I need to tokenize this and filter rows with the large sequence.
My tokenization took roughly 12mins. I used `map()` with batch size 1024 and multi-process with 96 processes.
When I applied the `filter()` function it is taking too much time. I need to filter se... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1796/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1796/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/5798 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5798/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5798/comments | https://api.github.com/repos/huggingface/datasets/issues/5798/events | https://github.com/huggingface/datasets/issues/5798 | 1,685,904,526 | I_kwDODunzps5kfNyO | 5,798 | Support parallelized downloading and processing in load_dataset with Spark | {
"avatar_url": "https://avatars.githubusercontent.com/u/12763339?v=4",
"events_url": "https://api.github.com/users/es94129/events{/privacy}",
"followers_url": "https://api.github.com/users/es94129/followers",
"following_url": "https://api.github.com/users/es94129/following{/other_user}",
"gists_url": "https:... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | null | [] | null | [
"Hi ! We're using process pools for parallelism right now. I was wondering if there's a package that implements the same API as a process pool but runs with Spark under the hood ? That or something similar would be cool because users could use whatever distributed framework they want this way.\r\n\r\nFeel free to p... | 2023-04-27T00:16:11Z | 2023-05-25T14:11:41Z | null | CONTRIBUTOR | null | null | null | ### Feature request
When calling `load_dataset` for datasets that have multiple files, support using Spark to distribute the downloading and processing job to worker nodes when `cache_dir` is a cloud file system shared among nodes.
```python
load_dataset(..., use_spark=True)
```
### Motivation
Further speed up ... | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5798/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5798/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/6124 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6124/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6124/comments | https://api.github.com/repos/huggingface/datasets/issues/6124/events | https://github.com/huggingface/datasets/issues/6124 | 1,837,868,112 | I_kwDODunzps5ti6RQ | 6,124 | Datasets crashing runs due to KeyError | {
"avatar_url": "https://avatars.githubusercontent.com/u/25208228?v=4",
"events_url": "https://api.github.com/users/conceptofmind/events{/privacy}",
"followers_url": "https://api.github.com/users/conceptofmind/followers",
"following_url": "https://api.github.com/users/conceptofmind/following{/other_user}",
"g... | [] | closed | false | null | [] | null | [
"i once had the same error and I could fix that by pushing a fake or a dummy commit on my hugging face dataset repo",
"Hi! We need a reproducer to fix this. Can you provide a link to the dataset (if it's public)?",
"> Hi! We need a reproducer to fix this. Can you provide a link to the dataset (if it's public)?\... | 2023-08-05T17:48:56Z | 2023-11-30T16:28:57Z | 2023-11-30T16:28:57Z | NONE | null | null | null | ### Describe the bug
Hi all,
I have been running into a pretty persistent issue recently when trying to load datasets.
```python
train_dataset = load_dataset(
'llama-2-7b-tokenized',
split = 'train'
)
```
I receive a KeyError which crashes the runs.
```
Traceback (most recent call... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6124/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6124/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/4675 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4675/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4675/comments | https://api.github.com/repos/huggingface/datasets/issues/4675/events | https://github.com/huggingface/datasets/issues/4675 | 1,302,193,649 | I_kwDODunzps5NneXx | 4,675 | Unable to use dataset with PyTorch dataloader | {
"avatar_url": "https://avatars.githubusercontent.com/u/25421460?v=4",
"events_url": "https://api.github.com/users/BlueskyFR/events{/privacy}",
"followers_url": "https://api.github.com/users/BlueskyFR/followers",
"following_url": "https://api.github.com/users/BlueskyFR/following{/other_user}",
"gists_url": "... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | open | false | null | [] | null | [
"Hi! `para_crawl` has a single column of type `Translation`, which stores translation dictionaries. These dictionaries can be stored in a NumPy array but not in a PyTorch tensor since PyTorch only supports numeric types. In `datasets`, the conversion to `torch` works as follows: \r\n1. convert PyArrow table to NumP... | 2022-07-12T15:04:04Z | 2022-07-14T14:17:46Z | null | NONE | null | null | null | ## Describe the bug
When using `.with_format("torch")`, an arrow table is returned and I am unable to use it by passing it to a PyTorch DataLoader: please see the code below.
## Steps to reproduce the bug
```python
from datasets import load_dataset
from torch.utils.data import DataLoader
ds = load_dataset(
... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4675/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4675/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/417 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/417/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/417/comments | https://api.github.com/repos/huggingface/datasets/issues/417/events | https://github.com/huggingface/datasets/pull/417 | 661,804,054 | MDExOlB1bGxSZXF1ZXN0NDUzNDMyODE5 | 417 | Fix docstrins multiple metrics instances | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [] | 2020-07-20T13:08:59Z | 2020-07-22T09:51:00Z | 2020-07-22T09:50:59Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/417.diff",
"html_url": "https://github.com/huggingface/datasets/pull/417",
"merged_at": "2020-07-22T09:50:58Z",
"patch_url": "https://github.com/huggingface/datasets/pull/417.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/417... | We change the docstrings of `nlp.Metric.compute`, `nlp.Metric.add` and `nlp.Metric.add_batch` depending on which metric is instantiated. However we had issues when instantiating multiple metrics (docstrings were duplicated).
This should fix #304 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/417/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/417/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2969 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2969/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2969/comments | https://api.github.com/repos/huggingface/datasets/issues/2969/events | https://github.com/huggingface/datasets/issues/2969 | 1,007,217,867 | I_kwDODunzps48COzL | 2,969 | medical-dialog error | {
"avatar_url": "https://avatars.githubusercontent.com/u/43877130?v=4",
"events_url": "https://api.github.com/users/smeyerhot/events{/privacy}",
"followers_url": "https://api.github.com/users/smeyerhot/followers",
"following_url": "https://api.github.com/users/smeyerhot/following{/other_user}",
"gists_url": "... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/o... | null | [
"Hi @smeyerhot, thanks for reporting.\r\n\r\nYou are right: there is an issue with the dataset metadata. I'm fixing it.\r\n\r\nIn the meantime, you can circumvent the issue by passing `ignore_verifications=True`:\r\n```python\r\nraw_datasets = load_dataset(\"medical_dialog\", \"en\", split=\"train\", download_mode=... | 2021-09-25T23:08:44Z | 2021-10-11T07:46:42Z | 2021-10-11T07:46:42Z | NONE | null | null | null | ## Describe the bug
A clear and concise description of what the bug is.
When I attempt to download the huggingface datatset medical_dialog it errors out midway through
## Steps to reproduce the bug
```python
raw_datasets = load_dataset("medical_dialog", "en", split="train", download_mode="force_redownload", data_d... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2969/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2969/timeline | null | completed | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.