url stringlengths 58 61 | repository_url stringclasses 1
value | labels_url stringlengths 72 75 | comments_url stringlengths 67 70 | events_url stringlengths 65 68 | html_url stringlengths 46 51 | id int64 600M 2.05B | node_id stringlengths 18 32 | number int64 2 6.51k | title stringlengths 1 290 | user dict | labels listlengths 0 4 | state stringclasses 2
values | locked bool 1
class | assignee dict | assignees listlengths 0 4 | milestone dict | comments listlengths 0 30 | created_at timestamp[ns, tz=UTC] | updated_at timestamp[ns, tz=UTC] | closed_at timestamp[ns, tz=UTC] | author_association stringclasses 3
values | active_lock_reason float64 | draft float64 0 1 ⌀ | pull_request dict | body stringlengths 0 228k ⌀ | reactions dict | timeline_url stringlengths 67 70 | performed_via_github_app float64 | state_reason stringclasses 3
values | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/3863 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3863/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3863/comments | https://api.github.com/repos/huggingface/datasets/issues/3863/events | https://github.com/huggingface/datasets/pull/3863 | 1,162,802,857 | PR_kwDODunzps40HP-A | 3,863 | Update code blocks | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_3863). All of your documentation changes will be reflected on that endpoint."
] | 2022-03-08T15:34:43Z | 2022-03-09T16:45:30Z | 2022-03-09T16:45:29Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3863.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3863",
"merged_at": "2022-03-09T16:45:29Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3863.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Following https://github.com/huggingface/datasets/pull/3860#issuecomment-1061756712 and https://github.com/huggingface/datasets/pull/3690 we need to update the code blocks to use markdown instead of sphinx | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3863/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3863/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1031 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1031/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1031/comments | https://api.github.com/repos/huggingface/datasets/issues/1031/events | https://github.com/huggingface/datasets/pull/1031 | 755,844,004 | MDExOlB1bGxSZXF1ZXN0NTMxNDgyMzEy | 1,031 | add crows_pairs | {
"avatar_url": "https://avatars.githubusercontent.com/u/27137566?v=4",
"events_url": "https://api.github.com/users/patil-suraj/events{/privacy}",
"followers_url": "https://api.github.com/users/patil-suraj/followers",
"following_url": "https://api.github.com/users/patil-suraj/following{/other_user}",
"gists_u... | [] | closed | false | null | [] | null | [
"looks good now :) wdyt @yjernite ?",
"Looks good to merge for me, can edit the dataset card later if required. Merging"
] | 2020-12-03T05:05:11Z | 2020-12-03T18:29:52Z | 2020-12-03T18:29:39Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1031.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1031",
"merged_at": "2020-12-03T18:29:39Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1031.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This PR adds CrowS-Pairs datasets.
More info:
https://github.com/nyu-mll/crows-pairs/
https://arxiv.org/pdf/2010.00133.pdf | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1031/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1031/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/481 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/481/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/481/comments | https://api.github.com/repos/huggingface/datasets/issues/481/events | https://github.com/huggingface/datasets/pull/481 | 674,567,389 | MDExOlB1bGxSZXF1ZXN0NDY0MjM2MTA1 | 481 | Apply utf-8 encoding to all datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4",
"events_url": "https://api.github.com/users/lewtun/events{/privacy}",
"followers_url": "https://api.github.com/users/lewtun/followers",
"following_url": "https://api.github.com/users/lewtun/following{/other_user}",
"gists_url": "https://a... | [] | closed | false | null | [] | null | [
"Not sure why the AWS test is failing - perhaps I made too many concurrent CI builds 😢. Can someone please rerun the CI to check the error is not on my end?",
"I pushed an improved docstring and the unit tests now pass, which suggests the previous failure on AWS was simply a timeout error. \r\n\r\nFor some reaso... | 2020-08-06T20:02:09Z | 2020-08-20T08:16:08Z | 2020-08-20T08:16:08Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/481.diff",
"html_url": "https://github.com/huggingface/datasets/pull/481",
"merged_at": "2020-08-20T08:16:08Z",
"patch_url": "https://github.com/huggingface/datasets/pull/481.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/481... | ## Description
This PR applies utf-8 encoding for all instances of `with open(...) as f` to all Python files in `datasets/`. As suggested by @thomwolf in #468 , we use regular expressions and the following function
```python
def apply_encoding_on_file_open(filepath: str):
"""Apply UTF-8 encoding for all insta... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/481/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/481/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3591 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3591/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3591/comments | https://api.github.com/repos/huggingface/datasets/issues/3591/events | https://github.com/huggingface/datasets/pull/3591 | 1,106,928,613 | PR_kwDODunzps4xNDoB | 3,591 | Add support for time, date, duration, and decimal dtypes | {
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [
"Is there a dataset which uses these four datatypes for tests purposes?\r\n",
"@severo Not yet. I'll let you know if that changes."
] | 2022-01-18T13:46:05Z | 2022-01-31T18:29:34Z | 2022-01-20T17:37:33Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3591.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3591",
"merged_at": "2022-01-20T17:37:33Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3591.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Add support for the pyarrow time (maps to `datetime.time` in python), date (maps to `datetime.time` in python), duration (maps to `datetime.timedelta` in python), and decimal (maps to `decimal.decimal` in python) dtypes. This should be helpful when writing scripts for time-series datasets. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3591/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3591/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2992 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2992/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2992/comments | https://api.github.com/repos/huggingface/datasets/issues/2992/events | https://github.com/huggingface/datasets/pull/2992 | 1,012,325,594 | PR_kwDODunzps4sg4ZP | 2,992 | Fix f1 metric with None average | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [] | 2021-09-30T15:31:57Z | 2021-10-01T14:17:39Z | 2021-10-01T14:17:38Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2992.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2992",
"merged_at": "2021-10-01T14:17:38Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2992.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Fix #2979. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2992/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2992/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3699 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3699/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3699/comments | https://api.github.com/repos/huggingface/datasets/issues/3699/events | https://github.com/huggingface/datasets/pull/3699 | 1,130,200,593 | PR_kwDODunzps4yY49I | 3,699 | Add dev-only config to Natural Questions dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"Great thanks ! I think we can fix the CI by copying the NQ folder on gcs to 0.0.3. Does that sound good ?",
"I've copied the 0.0.2 folder content to 0.0.3, as suggested.\r\n\r\nI'm updating the dataset card..."
] | 2022-02-10T14:42:24Z | 2022-02-11T09:50:22Z | 2022-02-11T09:50:21Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3699.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3699",
"merged_at": "2022-02-11T09:50:21Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3699.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | As suggested by @lhoestq and @thomwolf, a new config has been added to Natural Questions dataset, so that only dev split can be downloaded.
Fix #413. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3699/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3699/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5658 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5658/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5658/comments | https://api.github.com/repos/huggingface/datasets/issues/5658/events | https://github.com/huggingface/datasets/pull/5658 | 1,634,867,204 | PR_kwDODunzps5MmJe0 | 5,658 | docs: Update num_shards docs to mention num_proc on Dataset and DatasetDict | {
"avatar_url": "https://avatars.githubusercontent.com/u/78612354?v=4",
"events_url": "https://api.github.com/users/connor-henderson/events{/privacy}",
"followers_url": "https://api.github.com/users/connor-henderson/followers",
"following_url": "https://api.github.com/users/connor-henderson/following{/other_use... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... | 2023-03-22T00:12:18Z | 2023-03-24T16:43:34Z | 2023-03-24T16:36:21Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/5658.diff",
"html_url": "https://github.com/huggingface/datasets/pull/5658",
"merged_at": "2023-03-24T16:36:21Z",
"patch_url": "https://github.com/huggingface/datasets/pull/5658.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Closes #5653
@mariosasko | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5658/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5658/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2841 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2841/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2841/comments | https://api.github.com/repos/huggingface/datasets/issues/2841/events | https://github.com/huggingface/datasets/issues/2841 | 980,497,321 | MDU6SXNzdWU5ODA0OTczMjE= | 2,841 | Adding GLUECoS Hinglish and Spanglish code-switching bemchmark | {
"avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4",
"events_url": "https://api.github.com/users/yjernite/events{/privacy}",
"followers_url": "https://api.github.com/users/yjernite/followers",
"following_url": "https://api.github.com/users/yjernite/following{/other_user}",
"gists_url": "htt... | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | open | false | null | [] | null | [
"Hi @yjernite I am interested in adding this dataset. \r\nIn the repo they have also added a code mixed MT task from English to Hinglish [here](https://github.com/microsoft/GLUECoS#code-mixed-machine-translation-task). I think this could be a good dataset addition in itself and then I can add the rest of the GLUECo... | 2021-08-26T17:47:39Z | 2021-10-20T18:41:20Z | null | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** GLUECoS
- **Description:** a Microsoft Benchmark to evaluate code-switching for only two language pairs but a variety of tasks
- **Paper:** https://aclanthology.org/2020.acl-main.329/
- **Data:** https://github.com/microsoft/GLUECoS
- **Motivation:** We currently only have [one othe... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2841/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2841/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/6488 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6488/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6488/comments | https://api.github.com/repos/huggingface/datasets/issues/6488/events | https://github.com/huggingface/datasets/issues/6488 | 2,035,899,898 | I_kwDODunzps55WV36 | 6,488 | 429 Client Error | {
"avatar_url": "https://avatars.githubusercontent.com/u/7882383?v=4",
"events_url": "https://api.github.com/users/sasaadi/events{/privacy}",
"followers_url": "https://api.github.com/users/sasaadi/followers",
"following_url": "https://api.github.com/users/sasaadi/following{/other_user}",
"gists_url": "https:/... | [] | open | false | null | [] | null | [
"Transferring repos as this is a datasets issue "
] | 2023-12-11T15:06:01Z | 2023-12-11T15:34:23Z | null | NONE | null | null | null | Hello, I was downloading the following dataset and after 20% of data was downloaded, I started getting error 429. It is not resolved since a few days. How should I resolve it?
Thanks
Dataset:
https://huggingface.co/datasets/cerebras/SlimPajama-627B
Error:
`requests.exceptions.HTTPError: 429 Client Error: Too M... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6488/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6488/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/1447 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1447/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1447/comments | https://api.github.com/repos/huggingface/datasets/issues/1447/events | https://github.com/huggingface/datasets/pull/1447 | 761,067,955 | MDExOlB1bGxSZXF1ZXN0NTM1NzkxODk1 | 1,447 | Update step-by-step guide for windows | {
"avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4",
"events_url": "https://api.github.com/users/thomwolf/events{/privacy}",
"followers_url": "https://api.github.com/users/thomwolf/followers",
"following_url": "https://api.github.com/users/thomwolf/following{/other_user}",
"gists_url": "http... | [] | closed | false | null | [] | null | [
"Hi @thomwolf, for simplification purposes, I think you could remove the \"`pip install ...`\" steps from this commit, 'cause these deps (black, isort, flake8) are already installed on `pip install -e \".[dev]\"` on the [Start by preparing your environment](https://github.com/huggingface/datasets/blob/704107f924e74... | 2020-12-10T09:30:59Z | 2020-12-10T12:18:47Z | 2020-12-10T09:31:14Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1447.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1447",
"merged_at": "2020-12-10T09:31:14Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1447.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Update step-by-step guide for windows to give an alternative to `make style`. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1447/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1447/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5241 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5241/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5241/comments | https://api.github.com/repos/huggingface/datasets/issues/5241/events | https://github.com/huggingface/datasets/pull/5241 | 1,448,510,407 | PR_kwDODunzps5C3MTG | 5,241 | Support hfh rc version | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-11-14T18:05:47Z | 2022-11-15T16:11:30Z | 2022-11-15T16:09:31Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/5241.diff",
"html_url": "https://github.com/huggingface/datasets/pull/5241",
"merged_at": "2022-11-15T16:09:31Z",
"patch_url": "https://github.com/huggingface/datasets/pull/5241.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | otherwise the code doesn't work for hfh 0.11.0rc0
following #5237 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5241/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5241/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1249 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1249/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1249/comments | https://api.github.com/repos/huggingface/datasets/issues/1249/events | https://github.com/huggingface/datasets/pull/1249 | 758,472,863 | MDExOlB1bGxSZXF1ZXN0NTMzNjQwNjA1 | 1,249 | Add doc2dial dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/35491698?v=4",
"events_url": "https://api.github.com/users/KMFODA/events{/privacy}",
"followers_url": "https://api.github.com/users/KMFODA/followers",
"following_url": "https://api.github.com/users/KMFODA/following{/other_user}",
"gists_url": "https://a... | [] | closed | false | null | [] | null | [
"It not always practical to use nested `Sequence`. If you have troubles with sequence you can use lists instead. \r\n\r\nFor example\r\n```python\r\n\r\nfeatures=datasets.Features(\r\n {\r\n \"dial_id\": datasets.Value(\"string\"),\r\n \"doc_id\": datasets.Value(\"string\"),\r\n \"domain\": ... | 2020-12-07T12:39:09Z | 2020-12-14T16:17:14Z | 2020-12-14T16:17:14Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1249.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1249",
"merged_at": "2020-12-14T16:17:14Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1249.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | ### Doc2dial: A Goal-Oriented Document-Grounded Dialogue Dataset v0.9
Once complete this will add the [Doc2dial](https://doc2dial.github.io/data.html) dataset from the generic data sets list. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1249/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1249/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1281 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1281/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1281/comments | https://api.github.com/repos/huggingface/datasets/issues/1281/events | https://github.com/huggingface/datasets/pull/1281 | 759,203,317 | MDExOlB1bGxSZXF1ZXN0NTM0MjQ0MTA1 | 1,281 | adding hybrid_qa | {
"avatar_url": "https://avatars.githubusercontent.com/u/27137566?v=4",
"events_url": "https://api.github.com/users/patil-suraj/events{/privacy}",
"followers_url": "https://api.github.com/users/patil-suraj/followers",
"following_url": "https://api.github.com/users/patil-suraj/following{/other_user}",
"gists_u... | [] | closed | false | null | [] | null | [] | 2020-12-08T08:10:19Z | 2020-12-08T18:09:28Z | 2020-12-08T18:07:00Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1281.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1281",
"merged_at": "2020-12-08T18:07:00Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1281.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Adding HybridQA: A Dataset of Multi-Hop Question Answering over Tabular and Textual Data
https://github.com/wenhuchen/HybridQA | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1281/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1281/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2175 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2175/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2175/comments | https://api.github.com/repos/huggingface/datasets/issues/2175/events | https://github.com/huggingface/datasets/issues/2175 | 851,836,096 | MDU6SXNzdWU4NTE4MzYwOTY= | 2,175 | dataset.search_batch() function outputs all -1 indices sometime. | {
"avatar_url": "https://avatars.githubusercontent.com/u/16892570?v=4",
"events_url": "https://api.github.com/users/shamanez/events{/privacy}",
"followers_url": "https://api.github.com/users/shamanez/followers",
"following_url": "https://api.github.com/users/shamanez/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [
"Actually, I found the answer [here](https://github.com/facebookresearch/faiss/wiki/FAQ#what-does-it-mean-when-a-search-returns--1-ids). \r\n\r\nSo we have to do some modifications to the code for instances where the index doesn't retrieve any IDs.",
"@lhoestq @patrickvonplaten \r\n\r\nI also found another short... | 2021-04-06T21:50:49Z | 2021-04-16T12:21:16Z | 2021-04-16T12:21:15Z | NONE | null | null | null | I am working with RAG and playing around with different faiss indexes. At the moment I use **index = faiss.index_factory(768, "IVF65536_HNSW32,Flat")**.
During the retrieval phase exactly in [this line of retrieval_rag.py](https://github.com/huggingface/transformers/blob/master/src/transformers/models/rag/retrieval_... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2175/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2175/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/3130 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3130/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3130/comments | https://api.github.com/repos/huggingface/datasets/issues/3130/events | https://github.com/huggingface/datasets/pull/3130 | 1,032,299,417 | PR_kwDODunzps4tfBJU | 3,130 | Create SECURITY.md | {
"avatar_url": "https://avatars.githubusercontent.com/u/28839565?v=4",
"events_url": "https://api.github.com/users/zidingz/events{/privacy}",
"followers_url": "https://api.github.com/users/zidingz/followers",
"following_url": "https://api.github.com/users/zidingz/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"Hi @zidingz, thanks for your contribution.\r\n\r\nHowever I am closing it because it is a duplicate of a previous PR:\r\n - #2958\r\n\r\n"
] | 2021-10-21T10:03:03Z | 2021-10-21T14:33:28Z | 2021-10-21T14:31:50Z | NONE | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3130.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3130",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/3130.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3130"
} | To let the repository confirm feedback@huggingface.co as its security contact. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3130/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3130/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1744 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1744/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1744/comments | https://api.github.com/repos/huggingface/datasets/issues/1744/events | https://github.com/huggingface/datasets/pull/1744 | 787,649,811 | MDExOlB1bGxSZXF1ZXN0NTU2MzA0MjU4 | 1,744 | Add missing "brief" entries to reuters | {
"avatar_url": "https://avatars.githubusercontent.com/u/2238344?v=4",
"events_url": "https://api.github.com/users/jbragg/events{/privacy}",
"followers_url": "https://api.github.com/users/jbragg/followers",
"following_url": "https://api.github.com/users/jbragg/following{/other_user}",
"gists_url": "https://ap... | [] | closed | false | null | [] | null | [
"@lhoestq I ran `make style` but CI code quality still failing and I don't have access to logs",
"It's also likely that due to the previous placement of the field initialization, much of the data about topics etc was simply wrong and carried over from previous entries. Model scores seem to improve significantly w... | 2021-01-17T07:58:49Z | 2021-01-18T11:26:09Z | 2021-01-18T11:26:09Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1744.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1744",
"merged_at": "2021-01-18T11:26:09Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1744.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This brings the number of examples for ModApte to match the stated `Training set (9,603 docs)...Test Set (3,299 docs)` | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1744/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1744/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4692 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4692/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4692/comments | https://api.github.com/repos/huggingface/datasets/issues/4692/events | https://github.com/huggingface/datasets/issues/4692 | 1,306,609,680 | I_kwDODunzps5N4UgQ | 4,692 | Unable to cast a column with `Image()` by using the `cast_column()` feature | {
"avatar_url": "https://avatars.githubusercontent.com/u/28833916?v=4",
"events_url": "https://api.github.com/users/skrishnan99/events{/privacy}",
"followers_url": "https://api.github.com/users/skrishnan99/followers",
"following_url": "https://api.github.com/users/skrishnan99/following{/other_user}",
"gists_u... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"Hi, thanks for reporting! A PR (https://github.com/huggingface/datasets/pull/4614) has already been opened to address this issue."
] | 2022-07-15T22:56:03Z | 2022-07-19T13:36:24Z | 2022-07-19T13:36:24Z | NONE | null | null | null | ## Describe the bug
A clear and concise description of what the bug is.
When I create a dataset, then add a column to the created dataset through the `dataset.add_column` feature and then try to cast a column of the dataset (this column contains image paths) with `Image()` by using the `cast_column()` feature, I ge... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4692/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4692/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/2967 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2967/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2967/comments | https://api.github.com/repos/huggingface/datasets/issues/2967/events | https://github.com/huggingface/datasets/issues/2967 | 1,007,194,837 | I_kwDODunzps48CJLV | 2,967 | Adding vision-and-language datasets (e.g., VQA, VCR) to Datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/42200725?v=4",
"events_url": "https://api.github.com/users/WadeYin9712/events{/privacy}",
"followers_url": "https://api.github.com/users/WadeYin9712/followers",
"following_url": "https://api.github.com/users/WadeYin9712/following{/other_user}",
"gists_u... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | [] | null | [] | 2021-09-25T20:58:15Z | 2021-10-03T20:34:22Z | 2021-10-03T20:34:22Z | NONE | null | null | null | **Is your feature request related to a problem? Please describe.**
Would you like to add any vision-and-language datasets (e.g., VQA, VCR) to Huggingface Datasets?
**Describe the solution you'd like**
N/A
**Describe alternatives you've considered**
N/A
**Additional context**
This is Da Yin at UCLA. Recentl... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2967/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2967/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/864 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/864/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/864/comments | https://api.github.com/repos/huggingface/datasets/issues/864/events | https://github.com/huggingface/datasets/issues/864 | 745,322,357 | MDU6SXNzdWU3NDUzMjIzNTc= | 864 | Unable to download cnn_dailymail dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/46031058?v=4",
"events_url": "https://api.github.com/users/rohitashwa1907/events{/privacy}",
"followers_url": "https://api.github.com/users/rohitashwa1907/followers",
"following_url": "https://api.github.com/users/rohitashwa1907/following{/other_user}",
... | [
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists... | null | [
"Same error here!\r\n",
"Same here! My kaggle notebook stopped working like yesterday. It's strange because I have fixed version of datasets==1.1.2",
"I'm looking at it right now",
"I couldn't reproduce unfortunately. I tried\r\n```python\r\nfrom datasets import load_dataset\r\n\r\nload_dataset(\"cnn_dailymai... | 2020-11-18T04:38:02Z | 2020-11-20T05:22:11Z | 2020-11-20T05:22:10Z | NONE | null | null | null | ### Script to reproduce the error
```
from datasets import load_dataset
train_dataset = load_dataset("cnn_dailymail", "3.0.0", split= 'train[:10%')
valid_dataset = load_dataset("cnn_dailymail","3.0.0", split="validation[:5%]")
```
### Error
```
-------------------------------------------------------------... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/864/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/864/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/5183 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5183/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5183/comments | https://api.github.com/repos/huggingface/datasets/issues/5183/events | https://github.com/huggingface/datasets/issues/5183 | 1,431,418,066 | I_kwDODunzps5VUbTS | 5,183 | Loading an external dataset in a format similar to conll2003 | {
"avatar_url": "https://avatars.githubusercontent.com/u/112555442?v=4",
"events_url": "https://api.github.com/users/Taghreed7878/events{/privacy}",
"followers_url": "https://api.github.com/users/Taghreed7878/followers",
"following_url": "https://api.github.com/users/Taghreed7878/following{/other_user}",
"gis... | [] | closed | false | null | [] | null | [] | 2022-11-01T13:18:29Z | 2022-11-02T11:57:50Z | 2022-11-02T11:57:50Z | NONE | null | null | null | I'm trying to load a custom dataset in a Dataset object, it's similar to conll2003 but with 2 columns only (word entity), I used the following script:
features = datasets.Features(
{"tokens": datasets.Sequence(datasets.Value("string")),
"ner_tags": datasets.Sequence(
datasets.featu... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5183/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5183/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/3709 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3709/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3709/comments | https://api.github.com/repos/huggingface/datasets/issues/3709/events | https://github.com/huggingface/datasets/pull/3709 | 1,132,997,904 | PR_kwDODunzps4yi0J4 | 3,709 | Set base path to hub url for canonical datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"If we agree to have data files in a dedicated directory \"data/\" then we should be fine. You're right we should not try to edit a dataset script from the repository directly, but from github, in order to avoid conflicts"
] | 2022-02-11T19:23:20Z | 2022-02-16T14:02:28Z | 2022-02-16T14:02:27Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3709.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3709",
"merged_at": "2022-02-16T14:02:27Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3709.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This should allow canonical datasets to use relative paths to download data files from the Hub
cc @polinaeterna this will be useful if we have audio datasets that are canonical and for which you'd like to host data files | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 2,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 3,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3709/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3709/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2420 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2420/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2420/comments | https://api.github.com/repos/huggingface/datasets/issues/2420/events | https://github.com/huggingface/datasets/pull/2420 | 904,821,772 | MDExOlB1bGxSZXF1ZXN0NjU1OTQ1ODgw | 2,420 | Updated Dataset Description | {
"avatar_url": "https://avatars.githubusercontent.com/u/10741860?v=4",
"events_url": "https://api.github.com/users/binny-mathew/events{/privacy}",
"followers_url": "https://api.github.com/users/binny-mathew/followers",
"following_url": "https://api.github.com/users/binny-mathew/following{/other_user}",
"gist... | [] | closed | false | null | [] | null | [] | 2021-05-28T07:10:51Z | 2021-06-10T12:11:35Z | 2021-06-10T12:11:35Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2420.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2420",
"merged_at": "2021-06-10T12:11:35Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2420.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Added Point of contact information and several other details about the dataset. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2420/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2420/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4269 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4269/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4269/comments | https://api.github.com/repos/huggingface/datasets/issues/4269/events | https://github.com/huggingface/datasets/pull/4269 | 1,223,865,145 | PR_kwDODunzps43Nzwh | 4,269 | Add license and point of contact to big_patent dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-05-03T09:24:07Z | 2022-05-06T08:38:09Z | 2022-05-03T11:16:19Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4269.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4269",
"merged_at": "2022-05-03T11:16:19Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4269.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Update metadata of big_patent dataset with:
- license
- point of contact | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4269/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4269/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3426 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3426/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3426/comments | https://api.github.com/repos/huggingface/datasets/issues/3426/events | https://github.com/huggingface/datasets/pull/3426 | 1,078,670,031 | PR_kwDODunzps4vxEN5 | 3,426 | Update disaster_response_messages download urls (+ add validation split) | {
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [] | 2021-12-13T15:30:12Z | 2021-12-14T14:38:30Z | 2021-12-14T14:38:29Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3426.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3426",
"merged_at": "2021-12-14T14:38:29Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3426.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Fixes #3240, fixes #3416 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3426/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3426/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3342 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3342/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3342/comments | https://api.github.com/repos/huggingface/datasets/issues/3342/events | https://github.com/huggingface/datasets/pull/3342 | 1,067,481,390 | PR_kwDODunzps4vM3wh | 3,342 | Fix ASSET dataset data URLs | {
"avatar_url": "https://avatars.githubusercontent.com/u/4812544?v=4",
"events_url": "https://api.github.com/users/tianjianjiang/events{/privacy}",
"followers_url": "https://api.github.com/users/tianjianjiang/followers",
"following_url": "https://api.github.com/users/tianjianjiang/following{/other_user}",
"gi... | [] | closed | false | null | [] | null | [
"> Hi @tianjianjiang, thanks for the fix.\r\n> The links should also be updated in the `dataset_infos.json` file.\r\n> The failing tests are due to the missing tag in the header of the `README.md` file:\r\n\r\nHi @albertvillanova, thank you for the info! My apologies for the messy PR.\r\n"
] | 2021-11-30T17:13:30Z | 2021-12-14T14:50:00Z | 2021-12-14T14:50:00Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3342.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3342",
"merged_at": "2021-12-14T14:50:00Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3342.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Change the branch name "master" to "main" in the data URLs, since facebookresearch has changed that. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3342/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3342/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/795 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/795/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/795/comments | https://api.github.com/repos/huggingface/datasets/issues/795/events | https://github.com/huggingface/datasets/issues/795 | 735,198,265 | MDU6SXNzdWU3MzUxOTgyNjU= | 795 | Descriptions of raw and processed versions of wikitext are inverted | {
"avatar_url": "https://avatars.githubusercontent.com/u/16835358?v=4",
"events_url": "https://api.github.com/users/fraboniface/events{/privacy}",
"followers_url": "https://api.github.com/users/fraboniface/followers",
"following_url": "https://api.github.com/users/fraboniface/following{/other_user}",
"gists_u... | [
{
"color": "2edb81",
"default": false,
"description": "A bug in a dataset script provided in the library",
"id": 2067388877,
"name": "dataset bug",
"node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug"
}
] | closed | false | null | [] | null | [
"Yes indeed ! Thanks for reporting",
"Fixed by:\r\n- #3241"
] | 2020-11-03T10:24:51Z | 2022-02-14T15:46:21Z | 2022-02-14T15:46:21Z | NONE | null | null | null | Nothing of importance, but it looks like the descriptions of wikitext-n-v1 and wikitext-n-raw-v1 are inverted for both n=2 and n=103. I just verified by loading them and the `<unk>` tokens are present in the non-raw versions, which confirms that it's a mere inversion of the descriptions and not of the datasets themselv... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/795/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/795/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/2632 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2632/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2632/comments | https://api.github.com/repos/huggingface/datasets/issues/2632/events | https://github.com/huggingface/datasets/pull/2632 | 942,293,727 | MDExOlB1bGxSZXF1ZXN0Njg4MDQyMjcw | 2,632 | add image-classification task template | {
"avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4",
"events_url": "https://api.github.com/users/nateraw/events{/privacy}",
"followers_url": "https://api.github.com/users/nateraw/followers",
"following_url": "https://api.github.com/users/nateraw/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"Awesome!",
"Thanks for adding a new task template - great work @nateraw 🚀 !"
] | 2021-07-12T17:41:03Z | 2021-07-13T15:44:28Z | 2021-07-13T15:28:16Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2632.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2632",
"merged_at": "2021-07-13T15:28:15Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2632.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Snippet below is the tl;dr, but you can try it out directly here:
[](https://colab.research.google.com/gist/nateraw/005c025d41f0e48ae3d4ee61c0f20b70/image-classification-task-template-demo.ipynb)
```python
from datasets import load_datase... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 3,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 3,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2632/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2632/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4718 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4718/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4718/comments | https://api.github.com/repos/huggingface/datasets/issues/4718/events | https://github.com/huggingface/datasets/pull/4718 | 1,309,520,453 | PR_kwDODunzps47prWR | 4,718 | Make Extractor accept Path as input | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-07-19T13:25:06Z | 2022-07-22T13:42:27Z | 2022-07-22T13:29:43Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4718.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4718",
"merged_at": "2022-07-22T13:29:43Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4718.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This PR:
- Makes `Extractor` accept instance of `Path` as input
- Removes unnecessary castings of `Path` to `str` | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4718/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4718/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2949 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2949/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2949/comments | https://api.github.com/repos/huggingface/datasets/issues/2949/events | https://github.com/huggingface/datasets/pull/2949 | 1,001,026,680 | PR_kwDODunzps4r90Pt | 2,949 | Introduce web and wiki config in triviaqa dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/1706443?v=4",
"events_url": "https://api.github.com/users/shirte/events{/privacy}",
"followers_url": "https://api.github.com/users/shirte/followers",
"following_url": "https://api.github.com/users/shirte/following{/other_user}",
"gists_url": "https://ap... | [] | closed | false | null | [] | null | [
"I just made the dummy data smaller :)\r\nOnce github refreshes the change I think we can merge !",
"Thank you so much for reviewing and accepting my pull request!! :)\r\n\r\nI created these rather large dummy data sets to cover all different cases for the row structure. E.g. in the web configuration, it's possib... | 2021-09-20T14:17:23Z | 2021-10-05T13:20:52Z | 2021-10-01T15:39:29Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2949.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2949",
"merged_at": "2021-10-01T15:39:29Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2949.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | The TriviaQA paper suggests that the two subsets (Wikipedia and Web)
should be treated differently. There are also different leaderboards
for the two sets on CodaLab. For that reason, introduce additional
builder configs in the trivia_qa dataset. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2949/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2949/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2849 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2849/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2849/comments | https://api.github.com/repos/huggingface/datasets/issues/2849/events | https://github.com/huggingface/datasets/issues/2849 | 982,631,420 | MDU6SXNzdWU5ODI2MzE0MjA= | 2,849 | Add Open Catalyst Project Dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/7246357?v=4",
"events_url": "https://api.github.com/users/osanseviero/events{/privacy}",
"followers_url": "https://api.github.com/users/osanseviero/followers",
"following_url": "https://api.github.com/users/osanseviero/following{/other_user}",
"gists_ur... | [
{
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset",
"id": 2067376369,
"name": "dataset request",
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request"
}
] | open | false | null | [] | null | [] | 2021-08-30T10:14:39Z | 2021-08-30T10:14:39Z | null | MEMBER | null | null | null | ## Adding a Dataset
- **Name:** Open Catalyst 2020 (OC20) Dataset
- **Website:** https://opencatalystproject.org/
- **Data:** https://github.com/Open-Catalyst-Project/ocp/blob/master/DATASET.md
Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATAS... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2849/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2849/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/87 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/87/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/87/comments | https://api.github.com/repos/huggingface/datasets/issues/87/events | https://github.com/huggingface/datasets/pull/87 | 617,267,118 | MDExOlB1bGxSZXF1ZXN0NDE3MjE1NzA0 | 87 | Add Flores | {
"avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4",
"events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}",
"followers_url": "https://api.github.com/users/patrickvonplaten/followers",
"following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use... | [] | closed | false | null | [] | null | [] | 2020-05-13T08:51:29Z | 2020-05-13T09:23:34Z | 2020-05-13T09:23:33Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/87.diff",
"html_url": "https://github.com/huggingface/datasets/pull/87",
"merged_at": "2020-05-13T09:23:33Z",
"patch_url": "https://github.com/huggingface/datasets/pull/87.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/87"
} | Beautiful language for sure! | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/87/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/87/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2379 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2379/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2379/comments | https://api.github.com/repos/huggingface/datasets/issues/2379/events | https://github.com/huggingface/datasets/pull/2379 | 895,252,597 | MDExOlB1bGxSZXF1ZXN0NjQ3NDk2ODUx | 2,379 | Disallow duplicate keys in yaml tags | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [] | 2021-05-19T10:10:07Z | 2021-05-19T10:45:32Z | 2021-05-19T10:45:31Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2379.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2379",
"merged_at": "2021-05-19T10:45:31Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2379.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Make sure that there's no duplidate keys in yaml tags.
I added the check in the yaml tree constructor's method, so that the verification is done at every level in the yaml structure.
cc @julien-c | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2379/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2379/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3948 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3948/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3948/comments | https://api.github.com/repos/huggingface/datasets/issues/3948/events | https://github.com/huggingface/datasets/pull/3948 | 1,171,460,560 | PR_kwDODunzps40jg1F | 3,948 | Google BLEU Metric Card | {
"avatar_url": "https://avatars.githubusercontent.com/u/27527747?v=4",
"events_url": "https://api.github.com/users/emibaylor/events{/privacy}",
"followers_url": "https://api.github.com/users/emibaylor/followers",
"following_url": "https://api.github.com/users/emibaylor/following{/other_user}",
"gists_url": "... | [] | closed | false | null | [] | null | [
"A few things that aren't clear for me:\r\n- \"Because it performs better on individual sentence pairs as compared to BLEU, Google BLEU has also been used in RL experiments.\" -- why is this the case? why would that make it more usable for RL? (also, you should put \"Reinforcement Learning\" explicitly, not just th... | 2022-03-16T19:27:17Z | 2022-03-21T16:04:26Z | 2022-03-21T16:04:25Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3948.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3948",
"merged_at": "2022-03-21T16:04:25Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3948.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Add metric card for Google BLEU (GLEU) metric
One thing I noticed while writing this up is that, while this metric was made specifically to be better than BLEU at the sentence level instead of the corpus level, the current implementation only allows the calculation of the corpus-level statistic. I think changing thi... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3948/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3948/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3984 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3984/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3984/comments | https://api.github.com/repos/huggingface/datasets/issues/3984/events | https://github.com/huggingface/datasets/issues/3984 | 1,175,822,117 | I_kwDODunzps5GFZ8l | 3,984 | Local and automatic tests fail | {
"avatar_url": "https://avatars.githubusercontent.com/u/20767068?v=4",
"events_url": "https://api.github.com/users/MarkusSagen/events{/privacy}",
"followers_url": "https://api.github.com/users/MarkusSagen/followers",
"following_url": "https://api.github.com/users/MarkusSagen/following{/other_user}",
"gists_u... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"Hi ! To be able to run the tests, you need to install all the test dependencies and additional ones with\r\n```\r\npip install -e .[tests]\r\npip install -r additional-tests-requirements.txt --no-deps\r\n```\r\n\r\nIn particular, you probably need to `sacrebleu`. It looks like it wasn't able to instantiate `sacreb... | 2022-03-21T19:07:37Z | 2023-07-25T15:18:40Z | 2023-07-25T15:18:40Z | NONE | null | null | null | ## Describe the bug
Running the tests from CircleCI on a PR or locally fails, even with no changes. Tests seem to fail on `test_metric_common.py`
## Steps to reproduce the bug
```shell
git clone https://huggingface/datasets.git
cd datasets
```
```python
python -m pip install -e .
pytest
```
## Expected... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3984/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3984/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/2440 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2440/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2440/comments | https://api.github.com/repos/huggingface/datasets/issues/2440/events | https://github.com/huggingface/datasets/issues/2440 | 908,521,954 | MDU6SXNzdWU5MDg1MjE5NTQ= | 2,440 | Remove `extended` field from dataset tagger | {
"avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4",
"events_url": "https://api.github.com/users/lewtun/events{/privacy}",
"followers_url": "https://api.github.com/users/lewtun/followers",
"following_url": "https://api.github.com/users/lewtun/following{/other_user}",
"gists_url": "https://a... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"The tagger also doesn't insert the value for the `size_categories` field automatically, so this should be fixed too",
"Thanks for reporting. Indeed the `extended` tag doesn't exist. Not sure why we had that in the tagger.\r\nThe repo of the tagger is here if someone wants to give this a try: https://github.com/h... | 2021-06-01T17:18:42Z | 2021-06-09T09:06:31Z | 2021-06-09T09:06:30Z | MEMBER | null | null | null | ## Describe the bug
While working on #2435 I used the [dataset tagger](https://huggingface.co/datasets/tagging/) to generate the missing tags for the YAML metadata of each README.md file. However, it seems that our CI raises an error when the `extended` field is included:
```
dataset_name = 'arcd'
@pytest.m... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 1,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2440/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2440/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/2545 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2545/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2545/comments | https://api.github.com/repos/huggingface/datasets/issues/2545/events | https://github.com/huggingface/datasets/pull/2545 | 929,016,580 | MDExOlB1bGxSZXF1ZXN0Njc2OTMxOTYw | 2,545 | Fix DuplicatedKeysError in drop dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [] | 2021-06-24T09:10:39Z | 2021-06-24T14:57:08Z | 2021-06-24T14:57:08Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2545.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2545",
"merged_at": "2021-06-24T14:57:08Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2545.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Close #2542.
cc: @VictorSanh. | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2545/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2545/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/6240 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6240/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6240/comments | https://api.github.com/repos/huggingface/datasets/issues/6240/events | https://github.com/huggingface/datasets/issues/6240 | 1,895,723,888 | I_kwDODunzps5w_nNw | 6,240 | Dataloader stuck on multiple GPUs | {
"avatar_url": "https://avatars.githubusercontent.com/u/40049003?v=4",
"events_url": "https://api.github.com/users/kuri54/events{/privacy}",
"followers_url": "https://api.github.com/users/kuri54/followers",
"following_url": "https://api.github.com/users/kuri54/following{/other_user}",
"gists_url": "https://a... | [] | closed | false | null | [] | null | [
"What type of dataset are you using in this script? `torch.utils.data.Dataset` or `datasets.Dataset`? Please share the `datasets` package version if it's the latter. Otherwise, it's better to move this issue to the `accelerate` repo.",
"Very sorry, I thought I had a repo in `accelerate!`\r\nI will close this issu... | 2023-09-14T05:30:30Z | 2023-09-14T23:54:42Z | 2023-09-14T23:54:42Z | NONE | null | null | null | ### Describe the bug
I am trying to get CLIP to fine-tuning with my code.
When I tried to run it on multiple GPUs using accelerate, I encountered the following phenomenon.
- Validation dataloader stuck in 2nd epoch only on multi-GPU
Specifically, when the "for inputs in valid_loader:" process is finished, it does... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6240/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6240/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/5789 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5789/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5789/comments | https://api.github.com/repos/huggingface/datasets/issues/5789/events | https://github.com/huggingface/datasets/issues/5789 | 1,682,611,179 | I_kwDODunzps5kSpvr | 5,789 | Support streaming datasets that use jsonlines | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | open | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/o... | null | [] | 2023-04-25T07:40:02Z | 2023-04-25T07:40:03Z | null | MEMBER | null | null | null | Extend support for streaming datasets that use `jsonlines.open`.
Currently, if `jsonlines` is installed, `datasets` raises a `FileNotFoundError`:
```
FileNotFoundError: [Errno 2] No such file or directory: 'https://...'
```
See:
- https://huggingface.co/datasets/masakhane/afriqa/discussions/1 | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5789/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5789/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/1946 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1946/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1946/comments | https://api.github.com/repos/huggingface/datasets/issues/1946/events | https://github.com/huggingface/datasets/pull/1946 | 816,526,294 | MDExOlB1bGxSZXF1ZXN0NTgwMTcyNzI2 | 1,946 | Implement Dataset from CSV | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"@lhoestq question about public API: `keep_in_memory` or just `in_memory`?",
"For consistence I'd say `keep_in_memory`, but no strong opinion.",
"@lhoestq done!"
] | 2021-02-25T15:10:13Z | 2021-03-12T09:42:48Z | 2021-03-12T09:42:48Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1946.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1946",
"merged_at": "2021-03-12T09:42:48Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1946.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Implement `Dataset.from_csv`.
Analogue to #1943.
If finally, the scripts should be used instead, at least we can reuse the tests here. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1946/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1946/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3108 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3108/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3108/comments | https://api.github.com/repos/huggingface/datasets/issues/3108/events | https://github.com/huggingface/datasets/pull/3108 | 1,030,405,618 | PR_kwDODunzps4tY8ID | 3,108 | Add Google BLEU (aka GLEU) metric | {
"avatar_url": "https://avatars.githubusercontent.com/u/44175589?v=4",
"events_url": "https://api.github.com/users/slowwavesleep/events{/privacy}",
"followers_url": "https://api.github.com/users/slowwavesleep/followers",
"following_url": "https://api.github.com/users/slowwavesleep/following{/other_user}",
"g... | [] | closed | false | null | [] | null | [] | 2021-10-19T14:48:38Z | 2021-10-25T14:07:04Z | 2021-10-25T14:07:04Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3108.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3108",
"merged_at": "2021-10-25T14:07:04Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3108.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | This PR adds the NLTK implementation of Google BLEU metric. This is also a part of an effort to resolve an unfortunate naming collision between GLEU for machine translation and GLEU for grammatical error correction.
I used [this page](https://huggingface.co/docs/datasets/add_metric.html) for reference. Please, point ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3108/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3108/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/19 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/19/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/19/comments | https://api.github.com/repos/huggingface/datasets/issues/19/events | https://github.com/huggingface/datasets/pull/19 | 606,400,645 | MDExOlB1bGxSZXF1ZXN0NDA4NjIwMjUw | 19 | Replace tf.constant for TF | {
"avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4",
"events_url": "https://api.github.com/users/jplu/events{/privacy}",
"followers_url": "https://api.github.com/users/jplu/followers",
"following_url": "https://api.github.com/users/jplu/following{/other_user}",
"gists_url": "https://api.githu... | [] | closed | false | null | [] | null | [
"Awesome!"
] | 2020-04-24T15:32:06Z | 2020-04-29T09:27:08Z | 2020-04-25T21:18:45Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/19.diff",
"html_url": "https://github.com/huggingface/datasets/pull/19",
"merged_at": "2020-04-25T21:18:45Z",
"patch_url": "https://github.com/huggingface/datasets/pull/19.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/19"
} | Replace simple tf.constant type of Tensor to tf.ragged.constant which allows to have examples of different size in a tf.data.Dataset.
Now the training works with TF. Here the same example than for the PT in collab:
```python
import tensorflow as tf
import nlp
from transformers import BertTokenizerFast, TFBertF... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/19/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/19/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2340 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2340/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2340/comments | https://api.github.com/repos/huggingface/datasets/issues/2340/events | https://github.com/huggingface/datasets/pull/2340 | 882,370,824 | MDExOlB1bGxSZXF1ZXN0NjM1OTExNzIx | 2,340 | More consistent copy logic | {
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [] | 2021-05-09T14:17:33Z | 2021-05-11T08:58:33Z | 2021-05-11T08:58:33Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2340.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2340",
"merged_at": "2021-05-11T08:58:33Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2340.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Use `info.copy()` instead of `copy.deepcopy(info)`.
`Features.copy` now creates a deep copy. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2340/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2340/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4471 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4471/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4471/comments | https://api.github.com/repos/huggingface/datasets/issues/4471/events | https://github.com/huggingface/datasets/issues/4471 | 1,267,475,268 | I_kwDODunzps5LjCNE | 4,471 | CI error with repo lhoestq/_dummy | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"fixed by https://github.com/huggingface/datasets/pull/4472"
] | 2022-06-10T12:26:06Z | 2022-06-10T13:24:53Z | 2022-06-10T13:24:53Z | MEMBER | null | null | null | ## Describe the bug
CI is failing because of repo "lhoestq/_dummy". See: https://app.circleci.com/pipelines/github/huggingface/datasets/12461/workflows/1b040b45-9578-4ab9-8c44-c643c4eb8691/jobs/74269
```
requests.exceptions.HTTPError: 401 Client Error: Unauthorized for url: https://huggingface.co/api/datasets/lhoest... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4471/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4471/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/793 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/793/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/793/comments | https://api.github.com/repos/huggingface/datasets/issues/793/events | https://github.com/huggingface/datasets/pull/793 | 735,105,907 | MDExOlB1bGxSZXF1ZXN0NTE0NTU2NzY5 | 793 | [Datasets] fix discofuse links | {
"avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4",
"events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}",
"followers_url": "https://api.github.com/users/patrickvonplaten/followers",
"following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use... | [] | closed | false | null | [] | null | [] | 2020-11-03T08:03:45Z | 2020-11-03T08:16:41Z | 2020-11-03T08:16:40Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/793.diff",
"html_url": "https://github.com/huggingface/datasets/pull/793",
"merged_at": "2020-11-03T08:16:40Z",
"patch_url": "https://github.com/huggingface/datasets/pull/793.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/793... | The discofuse links were changed: https://github.com/google-research-datasets/discofuse/commit/d27641016eb5b3eb2af03c7415cfbb2cbebe8558.
The old links are broken
I changed the links and created the new dataset_infos.json.
Pinging @thomwolf @lhoestq for notification. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/793/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/793/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4761 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4761/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4761/comments | https://api.github.com/repos/huggingface/datasets/issues/4761/events | https://github.com/huggingface/datasets/issues/4761 | 1,321,068,411 | I_kwDODunzps5Oved7 | 4,761 | parallel searching in multi-gpu setting using faiss | {
"avatar_url": "https://avatars.githubusercontent.com/u/48146603?v=4",
"events_url": "https://api.github.com/users/Jiaxin-Wen/events{/privacy}",
"followers_url": "https://api.github.com/users/Jiaxin-Wen/followers",
"following_url": "https://api.github.com/users/Jiaxin-Wen/following{/other_user}",
"gists_url"... | [] | open | false | null | [] | null | [
"And I don't see any speed up when increasing the number of GPUs while calling `get_nearest_examples_batch`.",
"Hi ! Yes search_batch uses FAISS search which happens in parallel across the GPUs\r\n\r\n> And I don't see any speed up when increasing the number of GPUs while calling get_nearest_examples_batch.\r\n\r... | 2022-07-28T14:57:03Z | 2023-07-21T02:07:10Z | null | CONTRIBUTOR | null | null | null | While I notice that `add_faiss_index` has supported assigning multiple GPUs, I am still confused about how it works.
Does the `search-batch` function automatically parallelizes the input queries to different gpus?https://github.com/huggingface/datasets/blob/d76599bdd4d186b2e7c4f468b05766016055a0a5/src/datasets/sea... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4761/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4761/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/532 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/532/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/532/comments | https://api.github.com/repos/huggingface/datasets/issues/532/events | https://github.com/huggingface/datasets/issues/532 | 685,540,614 | MDU6SXNzdWU2ODU1NDA2MTQ= | 532 | File exists error when used with TPU | {
"avatar_url": "https://avatars.githubusercontent.com/u/20531705?v=4",
"events_url": "https://api.github.com/users/go-inoue/events{/privacy}",
"followers_url": "https://api.github.com/users/go-inoue/followers",
"following_url": "https://api.github.com/users/go-inoue/following{/other_user}",
"gists_url": "htt... | [] | open | false | null | [] | null | [
"I am facing probably facing similar issues with \r\n\r\n`wiki40b_en_100_0`",
"Could you try to run `dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")` once before calling the script ?\r\n\r\nIt looks like several processes try to create the dataset in arrow format at the same time. If the d... | 2020-08-25T14:36:38Z | 2020-09-01T12:14:56Z | null | NONE | null | null | null | Hi,
I'm getting a "File exists" error when I use [text dataset](https://github.com/huggingface/nlp/tree/master/datasets/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8).
I modified [line 131 in the original `run_language_modeling.py`](https://github.com/... | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/532/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/532/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/5568 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5568/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5568/comments | https://api.github.com/repos/huggingface/datasets/issues/5568/events | https://github.com/huggingface/datasets/issues/5568 | 1,596,900,532 | I_kwDODunzps5fLsS0 | 5,568 | dataset.to_iterable_dataset() loses useful info like dataset features | {
"avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4",
"events_url": "https://api.github.com/users/Hubert-Bonisseur/events{/privacy}",
"followers_url": "https://api.github.com/users/Hubert-Bonisseur/followers",
"following_url": "https://api.github.com/users/Hubert-Bonisseur/following{/other_use... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "7057ff",
"default": true... | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4",
"events_url": "https://api.github.com/users/Hubert-Bonisseur/events{/privacy}",
"followers_url": "https://api.github.com/users/Hubert-Bonisseur/followers",
"following_url": "https://api.github.com/users/Hubert-Bonisseur/following{/other_use... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4",
"events_url": "https://api.github.com/users/Hubert-Bonisseur/events{/privacy}",
"followers_url": "https://api.github.com/users/Hubert-Bonisseur/followers",
"following_url": "https://api.github.com/users/Hubert-Bonisseur/followin... | null | [
"Hi ! Oh good catch. I think the features should be passed to `IterableDataset.from_generator()` in `to_iterable_dataset()` indeed.\r\n\r\nSetting this as a good first issue if someone would like to contribute, otherwise we can take care of it :)",
"#self-assign",
"seems like the feature parameter is missing fr... | 2023-02-23T13:45:33Z | 2023-02-24T13:22:36Z | 2023-02-24T13:22:36Z | CONTRIBUTOR | null | null | null | ### Describe the bug
Hello,
I like the new `to_iterable_dataset` feature but I noticed something that seems to be missing.
When using `to_iterable_dataset` to transform your map style dataset into iterable dataset, you lose valuable metadata like the features.
These metadata are useful if you want to interleav... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5568/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5568/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/3359 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3359/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3359/comments | https://api.github.com/repos/huggingface/datasets/issues/3359/events | https://github.com/huggingface/datasets/pull/3359 | 1,068,638,213 | PR_kwDODunzps4vQtI0 | 3,359 | Add The Pile Free Law subset | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"@albertvillanova Is there a specific reason you’re adding the Pile under “the” instead of under “pile”? That does not appear to be consistent with other datasets.",
"Hi @StellaAthena,\r\n\r\nI asked myself the same question, but at the end I decided to be consistent with previously added Pile subsets:\r\n- #2817... | 2021-12-01T16:46:04Z | 2021-12-06T10:12:17Z | 2021-12-01T17:30:44Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3359.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3359",
"merged_at": "2021-12-01T17:30:43Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3359.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Add:
- Free Law subset of The Pile: "free_law" config
Close bigscience-workshop/data_tooling#75.
CC: @StellaAthena | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3359/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3359/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/359 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/359/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/359/comments | https://api.github.com/repos/huggingface/datasets/issues/359/events | https://github.com/huggingface/datasets/issues/359 | 653,656,279 | MDU6SXNzdWU2NTM2NTYyNzk= | 359 | ArrowBasedBuilder _prepare_split parse_schema breaks on nested structures | {
"avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4",
"events_url": "https://api.github.com/users/timothyjlaurent/events{/privacy}",
"followers_url": "https://api.github.com/users/timothyjlaurent/followers",
"following_url": "https://api.github.com/users/timothyjlaurent/following{/other_user}",... | [] | closed | false | null | [] | null | [
"Hi, it depends on what it is in your `dataset_builder.py` file. Can you share it?\r\n\r\nIf you are just loading `json` files, you can also directly use the `json` script (which will find the schema/features from your JSON structure):\r\n\r\n```python\r\nfrom nlp import load_dataset\r\nds = load_dataset(\"json\", ... | 2020-07-08T23:24:05Z | 2020-07-10T14:52:06Z | 2020-07-10T14:52:06Z | NONE | null | null | null | I tried using the Json dataloader to load some JSON lines files. but get an exception in the parse_schema function.
```
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-23-9aecfbee53bd> in <mo... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/359/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/359/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1719 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1719/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1719/comments | https://api.github.com/repos/huggingface/datasets/issues/1719/events | https://github.com/huggingface/datasets/pull/1719 | 783,557,542 | MDExOlB1bGxSZXF1ZXN0NTUyODk3MzY4 | 1,719 | Fix column list comparison in transmit format | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [] | 2021-01-11T17:23:56Z | 2021-01-11T18:45:03Z | 2021-01-11T18:45:02Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1719.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1719",
"merged_at": "2021-01-11T18:45:02Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1719.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | As noticed in #1718 the cache might not reload the cache files when new columns were added.
This is because of an issue in `transmit_format` where the column list comparison fails because the order was not deterministic. This causes the `transmit_format` to apply an unnecessary `set_format` transform with shuffled col... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1719/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1719/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/6113 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6113/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6113/comments | https://api.github.com/repos/huggingface/datasets/issues/6113/events | https://github.com/huggingface/datasets/issues/6113 | 1,833,854,030 | I_kwDODunzps5tTmRO | 6,113 | load_dataset() fails with streamlit caching inside docker | {
"avatar_url": "https://avatars.githubusercontent.com/u/987574?v=4",
"events_url": "https://api.github.com/users/fierval/events{/privacy}",
"followers_url": "https://api.github.com/users/fierval/followers",
"following_url": "https://api.github.com/users/fierval/following{/other_user}",
"gists_url": "https://... | [] | closed | false | null | [] | null | [
"Hi! This should be fixed in the latest (patch) release (run `pip install -U datasets` to install it). This behavior was due to a bug in our authentication logic."
] | 2023-08-02T20:20:26Z | 2023-08-21T18:18:27Z | 2023-08-21T18:18:27Z | NONE | null | null | null | ### Describe the bug
When calling `load_dataset` in a streamlit application running within a docker container, get a failure with the error message:
EmptyDatasetError: The directory at hf://datasets/fetch-rewards/inc-rings-2000@bea27cf60842b3641eae418f38864a2ec4cde684 doesn't contain any data files
Traceback:
Fil... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6113/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6113/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/6055 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6055/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6055/comments | https://api.github.com/repos/huggingface/datasets/issues/6055/events | https://github.com/huggingface/datasets/issues/6055 | 1,813,524,145 | I_kwDODunzps5sGC6x | 6,055 | Fix host URL in The Pile datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/7540752?v=4",
"events_url": "https://api.github.com/users/nickovchinnikov/events{/privacy}",
"followers_url": "https://api.github.com/users/nickovchinnikov/followers",
"following_url": "https://api.github.com/users/nickovchinnikov/following{/other_user}",... | [] | open | false | null | [] | null | [] | 2023-07-20T09:08:52Z | 2023-07-20T09:09:37Z | null | NONE | null | null | null | ### Describe the bug
In #3627 and #5543, you tried to fix the host URL in The Pile datasets. But both URLs are not working now:
`HTTPError: 404 Client Error: Not Found for URL: https://the-eye.eu/public/AI/pile_preliminary_components/PUBMED_title_abstracts_2019_baseline.jsonl.zst`
And
`ConnectTimeout: HTTPSCo... | {
"+1": 3,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 3,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6055/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6055/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/5875 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5875/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5875/comments | https://api.github.com/repos/huggingface/datasets/issues/5875/events | https://github.com/huggingface/datasets/issues/5875 | 1,716,770,394 | I_kwDODunzps5mU9Za | 5,875 | Why split slicing doesn't behave like list slicing ? | {
"avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4",
"events_url": "https://api.github.com/users/astariul/events{/privacy}",
"followers_url": "https://api.github.com/users/astariul/followers",
"following_url": "https://api.github.com/users/astariul/following{/other_user}",
"gists_url": "htt... | [
{
"color": "cfd3d7",
"default": true,
"description": "This issue or pull request already exists",
"id": 1935892865,
"name": "duplicate",
"node_id": "MDU6TGFiZWwxOTM1ODkyODY1",
"url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate"
}
] | open | false | null | [] | null | [
"A duplicate of https://github.com/huggingface/datasets/issues/1774"
] | 2023-05-19T07:21:10Z | 2023-05-23T16:02:14Z | null | NONE | null | null | null | ### Describe the bug
If I want to get the first 10 samples of my dataset, I can do :
```
ds = datasets.load_dataset('mnist', split='train[:10]')
```
But if I exceed the number of samples in the dataset, an exception is raised :
```
ds = datasets.load_dataset('mnist', split='train[:999999999]')
```
> V... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5875/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5875/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/1581 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1581/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1581/comments | https://api.github.com/repos/huggingface/datasets/issues/1581/events | https://github.com/huggingface/datasets/issues/1581 | 768,320,594 | MDU6SXNzdWU3NjgzMjA1OTQ= | 1,581 | Installing datasets and transformers in a tensorflow docker image throws Permission Error on 'import transformers' | {
"avatar_url": "https://avatars.githubusercontent.com/u/702586?v=4",
"events_url": "https://api.github.com/users/eduardofv/events{/privacy}",
"followers_url": "https://api.github.com/users/eduardofv/followers",
"following_url": "https://api.github.com/users/eduardofv/following{/other_user}",
"gists_url": "ht... | [] | closed | false | null | [] | null | [
"Thanks for reporting !\r\nYou can override the directory in which cache file are stored using for example\r\n```\r\nENV HF_HOME=\"/root/cache/hf_cache_home\"\r\n```\r\n\r\nThis way both `transformers` and `datasets` will use this directory instead of the default `.cache`",
"Great, thanks. I didn't see documentat... | 2020-12-16T00:02:21Z | 2021-06-17T15:40:45Z | 2021-06-17T15:40:45Z | NONE | null | null | null | I am using a docker container, based on latest tensorflow-gpu image, to run transformers and datasets (4.0.1 and 1.1.3 respectively - Dockerfile attached below). Importing transformers throws a Permission Error to access `/.cache`:
```
$ docker run --gpus=all --rm -it -u $(id -u):$(id -g) -v $(pwd)/data:/root/data ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1581/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1581/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/6378 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6378/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6378/comments | https://api.github.com/repos/huggingface/datasets/issues/6378/events | https://github.com/huggingface/datasets/pull/6378 | 1,973,942,770 | PR_kwDODunzps5eaqhv | 6,378 | Support pyarrow 14.0.0 | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... | 2023-11-02T10:25:10Z | 2023-11-02T15:24:28Z | 2023-11-02T15:15:44Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/6378.diff",
"html_url": "https://github.com/huggingface/datasets/pull/6378",
"merged_at": "2023-11-02T15:15:44Z",
"patch_url": "https://github.com/huggingface/datasets/pull/6378.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Support `pyarrow` 14.0.0.
Fix #6377 and fix #6374 (root cause).
This fix is analog to a previous one:
- #6175 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6378/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6378/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2587 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2587/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2587/comments | https://api.github.com/repos/huggingface/datasets/issues/2587/events | https://github.com/huggingface/datasets/pull/2587 | 936,771,339 | MDExOlB1bGxSZXF1ZXN0NjgzNDI5NjQy | 2,587 | Add aiohttp to tests extras require | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [] | 2021-07-05T07:14:01Z | 2021-07-05T09:04:38Z | 2021-07-05T09:04:38Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2587.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2587",
"merged_at": "2021-07-05T09:04:38Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2587.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Currently, none of the streaming tests are runned within our CI test suite, because the streaming tests require aiohttp and this is missing from our tests extras require dependencies.
Our CI test suite should be exhaustive and test all the library functionalities. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2587/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2587/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5931 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5931/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5931/comments | https://api.github.com/repos/huggingface/datasets/issues/5931/events | https://github.com/huggingface/datasets/issues/5931 | 1,745,408,784 | I_kwDODunzps5oCNMQ | 5,931 | `datasets.map` not reusing cached copy by default | {
"avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4",
"events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}",
"followers_url": "https://api.github.com/users/bhavitvyamalik/followers",
"following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}",
... | [] | closed | false | null | [] | null | [
"This can happen when a map transform cannot be hashed deterministically (e.g., an object referenced by the transform changes its state after the first call - an issue with fast tokenizers). The solution is to provide `cache_file_name` in the `map` call to check this file for the cached result instead of relying on... | 2023-06-07T09:03:33Z | 2023-06-21T16:15:40Z | 2023-06-21T16:15:40Z | CONTRIBUTOR | null | null | null | ### Describe the bug
When I load the dataset from local directory, it's cached copy is picked up after first time. However, for `map` operation, the operation is applied again and cached copy is not picked up. Is there any way to pick cached copy instead of processing it again? The only solution I could think of was... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5931/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5931/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1048 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1048/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1048/comments | https://api.github.com/repos/huggingface/datasets/issues/1048/events | https://github.com/huggingface/datasets/pull/1048 | 756,133,072 | MDExOlB1bGxSZXF1ZXN0NTMxNzI3MDA0 | 1,048 | Adding NCHLT dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4",
"events_url": "https://api.github.com/users/Narsil/events{/privacy}",
"followers_url": "https://api.github.com/users/Narsil/followers",
"following_url": "https://api.github.com/users/Narsil/following{/other_user}",
"gists_url": "https://api... | [] | closed | false | null | [] | null | [
"merging since the CI is fixed on master"
] | 2020-12-03T11:59:25Z | 2020-12-04T13:29:57Z | 2020-12-04T13:29:57Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1048.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1048",
"merged_at": "2020-12-04T13:29:56Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1048.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | https://repo.sadilar.org/handle/20.500.12185/7/discover?filtertype_0=database&filtertype_1=title&filter_relational_operator_1=contains&filter_relational_operator_0=equals&filter_1=&filter_0=Monolingual+Text+Corpora%3A+Annotated&filtertype=project&filter_relational_operator=equals&filter=NCHLT+Text+II | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1048/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1048/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4255 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4255/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4255/comments | https://api.github.com/repos/huggingface/datasets/issues/4255/events | https://github.com/huggingface/datasets/pull/4255 | 1,221,142,899 | PR_kwDODunzps43FHgR | 4,255 | No google drive URL for pubmed_qa | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._",
"CI is failing because some sections are missing in the dataset card, but this is unrelated to this PR - Merging !"
] | 2022-04-29T15:55:46Z | 2022-04-29T16:24:55Z | 2022-04-29T16:18:56Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4255.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4255",
"merged_at": "2022-04-29T16:18:56Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4255.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | I hosted the data files in https://huggingface.co/datasets/pubmed_qa. This is allowed because the data is under the MIT license.
cc @stas00 | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4255/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4255/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5930 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5930/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5930/comments | https://api.github.com/repos/huggingface/datasets/issues/5930/events | https://github.com/huggingface/datasets/issues/5930 | 1,745,184,395 | I_kwDODunzps5oBWaL | 5,930 | loading private custom dataset script - authentication error | {
"avatar_url": "https://avatars.githubusercontent.com/u/103381497?v=4",
"events_url": "https://api.github.com/users/flckv/events{/privacy}",
"followers_url": "https://api.github.com/users/flckv/followers",
"following_url": "https://api.github.com/users/flckv/following{/other_user}",
"gists_url": "https://api... | [] | closed | false | null | [] | null | [
"This issue seems to have been resolved, so I'm closing it."
] | 2023-06-07T06:58:23Z | 2023-06-15T14:49:21Z | 2023-06-15T14:49:20Z | NONE | null | null | null | ### Describe the bug
Train model with my custom dataset stored in HuggingFace and loaded with the loading script requires authentication but I am not sure how ?
I am logged in in the terminal, in the browser. I receive this error:
/python3.8/site-packages/datasets/utils/file_utils.py", line 566, in get_from... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5930/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5930/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/3681 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3681/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3681/comments | https://api.github.com/repos/huggingface/datasets/issues/3681/events | https://github.com/huggingface/datasets/pull/3681 | 1,124,237,458 | PR_kwDODunzps4yFcpM | 3,681 | Fix TestCommand to move dataset_infos instead of copying | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"All the datasets that are loaded normally with `load_dataset`, if `dataset_infos.json` exists, have this file in the importable directory. So it's fine if we copy the file instead of moving it but it's not a big deal.\r\n\r\nAny reason to prefer moving it rather than copying it ?",
"@lvwerra reported than when g... | 2022-02-04T14:01:52Z | 2023-09-24T10:00:11Z | 2023-09-24T09:59:55Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3681.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3681",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/3681.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3681"
} | Why do we copy instead of moving the file?
CC: @lhoestq @lvwerra | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3681/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3681/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3755 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3755/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3755/comments | https://api.github.com/repos/huggingface/datasets/issues/3755/events | https://github.com/huggingface/datasets/issues/3755 | 1,143,032,961 | I_kwDODunzps5EIUyB | 3,755 | Cannot preview dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/2518789?v=4",
"events_url": "https://api.github.com/users/frascuchon/events{/privacy}",
"followers_url": "https://api.github.com/users/frascuchon/followers",
"following_url": "https://api.github.com/users/frascuchon/following{/other_user}",
"gists_url":... | [
{
"color": "E5583E",
"default": false,
"description": "Related to the dataset viewer on huggingface.co",
"id": 3470211881,
"name": "dataset-viewer",
"node_id": "LA_kwDODunzps7O1zsp",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer"
}
] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"events_url": "https://api.github.com/users/severo/events{/privacy}",
"followers_url": "https://api.github.com/users/severo/followers",
"following_url": "https://api.github.com/users/severo/following{/other_user}",
"gists_url": "https://ap... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"events_url": "https://api.github.com/users/severo/events{/privacy}",
"followers_url": "https://api.github.com/users/severo/followers",
"following_url": "https://api.github.com/users/severo/following{/other_user}",
"gists_url... | null | [
"Thanks for reporting. The dataset viewer depends on some backend treatments, and for now, they might take some hours to get processed. We're working on improving it.",
"It has finally been processed. Thanks for the patience.",
"Thanks for the info @severo !"
] | 2022-02-18T13:06:45Z | 2022-02-19T14:30:28Z | 2022-02-18T15:41:33Z | NONE | null | null | null | ## Dataset viewer issue for '*rubrix/news*'
**Link:https://huggingface.co/datasets/rubrix/news** *link to the dataset viewer page*
Cannot see the dataset preview:
```
Status code: 400
Exception: Status400Error
Message: Not found. Cache is waiting to be refreshed.
```
Am I the one who added thi... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3755/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3755/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1129 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1129/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1129/comments | https://api.github.com/repos/huggingface/datasets/issues/1129/events | https://github.com/huggingface/datasets/pull/1129 | 757,255,492 | MDExOlB1bGxSZXF1ZXN0NTMyNjYxNzM2 | 1,129 | Adding initial version of cord-19 dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/5583410?v=4",
"events_url": "https://api.github.com/users/ggdupont/events{/privacy}",
"followers_url": "https://api.github.com/users/ggdupont/followers",
"following_url": "https://api.github.com/users/ggdupont/following{/other_user}",
"gists_url": "http... | [] | closed | false | null | [] | null | [
"Hi @ggdupont !\r\nHave you had a chance to take a look at my suggestions ?\r\nFeel free to ping me if you have questions or when you're ready for a review",
"> Hi @ggdupont !\r\n> Have you had a chance to take a look at my suggestions ?\r\n> Feel free to ping me if you have questions or when you're ready for a r... | 2020-12-04T17:03:17Z | 2021-02-09T10:22:35Z | 2021-02-09T10:18:06Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1129.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1129",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1129.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1129"
} | Initial version only reading the metadata in CSV.
### Checklist:
- [x] Create the dataset script /datasets/my_dataset/my_dataset.py using the template
- [x] Fill the _DESCRIPTION and _CITATION variables
- [x] Implement _infos(), _split_generators() and _generate_examples()
- [x] Make sure that the BUILDER_CONFIG... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1129/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1129/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/898 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/898/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/898/comments | https://api.github.com/repos/huggingface/datasets/issues/898/events | https://github.com/huggingface/datasets/pull/898 | 752,148,284 | MDExOlB1bGxSZXF1ZXN0NTI4NTI4MDY1 | 898 | Adding SQA dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4",
"events_url": "https://api.github.com/users/thomwolf/events{/privacy}",
"followers_url": "https://api.github.com/users/thomwolf/followers",
"following_url": "https://api.github.com/users/thomwolf/following{/other_user}",
"gists_url": "http... | [] | closed | false | null | [] | null | [
"This dataset seems to have around 1000 configs. Therefore when creating the dummy data we end up with hundreds of MB of dummy data which we don't want to add in the repo.\r\nLet's make this PR on hold for now and find a solution after the sprint of next week",
"Closing in favor of #1566 "
] | 2020-11-27T10:29:18Z | 2020-12-15T12:54:40Z | 2020-12-15T12:54:19Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/898.diff",
"html_url": "https://github.com/huggingface/datasets/pull/898",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/898.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/898"
} | As discussed in #880
Seems like automatic dummy-data generation doesn't work if the builder is a `ArrowBasedBuilder`, do you think you could take a look @lhoestq ? | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/898/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/898/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3105 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3105/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3105/comments | https://api.github.com/repos/huggingface/datasets/issues/3105/events | https://github.com/huggingface/datasets/issues/3105 | 1,029,098,843 | I_kwDODunzps49Vs1b | 3,105 | download_mode=`force_redownload` does not work on removed datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"events_url": "https://api.github.com/users/severo/events{/privacy}",
"followers_url": "https://api.github.com/users/severo/followers",
"following_url": "https://api.github.com/users/severo/following{/other_user}",
"gists_url": "https://ap... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
},
{
"color": "E5583E",
"default": false,
"descrip... | open | false | null | [] | null | [] | 2021-10-18T13:12:38Z | 2021-10-22T09:36:10Z | null | CONTRIBUTOR | null | null | null | ## Describe the bug
If a cached dataset is removed from the library, I don't see how to delete it programmatically. I thought that using `force_redownload` would try to refresh the cache, then raise an exception, but it reuses the cache instead.
## Steps to reproduce the bug
_requires to already have `wit` in ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3105/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3105/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3632 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3632/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3632/comments | https://api.github.com/repos/huggingface/datasets/issues/3632/events | https://github.com/huggingface/datasets/issues/3632 | 1,115,027,185 | I_kwDODunzps5Cdfbx | 3,632 | Adding CC-100: Monolingual Datasets from Web Crawl Data (Datasets links are invalid) | {
"avatar_url": "https://avatars.githubusercontent.com/u/55232459?v=4",
"events_url": "https://api.github.com/users/AnzorGozalishvili/events{/privacy}",
"followers_url": "https://api.github.com/users/AnzorGozalishvili/followers",
"following_url": "https://api.github.com/users/AnzorGozalishvili/following{/other_... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/o... | null | [
"Hi @AnzorGozalishvili,\r\n\r\nMaybe their site was temporarily down, but it seems to work fine now.\r\n\r\nCould you please try again and confirm if the problem persists? ",
"Hi @albertvillanova \r\nI checked and it works. \r\nIt seems that it was really temporarily down.\r\nThanks!"
] | 2022-01-26T13:35:37Z | 2022-02-10T06:58:11Z | 2022-02-10T06:58:11Z | CONTRIBUTOR | null | null | null | ## Describe the bug
The dataset links are no longer valid for CC-100. It seems that the website which was keeping these files are no longer accessible and therefore this dataset became unusable.
Check out the dataset [homepage](http://data.statmt.org/cc-100/) which isn't accessible.
Also the URLs for dataset file ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3632/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3632/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/3176 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3176/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3176/comments | https://api.github.com/repos/huggingface/datasets/issues/3176/events | https://github.com/huggingface/datasets/pull/3176 | 1,039,068,312 | PR_kwDODunzps4t00xS | 3,176 | OpenSLR dataset: update generate_examples to properly extract data for SLR83 | {
"avatar_url": "https://avatars.githubusercontent.com/u/4561309?v=4",
"events_url": "https://api.github.com/users/tyrius02/events{/privacy}",
"followers_url": "https://api.github.com/users/tyrius02/followers",
"following_url": "https://api.github.com/users/tyrius02/following{/other_user}",
"gists_url": "http... | [] | closed | false | null | [] | null | [
"Also fix #3125."
] | 2021-10-29T00:59:27Z | 2021-11-04T16:20:45Z | 2021-10-29T10:04:09Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3176.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3176",
"merged_at": "2021-10-29T10:04:09Z",
"patch_url": "https://github.com/huggingface/datasets/pull/3176.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Fixed #3168.
The SLR38 indices are CSV files and there wasn't any code in openslr.py to process these files properly. The end result was an empty table.
I've added code to properly process these CSV files. | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3176/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3176/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/6073 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6073/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6073/comments | https://api.github.com/repos/huggingface/datasets/issues/6073/events | https://github.com/huggingface/datasets/issues/6073 | 1,822,167,804 | I_kwDODunzps5snBL8 | 6,073 | version2.3.2 load_dataset()data_files can't include .xxxx in path | {
"avatar_url": "https://avatars.githubusercontent.com/u/45893496?v=4",
"events_url": "https://api.github.com/users/BUAAChuanWang/events{/privacy}",
"followers_url": "https://api.github.com/users/BUAAChuanWang/followers",
"following_url": "https://api.github.com/users/BUAAChuanWang/following{/other_user}",
"g... | [] | closed | false | null | [] | null | [
"Version 2.3.2 is over one year old, so please use the latest release (2.14.0) to get the expected behavior. Version 2.3.2 does not contain some fixes we made to fix resolving hidden files/directories (starting with a dot)."
] | 2023-07-26T11:09:31Z | 2023-08-29T15:53:59Z | 2023-08-29T15:53:59Z | NONE | null | null | null | ### Describe the bug
First, I cd workdir.
Then, I just use load_dataset("json", data_file={"train":"/a/b/c/.d/train/train.json", "test":"/a/b/c/.d/train/test.json"})
that couldn't work and
<FileNotFoundError: Unable to find
'/a/b/c/.d/train/train.jsonl' at
/a/b/c/.d/>
And I debug, it is fine in version2.1.2... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6073/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6073/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/5245 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5245/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5245/comments | https://api.github.com/repos/huggingface/datasets/issues/5245/events | https://github.com/huggingface/datasets/issues/5245 | 1,450,376,433 | I_kwDODunzps5Wcvzx | 5,245 | Unable to rename columns in streaming dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/9079808?v=4",
"events_url": "https://api.github.com/users/peregilk/events{/privacy}",
"followers_url": "https://api.github.com/users/peregilk/followers",
"following_url": "https://api.github.com/users/peregilk/following{/other_user}",
"gists_url": "http... | [] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4",
"events_url": "https://api.github.com/users/alvarobartt/events{/privacy}",
"followers_url": "https://api.github.com/users/alvarobartt/followers",
"following_url": "https://api.github.com/users/alvarobartt/following{/other_user}",
"gists_u... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4",
"events_url": "https://api.github.com/users/alvarobartt/events{/privacy}",
"followers_url": "https://api.github.com/users/alvarobartt/followers",
"following_url": "https://api.github.com/users/alvarobartt/following{/other_user}"... | null | [
"Hi @peregilk this bug is directly related to https://github.com/huggingface/datasets/issues/3888, and still not fixed... But I'll try to have a look!",
"Thanks @alvarobartt. It is great if you are able to fix it, but when reading the explanation it seems like it is possible to work around it.\r\n\r\nWe also trie... | 2022-11-15T21:04:41Z | 2022-11-28T12:53:24Z | 2022-11-28T12:53:24Z | NONE | null | null | null | ### Describe the bug
Trying to rename column in a streaming datasets, destroys the features object.
### Steps to reproduce the bug
The following code illustrates the error:
```
from datasets import load_dataset
dataset = load_dataset('mc4', 'en', streaming=True, split='train')
dataset.info.features
# {'text':... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5245/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5245/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/5739 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5739/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5739/comments | https://api.github.com/repos/huggingface/datasets/issues/5739/events | https://github.com/huggingface/datasets/issues/5739 | 1,663,762,901 | I_kwDODunzps5jKwHV | 5,739 | weird result during dataset split when data path starts with `/data` | {
"avatar_url": "https://avatars.githubusercontent.com/u/1772912?v=4",
"events_url": "https://api.github.com/users/ericxsun/events{/privacy}",
"followers_url": "https://api.github.com/users/ericxsun/followers",
"following_url": "https://api.github.com/users/ericxsun/following{/other_user}",
"gists_url": "http... | [] | open | false | null | [] | null | [
"Same problem.",
"hi! \r\nI think you can run python from `/data/train/raw/` directory and load dataset as `load_dataset(\"code_contests\")` to mitigate this issue as a workaround. \r\n@ericxsun Do you want to open a PR to fix the regex? As you already found the solution :) ",
"> hi! I think you can run python ... | 2023-04-12T04:51:35Z | 2023-04-21T14:20:59Z | null | NONE | null | null | null | ### Describe the bug
The regex defined here https://github.com/huggingface/datasets/blob/f2607935c4e45c70c44fcb698db0363ca7ba83d4/src/datasets/utils/py_utils.py#L158
will cause a weird result during dataset split when data path starts with `/data`
### Steps to reproduce the bug
1. clone dataset into local path
... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5739/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5739/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/5786 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5786/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5786/comments | https://api.github.com/repos/huggingface/datasets/issues/5786/events | https://github.com/huggingface/datasets/issues/5786 | 1,680,957,070 | I_kwDODunzps5kMV6O | 5,786 | Multiprocessing in a `filter` or `map` function with a Pytorch model | {
"avatar_url": "https://avatars.githubusercontent.com/u/44556846?v=4",
"events_url": "https://api.github.com/users/HugoLaurencon/events{/privacy}",
"followers_url": "https://api.github.com/users/HugoLaurencon/followers",
"following_url": "https://api.github.com/users/HugoLaurencon/following{/other_user}",
"g... | [] | closed | false | null | [] | null | [
"Hi ! PyTorch may hang when calling `load_state_dict()` in a subprocess. To fix that, set the multiprocessing start method to \"spawn\". Since `datasets` uses `multiprocess`, you should do:\r\n\r\n```python\r\n# Required to avoid issues with pytorch (otherwise hangs during load_state_dict in multiprocessing)\r\nimp... | 2023-04-24T10:38:07Z | 2023-05-30T09:56:30Z | 2023-04-24T10:43:58Z | MEMBER | null | null | null | ### Describe the bug
I am trying to use a Pytorch model loaded on CPUs with multiple processes with a `.map` or a `.filter` method.
Usually, when dealing with models that are non-pickable, creating a class such that the `map` function is the method `__call__`, and adding `reduce` helps to solve the problem.
Howe... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5786/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5786/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1286 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1286/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1286/comments | https://api.github.com/repos/huggingface/datasets/issues/1286/events | https://github.com/huggingface/datasets/issues/1286 | 759,291,509 | MDU6SXNzdWU3NTkyOTE1MDk= | 1,286 | [libprotobuf FATAL /sentencepiece/src/../third_party/protobuf-lite/google/protobuf/repeated_field.h:1505] CHECK failed: (index) >= (0): terminate called after throwing an instance of 'google::protobuf::FatalException' what(): CHECK failed: (index) >= (0): Aborted | {
"avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4",
"events_url": "https://api.github.com/users/rabeehk/events{/privacy}",
"followers_url": "https://api.github.com/users/rabeehk/followers",
"following_url": "https://api.github.com/users/rabeehk/following{/other_user}",
"gists_url": "https:/... | [] | closed | false | null | [] | null | [
"I remember also getting the same issue for several other translation datasets like all the iwslt2017 group, this is blokcing me and I really need to fix it and I was wondering if you have an idea on this. @lhoestq thanks,. ",
"maybe there is an empty line or something inside these datasets? could you tell me wh... | 2020-12-08T09:44:15Z | 2020-12-12T19:36:22Z | 2020-12-12T16:22:36Z | CONTRIBUTOR | null | null | null | Hi
I am getting this error when evaluating on wmt16-ro-en using finetune_trainer.py of huggingface repo. thank for your help
{'epoch': 20.0}
100%|████████████████████████████... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1286/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1286/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1409 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1409/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1409/comments | https://api.github.com/repos/huggingface/datasets/issues/1409/events | https://github.com/huggingface/datasets/pull/1409 | 760,593,932 | MDExOlB1bGxSZXF1ZXN0NTM1Mzk5OTI1 | 1,409 | Adding the ASSIN dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/5097052?v=4",
"events_url": "https://api.github.com/users/jonatasgrosman/events{/privacy}",
"followers_url": "https://api.github.com/users/jonatasgrosman/followers",
"following_url": "https://api.github.com/users/jonatasgrosman/following{/other_user}",
... | [] | closed | false | null | [] | null | [
"I wrongly commited data from another branch in this PR, I'll close this a reopen another PR with the fixed branch"
] | 2020-12-09T19:07:00Z | 2020-12-09T19:18:12Z | 2020-12-09T19:15:52Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1409.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1409",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1409.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1409"
} | Adding the ASSIN dataset, a Portuguese language dataset for Natural Language Inference and Semantic Similarity Scoring | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1409/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1409/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4172 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4172/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4172/comments | https://api.github.com/repos/huggingface/datasets/issues/4172/events | https://github.com/huggingface/datasets/pull/4172 | 1,204,433,160 | PR_kwDODunzps42O7LW | 4,172 | Update assin2 dataset_infos.json | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-04-14T11:53:06Z | 2022-04-15T14:47:42Z | 2022-04-15T14:41:22Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4172.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4172",
"merged_at": "2022-04-15T14:41:22Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4172.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Following comments in https://github.com/huggingface/datasets/issues/4003 we found that it was outdated and casing an error when loading the dataset | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4172/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4172/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/3338 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3338/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3338/comments | https://api.github.com/repos/huggingface/datasets/issues/3338/events | https://github.com/huggingface/datasets/pull/3338 | 1,066,371,235 | PR_kwDODunzps4vJRFM | 3,338 | [WIP] Add doctests for tutorials | {
"avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4",
"events_url": "https://api.github.com/users/stevhliu/events{/privacy}",
"followers_url": "https://api.github.com/users/stevhliu/followers",
"following_url": "https://api.github.com/users/stevhliu/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [
"I manage to remove the mentions of ellipsis in the code by launching the command as follows:\r\n\r\n```\r\npython -m doctest -v docs/source/load_hub.rst -o=ELLIPSIS\r\n```\r\n\r\nThe way you put your ellipsis will only work on mac, I've adapted it for linux as well with the following:\r\n\r\n```diff\r\n >>> fro... | 2021-11-29T18:40:46Z | 2023-05-05T17:18:20Z | 2023-05-05T17:18:15Z | MEMBER | null | 1 | {
"diff_url": "https://github.com/huggingface/datasets/pull/3338.diff",
"html_url": "https://github.com/huggingface/datasets/pull/3338",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/3338.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3338"
} | Opening a PR as discussed with @LysandreJik for some help with doctest issues. The goal is to add doctests for each of the tutorials in the documentation to make sure the code samples work as shown.
### Issues
A doctest has been added in the docstring of the `load_dataset_builder` function in `load.py` to handle ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/3338/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/3338/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/5482 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5482/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5482/comments | https://api.github.com/repos/huggingface/datasets/issues/5482/events | https://github.com/huggingface/datasets/issues/5482 | 1,560,853,137 | I_kwDODunzps5dCLqR | 5,482 | Reload features from Parquet metadata | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
},
{
"color": "BDE59C",
"default": fals... | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/6368040?v=4",
"events_url": "https://api.github.com/users/MFreidank/events{/privacy}",
"followers_url": "https://api.github.com/users/MFreidank/followers",
"following_url": "https://api.github.com/users/MFreidank/following{/other_user}",
"gists_url": "h... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/6368040?v=4",
"events_url": "https://api.github.com/users/MFreidank/events{/privacy}",
"followers_url": "https://api.github.com/users/MFreidank/followers",
"following_url": "https://api.github.com/users/MFreidank/following{/other_user}",
"... | null | [
"I'd be happy to have a look, if nobody else has started working on this yet @lhoestq. \r\n\r\nIt seems to me that for the `arrow` format features are currently attached as metadata [in `datasets.arrow_writer`](https://github.com/huggingface/datasets/blob/5f810b7011a8a4ab077a1847c024d2d9e267b065/src/datasets/arrow_... | 2023-01-28T13:12:31Z | 2023-02-12T15:57:02Z | 2023-02-12T15:57:02Z | MEMBER | null | null | null | The idea would be to allow this :
```python
ds.to_parquet("my_dataset/ds.parquet")
reloaded = load_dataset("my_dataset")
assert ds.features == reloaded.features
```
And it should also work with Image and Audio types (right now they're reloaded as a dict type)
This can be implemented by storing and reading th... | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5482/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5482/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1045 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1045/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1045/comments | https://api.github.com/repos/huggingface/datasets/issues/1045/events | https://github.com/huggingface/datasets/pull/1045 | 756,120,760 | MDExOlB1bGxSZXF1ZXN0NTMxNzE2NzIy | 1,045 | Add xitsonga ner corpus | {
"avatar_url": "https://avatars.githubusercontent.com/u/7923902?v=4",
"events_url": "https://api.github.com/users/yvonnegitau/events{/privacy}",
"followers_url": "https://api.github.com/users/yvonnegitau/followers",
"following_url": "https://api.github.com/users/yvonnegitau/following{/other_user}",
"gists_ur... | [] | closed | false | null | [] | null | [
"Look like this PR includes changes to many other files than the ones related to xitsonga NER.\r\nCould you create another branch and another PR please ?"
] | 2020-12-03T11:40:48Z | 2020-12-03T17:20:03Z | 2020-12-03T17:19:32Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1045.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1045",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/1045.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/1045"
} | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1045/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1045/timeline | null | null | true | |
https://api.github.com/repos/huggingface/datasets/issues/1387 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1387/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1387/comments | https://api.github.com/repos/huggingface/datasets/issues/1387/events | https://github.com/huggingface/datasets/pull/1387 | 760,368,355 | MDExOlB1bGxSZXF1ZXN0NTM1MjExODQ1 | 1,387 | Add LIAR dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/1206395?v=4",
"events_url": "https://api.github.com/users/hugoabonizio/events{/privacy}",
"followers_url": "https://api.github.com/users/hugoabonizio/followers",
"following_url": "https://api.github.com/users/hugoabonizio/following{/other_user}",
"gists... | [] | closed | false | null | [] | null | [
"@lhoestq done! The failing testes don't seem to be related, it seems to be a connection issue, if I understand it correctly.",
"merging since the CI is fixed on master"
] | 2020-12-09T14:16:55Z | 2020-12-14T18:06:43Z | 2020-12-14T16:23:59Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1387.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1387",
"merged_at": "2020-12-14T16:23:59Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1387.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Add LIAR dataset from [“Liar, Liar Pants on Fire”: A New Benchmark Dataset for Fake News Detection](https://www.aclweb.org/anthology/P17-2067/). | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1387/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1387/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4851 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4851/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4851/comments | https://api.github.com/repos/huggingface/datasets/issues/4851/events | https://github.com/huggingface/datasets/pull/4851 | 1,339,085,917 | PR_kwDODunzps49L6ee | 4,851 | Fix license tag and Source Data section in billsum dataset card | {
"avatar_url": "https://avatars.githubusercontent.com/u/8100?v=4",
"events_url": "https://api.github.com/users/kashif/events{/privacy}",
"followers_url": "https://api.github.com/users/kashif/followers",
"following_url": "https://api.github.com/users/kashif/following{/other_user}",
"gists_url": "https://api.g... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._",
"thanks @albertvillanova done thank you!"
] | 2022-08-15T14:37:00Z | 2022-08-22T13:56:24Z | 2022-08-22T13:40:59Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4851.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4851",
"merged_at": "2022-08-22T13:40:59Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4851.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Fixed the data source and license fields | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4851/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4851/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4427 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4427/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4427/comments | https://api.github.com/repos/huggingface/datasets/issues/4427/events | https://github.com/huggingface/datasets/pull/4427 | 1,253,959,313 | PR_kwDODunzps44vyGg | 4,427 | Add HF.co for PRs/Issues for specific datasets | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | 2022-05-31T14:31:21Z | 2022-06-01T12:37:42Z | 2022-06-01T12:29:02Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4427.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4427",
"merged_at": "2022-06-01T12:29:02Z",
"patch_url": "https://github.com/huggingface/datasets/pull/4427.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | As in https://github.com/huggingface/transformers/pull/17485, issues and PR for datasets under a namespace have to be on the HF Hub | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4427/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4427/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2553 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2553/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2553/comments | https://api.github.com/repos/huggingface/datasets/issues/2553/events | https://github.com/huggingface/datasets/issues/2553 | 931,365,926 | MDU6SXNzdWU5MzEzNjU5MjY= | 2,553 | load_dataset("web_nlg") NonMatchingChecksumError | {
"avatar_url": "https://avatars.githubusercontent.com/u/33730312?v=4",
"events_url": "https://api.github.com/users/alxthm/events{/privacy}",
"followers_url": "https://api.github.com/users/alxthm/followers",
"following_url": "https://api.github.com/users/alxthm/following{/other_user}",
"gists_url": "https://a... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists... | null | [
"Hi ! Thanks for reporting. This is due to the WebNLG repository that got updated today.\r\nI just pushed a fix at #2558 - this shouldn't happen anymore in the future.",
"This is fixed on `master` now :)\r\nWe'll do a new release soon !"
] | 2021-06-28T09:26:46Z | 2021-06-28T17:23:39Z | 2021-06-28T17:23:16Z | NONE | null | null | null | Hi! It seems the WebNLG dataset gives a NonMatchingChecksumError.
## Steps to reproduce the bug
```python
from datasets import load_dataset
dataset = load_dataset('web_nlg', name="release_v3.0_en", split="dev")
```
Gives
```
NonMatchingChecksumError: Checksums didn't match for dataset source files:
['h... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2553/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2553/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/4522 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4522/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4522/comments | https://api.github.com/repos/huggingface/datasets/issues/4522/events | https://github.com/huggingface/datasets/issues/4522 | 1,274,929,328 | I_kwDODunzps5L_eCw | 4,522 | Try to reduce the number of datasets that require manual download | {
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"events_url": "https://api.github.com/users/severo/events{/privacy}",
"followers_url": "https://api.github.com/users/severo/followers",
"following_url": "https://api.github.com/users/severo/following{/other_user}",
"gists_url": "https://ap... | [] | open | false | {
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
"gists_url"... | [
{
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"events_url": "https://api.github.com/users/mariosasko/events{/privacy}",
"followers_url": "https://api.github.com/users/mariosasko/followers",
"following_url": "https://api.github.com/users/mariosasko/following{/other_user}",
... | null | [] | 2022-06-17T11:42:03Z | 2022-06-17T11:52:48Z | null | CONTRIBUTOR | null | null | null | > Currently, 41 canonical datasets require manual download. I checked their scripts and I'm pretty sure this number can be reduced to ≈ 30 by not relying on bash scripts to download data, hosting data directly on the Hub when the license permits, etc. Then, we will mostly be left with datasets with restricted access, w... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4522/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4522/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/5678 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/5678/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/5678/comments | https://api.github.com/repos/huggingface/datasets/issues/5678/events | https://github.com/huggingface/datasets/issues/5678 | 1,645,018,359 | I_kwDODunzps5iDPz3 | 5,678 | Add support to create a Dataset from spark dataframe | {
"avatar_url": "https://avatars.githubusercontent.com/u/38018689?v=4",
"events_url": "https://api.github.com/users/lu-wang-dl/events{/privacy}",
"followers_url": "https://api.github.com/users/lu-wang-dl/followers",
"following_url": "https://api.github.com/users/lu-wang-dl/following{/other_user}",
"gists_url"... | [
{
"color": "a2eeef",
"default": true,
"description": "New feature or request",
"id": 1935892871,
"name": "enhancement",
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement"
}
] | closed | false | null | [] | null | [
"if i read spark Dataframe , got an error on multi-node Spark cluster.\r\nDid the Api (Dataset.from_spark) support Spark cluster, read dataframe and save_to_disk?\r\n\r\nError: \r\n_pickle.PicklingError: Could not serialize object: RuntimeError: It appears that you are attempting to reference SparkContext from a b... | 2023-03-29T04:36:28Z | 2023-07-21T14:15:38Z | 2023-07-21T14:15:38Z | NONE | null | null | null | ### Feature request
Add a new API `Dataset.from_spark` to create a Dataset from Spark DataFrame.
### Motivation
Spark is a distributed computing framework that can handle large datasets. By supporting loading Spark DataFrames directly into Hugging Face Datasets, we enable take the advantages of spark to processing t... | {
"+1": 2,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 2,
"url": "https://api.github.com/repos/huggingface/datasets/issues/5678/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/5678/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/6212 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6212/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6212/comments | https://api.github.com/repos/huggingface/datasets/issues/6212/events | https://github.com/huggingface/datasets/issues/6212 | 1,880,399,516 | I_kwDODunzps5wFJ6c | 6,212 | Tilde (~) is not supported for data_files | {
"avatar_url": "https://avatars.githubusercontent.com/u/128361578?v=4",
"events_url": "https://api.github.com/users/exs-avianello/events{/privacy}",
"followers_url": "https://api.github.com/users/exs-avianello/followers",
"following_url": "https://api.github.com/users/exs-avianello/following{/other_user}",
"... | [] | open | false | null | [] | null | [
"Hi @exs-avianello, is it really needed? Note you can alternatively use `pathlib.Path` among others as it follows:\r\n\r\n```python\r\nimport datasets\r\nfrom pathlib import Path\r\n\r\n# save a parquet file at ~/path/to/data.parquet\r\n\r\ndata_files = Path.home() / \"path/to/data.parquet\"\r\ndataset = datasets.l... | 2023-09-04T14:23:49Z | 2023-09-05T08:28:39Z | null | NONE | null | null | null | ### Describe the bug
Attempting to `load_dataset` from a path starting with `~` (as a shorthand for the user's home directory) seems not to be fully working - at least as far as the `parquet` dataset builder is concerned.
(the same file can be loaded correctly if providing its absolute path instead)
I think that... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6212/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6212/timeline | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/1040 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1040/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1040/comments | https://api.github.com/repos/huggingface/datasets/issues/1040/events | https://github.com/huggingface/datasets/pull/1040 | 756,050,387 | MDExOlB1bGxSZXF1ZXN0NTMxNjU4MTU3 | 1,040 | Add UN Universal Declaration of Human Rights (UDHR) | {
"avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4",
"events_url": "https://api.github.com/users/joeddav/events{/privacy}",
"followers_url": "https://api.github.com/users/joeddav/followers",
"following_url": "https://api.github.com/users/joeddav/following{/other_user}",
"gists_url": "https:/... | [] | closed | false | null | [] | null | [] | 2020-12-03T10:04:58Z | 2020-12-03T19:20:15Z | 2020-12-03T19:20:11Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1040.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1040",
"merged_at": "2020-12-03T19:20:11Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1040.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Universal declaration of human rights with translations in 464 languages and dialects.
- UN page: https://www.ohchr.org/EN/UDHR/Pages/UDHRIndex.aspx
- Raw data source: https://unicode.org/udhr/index.html
Each instance of the dataset corresponds to one translation of the document. Since there's only one instance ... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1040/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1040/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1162 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1162/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1162/comments | https://api.github.com/repos/huggingface/datasets/issues/1162/events | https://github.com/huggingface/datasets/pull/1162 | 757,707,085 | MDExOlB1bGxSZXF1ZXN0NTMzMDM1MzEw | 1,162 | Add Mocha dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/46804938?v=4",
"events_url": "https://api.github.com/users/mattbui/events{/privacy}",
"followers_url": "https://api.github.com/users/mattbui/followers",
"following_url": "https://api.github.com/users/mattbui/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [] | 2020-12-05T15:45:14Z | 2020-12-07T10:09:39Z | 2020-12-07T10:09:39Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1162.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1162",
"merged_at": "2020-12-07T10:09:39Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1162.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | More information: https://allennlp.org/mocha | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1162/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1162/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/882 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/882/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/882/comments | https://api.github.com/repos/huggingface/datasets/issues/882/events | https://github.com/huggingface/datasets/pull/882 | 749,662,188 | MDExOlB1bGxSZXF1ZXN0NTI2NDQyMjA2 | 882 | Update README.md | {
"avatar_url": "https://avatars.githubusercontent.com/u/32997732?v=4",
"events_url": "https://api.github.com/users/vaibhavad/events{/privacy}",
"followers_url": "https://api.github.com/users/vaibhavad/followers",
"following_url": "https://api.github.com/users/vaibhavad/following{/other_user}",
"gists_url": "... | [] | closed | false | null | [] | null | [] | 2020-11-24T12:23:52Z | 2021-01-29T10:41:07Z | 2021-01-29T10:41:07Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/882.diff",
"html_url": "https://github.com/huggingface/datasets/pull/882",
"merged_at": "2021-01-29T10:41:06Z",
"patch_url": "https://github.com/huggingface/datasets/pull/882.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/882... | "no label" is "-" in the original dataset but "-1" in Huggingface distribution. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/882/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/882/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/2119 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2119/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2119/comments | https://api.github.com/repos/huggingface/datasets/issues/2119/events | https://github.com/huggingface/datasets/pull/2119 | 841,567,199 | MDExOlB1bGxSZXF1ZXN0NjAxMjg2MjIy | 2,119 | copy.deepcopy os.environ instead of copy | {
"avatar_url": "https://avatars.githubusercontent.com/u/5506053?v=4",
"events_url": "https://api.github.com/users/NihalHarish/events{/privacy}",
"followers_url": "https://api.github.com/users/NihalHarish/followers",
"following_url": "https://api.github.com/users/NihalHarish/following{/other_user}",
"gists_ur... | [] | closed | false | null | [] | null | [] | 2021-03-26T03:58:38Z | 2021-03-26T15:13:52Z | 2021-03-26T15:13:52Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2119.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2119",
"merged_at": "2021-03-26T15:13:52Z",
"patch_url": "https://github.com/huggingface/datasets/pull/2119.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Fixes: https://github.com/huggingface/datasets/issues/2115
- bug fix: using envrion.copy() returns a dict.
- using deepcopy(environ) returns an `_environ` object
- Changing the datatype of the _environ object can break code, if subsequent libraries perform operations using apis exclusive to the environ object, lik... | {
"+1": 2,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 2,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2119/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2119/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/6423 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6423/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6423/comments | https://api.github.com/repos/huggingface/datasets/issues/6423/events | https://github.com/huggingface/datasets/pull/6423 | 1,994,946,847 | PR_kwDODunzps5fhzD6 | 6,423 | Fix conda release by adding pyarrow-hotfix dependency | {
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"events_url": "https://api.github.com/users/albertvillanova/events{/privacy}",
"followers_url": "https://api.github.com/users/albertvillanova/followers",
"following_url": "https://api.github.com/users/albertvillanova/following{/other_user}",... | [] | closed | false | null | [] | null | [
"_The documentation is not available anymore as the PR was closed or merged._",
"<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea... | 2023-11-15T14:57:12Z | 2023-11-15T17:15:33Z | 2023-11-15T17:09:24Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/6423.diff",
"html_url": "https://github.com/huggingface/datasets/pull/6423",
"merged_at": "2023-11-15T17:09:24Z",
"patch_url": "https://github.com/huggingface/datasets/pull/6423.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Fix conda release by adding pyarrow-hotfix dependency.
Note that conda release failed in latest 2.14.7 release: https://github.com/huggingface/datasets/actions/runs/6874667214/job/18696761723
```
Traceback (most recent call last):
File "/usr/share/miniconda/envs/build-datasets/conda-bld/datasets_1700036460222/t... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6423/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6423/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4177 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4177/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4177/comments | https://api.github.com/repos/huggingface/datasets/issues/4177/events | https://github.com/huggingface/datasets/pull/4177 | 1,207,535,920 | PR_kwDODunzps42Yxca | 4,177 | Adding missing subsets to the `SemEval-2018 Task 1` dataset | {
"avatar_url": "https://avatars.githubusercontent.com/u/11460267?v=4",
"events_url": "https://api.github.com/users/micahcarroll/events{/privacy}",
"followers_url": "https://api.github.com/users/micahcarroll/followers",
"following_url": "https://api.github.com/users/micahcarroll/following{/other_user}",
"gist... | [
{
"color": "0e8a16",
"default": false,
"description": "Contribution to a dataset script",
"id": 4564477500,
"name": "dataset contribution",
"node_id": "LA_kwDODunzps8AAAABEBBmPA",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution"
}
] | open | false | null | [] | null | [
"Datasets are not tracked in this repository anymore. You should move this PR to the [discussions page of this dataset](https://huggingface.co/datasets/sem_eval_2018_task_1/discussions)"
] | 2022-04-18T22:59:30Z | 2022-10-05T10:38:16Z | null | NONE | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4177.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4177",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/4177.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4177"
} | This dataset for the [1st task of SemEval-2018](https://competitions.codalab.org/competitions/17751) competition was missing all subtasks except for subtask 5. I added another two subtasks (subtask 1 and 2), which are each comprised of 12 additional data subsets: for each language in En, Es, Ar, there are 4 datasets, b... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4177/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4177/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/262 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/262/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/262/comments | https://api.github.com/repos/huggingface/datasets/issues/262/events | https://github.com/huggingface/datasets/pull/262 | 636,702,849 | MDExOlB1bGxSZXF1ZXN0NDMyODI3Mzcz | 262 | Add new dataset ANLI Round 1 | {
"avatar_url": "https://avatars.githubusercontent.com/u/11016329?v=4",
"events_url": "https://api.github.com/users/easonnie/events{/privacy}",
"followers_url": "https://api.github.com/users/easonnie/followers",
"following_url": "https://api.github.com/users/easonnie/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [
"Hello ! Thanks for adding this one :)\r\n\r\nThis looks great, you just have to do the last steps to make the CI pass.\r\nI can see that two things are missing:\r\n1. the dummy data that is used to test that the script is working as expected\r\n2. the json file with all the infos about the dataset\r\n\r\nYou can s... | 2020-06-11T04:14:57Z | 2020-06-12T22:03:03Z | 2020-06-12T22:03:03Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/262.diff",
"html_url": "https://github.com/huggingface/datasets/pull/262",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/262.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/262"
} | Adding new dataset [ANLI](https://github.com/facebookresearch/anli/).
I'm not familiar with how to add new dataset. Let me know if there is any issue. I only include round 1 data here. There will be round 2, round 3 and more in the future with potentially different format. I think it will be better to separate them. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/262/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/262/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/951 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/951/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/951/comments | https://api.github.com/repos/huggingface/datasets/issues/951/events | https://github.com/huggingface/datasets/pull/951 | 754,349,979 | MDExOlB1bGxSZXF1ZXN0NTMwMjY1MTY0 | 951 | Prachathai67k | {
"avatar_url": "https://avatars.githubusercontent.com/u/15519308?v=4",
"events_url": "https://api.github.com/users/cstorm125/events{/privacy}",
"followers_url": "https://api.github.com/users/cstorm125/followers",
"following_url": "https://api.github.com/users/cstorm125/following{/other_user}",
"gists_url": "... | [] | closed | false | null | [] | null | [
"Wrongly branching from existing branch of wisesight_sentiment. Closing and opening another one specifically for prachathai67k"
] | 2020-12-01T12:21:52Z | 2020-12-01T12:29:53Z | 2020-12-01T12:28:26Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/951.diff",
"html_url": "https://github.com/huggingface/datasets/pull/951",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/951.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/951"
} | Add `prachathai-67k`: News Article Corpus and Multi-label Text Classificdation from Prachathai.com
The `prachathai-67k` dataset was scraped from the news site [Prachathai](prachathai.com). We filtered out those articles with less than 500 characters of body text, mostly images and cartoons. It contains 67,889 articl... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/951/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/951/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4798 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4798/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4798/comments | https://api.github.com/repos/huggingface/datasets/issues/4798/events | https://github.com/huggingface/datasets/pull/4798 | 1,330,699,942 | PR_kwDODunzps48wVEG | 4,798 | Shard generator | {
"avatar_url": "https://avatars.githubusercontent.com/u/43296932?v=4",
"events_url": "https://api.github.com/users/marianna13/events{/privacy}",
"followers_url": "https://api.github.com/users/marianna13/followers",
"following_url": "https://api.github.com/users/marianna13/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [
"Hi, thanks!\r\n\r\n> I was using Hugging Face datasets to process some very large datasets and found that it would be quite handy to have a feature that will allow to \"split\" these large datasets into chunks with equal size\r\n\r\n`map`, the method we use for processing in `datasets`, already does that if `batch... | 2022-08-06T09:14:06Z | 2022-10-03T15:35:10Z | 2022-10-03T15:35:10Z | NONE | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/4798.diff",
"html_url": "https://github.com/huggingface/datasets/pull/4798",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/4798.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4798"
} | Hi everyone! I was using Hugging Face datasets to process some very large datasets and found that it would be quite handy to have a feature that will allow to "split" these large datasets into chunks with equal size. Even better - be able to run through these chunks one by one in simple and convenient way. So I decided... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4798/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4798/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1741 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1741/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1741/comments | https://api.github.com/repos/huggingface/datasets/issues/1741/events | https://github.com/huggingface/datasets/issues/1741 | 787,327,060 | MDU6SXNzdWU3ODczMjcwNjA= | 1,741 | error when run fine_tuning on text_classification | {
"avatar_url": "https://avatars.githubusercontent.com/u/43234824?v=4",
"events_url": "https://api.github.com/users/XiaoYang66/events{/privacy}",
"followers_url": "https://api.github.com/users/XiaoYang66/followers",
"following_url": "https://api.github.com/users/XiaoYang66/following{/other_user}",
"gists_url"... | [] | closed | false | null | [] | null | [
"none"
] | 2021-01-16T02:23:19Z | 2021-01-16T02:39:28Z | 2021-01-16T02:39:18Z | NONE | null | null | null | dataset:sem_eval_2014_task_1
pretrained_model:bert-base-uncased
error description:
when i use these resoruce to train fine_tuning a text_classification on sem_eval_2014_task_1,there always be some problem(when i use other dataset ,there exist the error too). And i followed the colab code (url:https://colab.researc... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1741/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1741/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/216 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/216/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/216/comments | https://api.github.com/repos/huggingface/datasets/issues/216/events | https://github.com/huggingface/datasets/issues/216 | 626,896,890 | MDU6SXNzdWU2MjY4OTY4OTA= | 216 | ❓ How to get ROUGE-2 with the ROUGE metric ? | {
"avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4",
"events_url": "https://api.github.com/users/astariul/events{/privacy}",
"followers_url": "https://api.github.com/users/astariul/followers",
"following_url": "https://api.github.com/users/astariul/following{/other_user}",
"gists_url": "htt... | [] | closed | false | null | [] | null | [
"ROUGE-1 and ROUGE-L shouldn't return the same thing. This is weird",
"For the rouge2 metric you can do\r\n\r\n```python\r\nrouge = nlp.load_metric('rouge')\r\nwith open(\"pred.txt\") as p, open(\"ref.txt\") as g:\r\n for lp, lg in zip(p, g):\r\n rouge.add(lp, lg)\r\nscore = rouge.compute(rouge_types=[\... | 2020-05-28T23:47:32Z | 2020-06-01T00:04:35Z | 2020-06-01T00:04:35Z | NONE | null | null | null | I'm trying to use ROUGE metric, but I don't know how to get the ROUGE-2 metric.
---
I compute scores with :
```python
import nlp
rouge = nlp.load_metric('rouge')
with open("pred.txt") as p, open("ref.txt") as g:
for lp, lg in zip(p, g):
rouge.add([lp], [lg])
score = rouge.compute()
```
... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/216/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/216/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/2395 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/2395/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/2395/comments | https://api.github.com/repos/huggingface/datasets/issues/2395/events | https://github.com/huggingface/datasets/pull/2395 | 898,762,730 | MDExOlB1bGxSZXF1ZXN0NjUwNTk3NjI0 | 2,395 | `pretty_name` for dataset in YAML tags | {
"avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4",
"events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}",
"followers_url": "https://api.github.com/users/bhavitvyamalik/followers",
"following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}",
... | [
{
"color": "0e8a16",
"default": false,
"description": "Contribution to a dataset script",
"id": 4564477500,
"name": "dataset contribution",
"node_id": "LA_kwDODunzps8AAAABEBBmPA",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution"
}
] | closed | false | null | [] | null | [
"Initially I removed the ` - ` since there was only one `pretty_name` per config but turns out it was breaking here in `from_yaml_string`https://github.com/huggingface/datasets/blob/74751e3f98c74d22c48c6beb1fab0c13b5dfd075/src/datasets/utils/metadata.py#L197 in `/utils/metadata.py`",
"@lhoestq I guess this will a... | 2021-05-22T09:24:45Z | 2022-09-23T13:29:14Z | 2022-09-23T13:29:13Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/2395.diff",
"html_url": "https://github.com/huggingface/datasets/pull/2395",
"merged_at": null,
"patch_url": "https://github.com/huggingface/datasets/pull/2395.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/2395"
} | I'm updating `pretty_name` for datasets in YAML tags as discussed with @lhoestq. Here are the first 10, please let me know if they're looking good.
If dataset has 1 config, I've added `pretty_name` as `config_name: full_name_of_dataset` as config names were `plain_text`, `default`, `squad` etc (not so important in t... | {
"+1": 1,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 1,
"url": "https://api.github.com/repos/huggingface/datasets/issues/2395/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/2395/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/6363 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/6363/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/6363/comments | https://api.github.com/repos/huggingface/datasets/issues/6363/events | https://github.com/huggingface/datasets/issues/6363 | 1,968,891,277 | I_kwDODunzps51WuWN | 6,363 | dataset.transform() hangs indefinitely while finetuning the stable diffusion XL | {
"avatar_url": "https://avatars.githubusercontent.com/u/10846405?v=4",
"events_url": "https://api.github.com/users/bhosalems/events{/privacy}",
"followers_url": "https://api.github.com/users/bhosalems/followers",
"following_url": "https://api.github.com/users/bhosalems/following{/other_user}",
"gists_url": "... | [] | closed | false | null | [] | null | [
"I think the code hangs on the `accelerator.main_process_first()` context manager exit. To verify this, you can append a print statement to the end of the `accelerator.main_process_first()` block. \r\n\r\n\r\nIf the problem is in `with_transform`, it would help if you could share the error stack trace printed when... | 2023-10-30T17:34:05Z | 2023-11-22T00:29:21Z | 2023-11-22T00:29:21Z | NONE | null | null | null | ### Describe the bug
Multi-GPU fine-tuning the stable diffusion X by following https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/README_sdxl.md hangs indefinitely.
### Steps to reproduce the bug
accelerate launch train_text_to_image_sdxl.py --pretrained_model_name_or_path=$MODEL_NAME --... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/6363/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/6363/timeline | null | completed | false |
https://api.github.com/repos/huggingface/datasets/issues/1702 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1702/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1702/comments | https://api.github.com/repos/huggingface/datasets/issues/1702/events | https://github.com/huggingface/datasets/pull/1702 | 781,383,277 | MDExOlB1bGxSZXF1ZXN0NTUxMTE2NDc0 | 1,702 | Fix importlib metdata import in py38 | {
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"events_url": "https://api.github.com/users/lhoestq/events{/privacy}",
"followers_url": "https://api.github.com/users/lhoestq/followers",
"following_url": "https://api.github.com/users/lhoestq/following{/other_user}",
"gists_url": "https:... | [] | closed | false | null | [] | null | [] | 2021-01-07T15:10:30Z | 2021-01-08T10:47:15Z | 2021-01-08T10:47:15Z | MEMBER | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1702.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1702",
"merged_at": "2021-01-08T10:47:14Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1702.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | In Python 3.8 there's no need to install `importlib_metadata` since it already exists as `importlib.metadata` in the standard lib. | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1702/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1702/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/1056 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/1056/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/1056/comments | https://api.github.com/repos/huggingface/datasets/issues/1056/events | https://github.com/huggingface/datasets/pull/1056 | 756,309,828 | MDExOlB1bGxSZXF1ZXN0NTMxODc1MjA2 | 1,056 | Add deal_or_no_dialog | {
"avatar_url": "https://avatars.githubusercontent.com/u/28675016?v=4",
"events_url": "https://api.github.com/users/moussaKam/events{/privacy}",
"followers_url": "https://api.github.com/users/moussaKam/followers",
"following_url": "https://api.github.com/users/moussaKam/following{/other_user}",
"gists_url": "... | [] | closed | false | null | [] | null | [] | 2020-12-03T15:38:07Z | 2020-12-03T18:13:45Z | 2020-12-03T18:13:45Z | CONTRIBUTOR | null | 0 | {
"diff_url": "https://github.com/huggingface/datasets/pull/1056.diff",
"html_url": "https://github.com/huggingface/datasets/pull/1056",
"merged_at": "2020-12-03T18:13:45Z",
"patch_url": "https://github.com/huggingface/datasets/pull/1056.patch",
"url": "https://api.github.com/repos/huggingface/datasets/pulls/... | Add deal_or_no_dialog Dataset
github: https://github.com/facebookresearch/end-to-end-negotiator
Paper: [Deal or No Deal? End-to-End Learning for Negotiation Dialogues](https://arxiv.org/abs/1706.05125) | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/1056/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/1056/timeline | null | null | true |
https://api.github.com/repos/huggingface/datasets/issues/4149 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/4149/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/4149/comments | https://api.github.com/repos/huggingface/datasets/issues/4149/events | https://github.com/huggingface/datasets/issues/4149 | 1,201,389,221 | I_kwDODunzps5Hm76l | 4,149 | load_dataset for winoground returning decoding error | {
"avatar_url": "https://avatars.githubusercontent.com/u/4686956?v=4",
"events_url": "https://api.github.com/users/odellus/events{/privacy}",
"followers_url": "https://api.github.com/users/odellus/followers",
"following_url": "https://api.github.com/users/odellus/following{/other_user}",
"gists_url": "https:/... | [
{
"color": "d73a4a",
"default": true,
"description": "Something isn't working",
"id": 1935892857,
"name": "bug",
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug"
}
] | closed | false | null | [] | null | [
"I thought I had fixed it with this after some helpful hints from @severo\r\n```python\r\nimport datasets \r\ntoken = 'hf_XXXXX'\r\ndataset = datasets.load_dataset(\r\n 'facebook/winoground', \r\n name='facebook--winoground', \r\n split='train', \r\n streaming=True,\r\n use_auth_token=token,\r\n)\r\n... | 2022-04-12T08:16:16Z | 2022-05-04T23:40:38Z | 2022-05-04T23:40:38Z | CONTRIBUTOR | null | null | null | ## Describe the bug
I am trying to use datasets to load winoground and I'm getting a JSON decoding error.
## Steps to reproduce the bug
```python
from datasets import load_dataset
token = 'hf_XXXXX' # my HF access token
datasets = load_dataset('facebook/winoground', use_auth_token=token)
```
## Expected res... | {
"+1": 0,
"-1": 0,
"confused": 0,
"eyes": 0,
"heart": 0,
"hooray": 0,
"laugh": 0,
"rocket": 0,
"total_count": 0,
"url": "https://api.github.com/repos/huggingface/datasets/issues/4149/reactions"
} | https://api.github.com/repos/huggingface/datasets/issues/4149/timeline | null | completed | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.