url stringlengths 58 61 | repository_url stringclasses 1
value | labels_url stringlengths 72 75 | comments_url stringlengths 67 70 | events_url stringlengths 65 68 | html_url stringlengths 46 51 | id int64 599M 1.07B | node_id stringlengths 18 32 | number int64 1 3.39k | title stringlengths 1 276 | user dict | labels list | state stringclasses 1
value | locked bool 1
class | assignee dict | assignees list | milestone dict | comments list | created_at int64 1,587B 1,639B | updated_at int64 1,587B 1,639B | closed_at int64 1,587B 1,639B | author_association stringclasses 3
values | active_lock_reason null | body stringlengths 0 228k ⌀ | reactions dict | timeline_url stringlengths 67 70 | performed_via_github_app null | draft bool 2
classes | pull_request dict | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/huggingface/datasets/issues/3169 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3169/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3169/comments | https://api.github.com/repos/huggingface/datasets/issues/3169/events | https://github.com/huggingface/datasets/pull/3169 | 1,036,773,357 | PR_kwDODunzps4ttYmZ | 3,169 | Configurable max filename length in file locks | {
"login": "lmmx",
"id": 2979452,
"node_id": "MDQ6VXNlcjI5Nzk0NTI=",
"avatar_url": "https://avatars.githubusercontent.com/u/2979452?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lmmx",
"html_url": "https://github.com/lmmx",
"followers_url": "https://api.github.com/users/lmmx/followers",
... | [] | closed | false | null | [] | null | [
"I've also added environment variable configuration so that this can be configured once per machine (e.g. in a `.bashrc` file), as is already done for a few other config variables here.",
"Cancelling PR in favour of @mariosasko's in #3173"
] | 1,635,285,175,000 | 1,635,437,654,000 | 1,635,437,653,000 | NONE | null | Resolve #2924 (https://github.com/huggingface/datasets/issues/2924#issuecomment-952330956) wherein the assumption of file lock maximum filename length to be 255 raises an OSError on encrypted drives (ecryptFS on Linux uses part of the lower filename, reducing the maximum filename size to 143). Allowing this limit to be... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3169/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3169/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3169",
"html_url": "https://github.com/huggingface/datasets/pull/3169",
"diff_url": "https://github.com/huggingface/datasets/pull/3169.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3169.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3168 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3168/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3168/comments | https://api.github.com/repos/huggingface/datasets/issues/3168/events | https://github.com/huggingface/datasets/issues/3168 | 1,036,673,263 | I_kwDODunzps49ymDv | 3,168 | OpenSLR/83 is empty | {
"login": "tyrius02",
"id": 4561309,
"node_id": "MDQ6VXNlcjQ1NjEzMDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4561309?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tyrius02",
"html_url": "https://github.com/tyrius02",
"followers_url": "https://api.github.com/users/tyriu... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "tyrius02",
"id": 4561309,
"node_id": "MDQ6VXNlcjQ1NjEzMDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4561309?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tyrius02",
"html_url": "https://github.com/tyrius02",
"followers_url": "https://api.github.com/users/tyriu... | [
{
"login": "tyrius02",
"id": 4561309,
"node_id": "MDQ6VXNlcjQ1NjEzMDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4561309?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tyrius02",
"html_url": "https://github.com/tyrius02",
"followers_url": "https://api.gi... | null | [
"Hi @tyrius02, thanks for reporting. I see you self-assigned this issue: are you working on this?",
"@albertvillanova Yes. Figured I introduced the broken config, I should fix it too.\r\n\r\nI've got it working, but I'm struggling with one of the tests. I've started a PR so I/we can work through it.",
"Looks li... | 1,635,277,341,000 | 1,635,501,849,000 | 1,635,501,849,000 | CONTRIBUTOR | null | ## Describe the bug
As the summary says, openslr / SLR83 / train is empty.
The dataset returned after loading indicates there are **zero** rows. The correct number should be **17877**.
## Steps to reproduce the bug
```python
import datasets
datasets.load_dataset('openslr', 'SLR83')
```
## Expected resul... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3168/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3168/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3167 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3167/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3167/comments | https://api.github.com/repos/huggingface/datasets/issues/3167/events | https://github.com/huggingface/datasets/issues/3167 | 1,036,488,992 | I_kwDODunzps49x5Eg | 3,167 | bookcorpusopen no longer works | {
"login": "lucadiliello",
"id": 23355969,
"node_id": "MDQ6VXNlcjIzMzU1OTY5",
"avatar_url": "https://avatars.githubusercontent.com/u/23355969?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lucadiliello",
"html_url": "https://github.com/lucadiliello",
"followers_url": "https://api.github.c... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"Hi ! Thanks for reporting :) I think #3280 should fix this",
"I tried with the latest changes from #3280 on google colab and it worked fine :)\r\nWe'll do a new release soon, in the meantime you can use the updated version with:\r\n```python\r\nload_dataset(\"bookcorpusopen\", revision=\"master\")\r\n```",
"Fi... | 1,635,264,375,000 | 1,637,164,426,000 | 1,637,164,426,000 | CONTRIBUTOR | null | ## Describe the bug
When using the latest version of datasets (1.14.0), I cannot use the `bookcorpusopen` dataset. The process blocks always around `9924 examples [00:06, 1439.61 examples/s]` when preparing the dataset. I also noticed that after half an hour the process is automatically killed because of the RAM usa... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3167/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3167/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3166 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3166/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3166/comments | https://api.github.com/repos/huggingface/datasets/issues/3166/events | https://github.com/huggingface/datasets/pull/3166 | 1,036,450,283 | PR_kwDODunzps4tsVQJ | 3,166 | Deprecate prepare_module | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"Sounds good, thanks !"
] | 1,635,262,104,000 | 1,636,104,457,000 | 1,636,104,456,000 | MEMBER | null | In version 1.13, `prepare_module` was deprecated.
This PR adds a deprecation warning and removes it from all the library, using `dataset_module_factory` or `metric_module_factory` instead.
Fix #3165. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3166/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3166/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3166",
"html_url": "https://github.com/huggingface/datasets/pull/3166",
"diff_url": "https://github.com/huggingface/datasets/pull/3166.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3166.patch",
"merged_at": 1636104456000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3165 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3165/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3165/comments | https://api.github.com/repos/huggingface/datasets/issues/3165/events | https://github.com/huggingface/datasets/issues/3165 | 1,036,448,998 | I_kwDODunzps49xvTm | 3,165 | Deprecate prepare_module | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,635,262,035,000 | 1,636,104,456,000 | 1,636,104,456,000 | MEMBER | null | In version 1.13, `prepare_module` was deprecated.
Add deprecation warning and remove its usage from all the library. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3165/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3165/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3164 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3164/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3164/comments | https://api.github.com/repos/huggingface/datasets/issues/3164/events | https://github.com/huggingface/datasets/issues/3164 | 1,035,662,830 | I_kwDODunzps49uvXu | 3,164 | Add raw data files to the Hub with GitHub LFS for canonical dataset | {
"login": "zlucia",
"id": 40370937,
"node_id": "MDQ6VXNlcjQwMzcwOTM3",
"avatar_url": "https://avatars.githubusercontent.com/u/40370937?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zlucia",
"html_url": "https://github.com/zlucia",
"followers_url": "https://api.github.com/users/zlucia/fo... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Hi @zlucia, I would actually suggest hosting the dataset as a huggingface.co-hosted dataset.\r\n\r\nThe only difference with a \"canonical\"/legacy dataset is that it's nested under an organization (here `stanford` or `stanfordnlp` for instance – completely up to you) but then you can upload your data using git-lf... | 1,635,204,501,000 | 1,635,623,691,000 | 1,635,623,691,000 | NONE | null | I'm interested in sharing the CaseHOLD dataset (https://arxiv.org/abs/2104.08671) as a canonical dataset on the HuggingFace Hub and would like to add the raw data files to the Hub with GitHub LFS, since it seems like a more sustainable long term storage solution, compared to other storage solutions available to my team... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3164/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3164/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3163 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3163/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3163/comments | https://api.github.com/repos/huggingface/datasets/issues/3163/events | https://github.com/huggingface/datasets/pull/3163 | 1,035,475,061 | PR_kwDODunzps4tpI44 | 3,163 | Add Image feature | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [
"Awesome, looking forward to using it :)",
"Few additional comments:\r\n* the current API doesn't meet the requirements mentioned in #3145 (e.g. image mime-type). However, this will be doable soon as we also plan to store image bytes alongside paths in arrow files (see https://github.com/huggingface/datasets/pull... | 1,635,188,868,000 | 1,638,815,744,000 | 1,638,812,942,000 | CONTRIBUTOR | null | Adds the Image feature. This feature is heavily inspired by the recently added Audio feature (#2324). Currently, this PR is pretty simple.
Some considerations that need further discussion:
* I've decided to use `Pillow`/`PIL` as the image decoding library. Another candidate I considered is `torchvision`, mostly bec... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3163/reactions",
"total_count": 8,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 7,
"rocket": 0,
"eyes": 1
} | https://api.github.com/repos/huggingface/datasets/issues/3163/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3163",
"html_url": "https://github.com/huggingface/datasets/pull/3163",
"diff_url": "https://github.com/huggingface/datasets/pull/3163.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3163.patch",
"merged_at": 1638812942000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3161 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3161/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3161/comments | https://api.github.com/repos/huggingface/datasets/issues/3161/events | https://github.com/huggingface/datasets/pull/3161 | 1,035,444,292 | PR_kwDODunzps4tpCsm | 3,161 | Add riddle_sense dataset | {
"login": "ziyiwu9494",
"id": 44691149,
"node_id": "MDQ6VXNlcjQ0NjkxMTQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/44691149?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ziyiwu9494",
"html_url": "https://github.com/ziyiwu9494",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [
"@lhoestq \r\nI address all the comments, I think. Thanks! \r\n",
"The five test fails are unrelated to this PR and fixed on master so we can ignore them"
] | 1,635,186,656,000 | 1,636,034,475,000 | 1,636,034,475,000 | CONTRIBUTOR | null | Adding a new dataset for QA with riddles. I'm confused about the tagging process because it looks like the streamlit app loads data from the current repo, so is it something that should be done after merging or off my fork? | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3161/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3161/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3161",
"html_url": "https://github.com/huggingface/datasets/pull/3161",
"diff_url": "https://github.com/huggingface/datasets/pull/3161.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3161.patch",
"merged_at": 1636034474000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3160 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3160/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3160/comments | https://api.github.com/repos/huggingface/datasets/issues/3160/events | https://github.com/huggingface/datasets/pull/3160 | 1,035,274,640 | PR_kwDODunzps4tofO0 | 3,160 | Better error msg if `len(predictions)` doesn't match `len(references)` in metrics | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [
"Can't test this now but it may be a good improvement indeed.",
"I added a function, but it only works with the `list` type. For arrays/tensors, we delegate formatting to the frameworks. "
] | 1,635,175,505,000 | 1,636,112,699,000 | 1,636,104,662,000 | CONTRIBUTOR | null | Improve the error message in `Metric.add_batch` if `len(predictions)` doesn't match `len(references)`.
cc: @BramVanroy (feel free to test this code on your examples and review this PR) | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3160/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3160/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3160",
"html_url": "https://github.com/huggingface/datasets/pull/3160",
"diff_url": "https://github.com/huggingface/datasets/pull/3160.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3160.patch",
"merged_at": 1636104662000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3159 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3159/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3159/comments | https://api.github.com/repos/huggingface/datasets/issues/3159/events | https://github.com/huggingface/datasets/pull/3159 | 1,035,174,560 | PR_kwDODunzps4toKD5 | 3,159 | Make inspect.get_dataset_config_names always return a non-empty list | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"This PR is already working (although not very beautiful; see below): the idea was to have the `DatasetModule.builder_kwargs` accessible from the `builder_cls`, so that this can generate the default builder config (at the class level, without requiring the builder to be instantiated).\r\n\r\nI have a plan for a fol... | 1,635,170,383,000 | 1,635,513,277,000 | 1,635,399,889,000 | MEMBER | null | Make all named configs cases, so that no special unnamed config case needs to be handled differently.
Fix #3135. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3159/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3159/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3159",
"html_url": "https://github.com/huggingface/datasets/pull/3159",
"diff_url": "https://github.com/huggingface/datasets/pull/3159.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3159.patch",
"merged_at": 1635399889000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3158 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3158/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3158/comments | https://api.github.com/repos/huggingface/datasets/issues/3158/events | https://github.com/huggingface/datasets/pull/3158 | 1,035,158,070 | PR_kwDODunzps4toGpe | 3,158 | Fix string encoding for Value type | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"That was fast! \r\n"
] | 1,635,169,453,000 | 1,635,171,126,000 | 1,635,171,125,000 | MEMBER | null | Some metrics have `string` features but currently it fails if users pass integers instead. Indeed feature encoding that handles the conversion of the user's objects to the right python type is missing a case for `string`, while it already works as expected for integers, floats and booleans
Here is an example code th... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3158/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 2,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3158/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3158",
"html_url": "https://github.com/huggingface/datasets/pull/3158",
"diff_url": "https://github.com/huggingface/datasets/pull/3158.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3158.patch",
"merged_at": 1635171125000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3157 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3157/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3157/comments | https://api.github.com/repos/huggingface/datasets/issues/3157/events | https://github.com/huggingface/datasets/pull/3157 | 1,034,775,165 | PR_kwDODunzps4tm3_I | 3,157 | Fixed: duplicate parameter and missing parameter in docstring | {
"login": "PanQiWei",
"id": 46810637,
"node_id": "MDQ6VXNlcjQ2ODEwNjM3",
"avatar_url": "https://avatars.githubusercontent.com/u/46810637?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/PanQiWei",
"html_url": "https://github.com/PanQiWei",
"followers_url": "https://api.github.com/users/Pan... | [] | closed | false | null | [] | null | [] | 1,635,146,760,000 | 1,635,170,539,000 | 1,635,170,539,000 | CONTRIBUTOR | null | changing duplicate parameter `data_files` in `DatasetBuilder.__init__` to the missing parameter `data_dir` | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3157/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3157/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3157",
"html_url": "https://github.com/huggingface/datasets/pull/3157",
"diff_url": "https://github.com/huggingface/datasets/pull/3157.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3157.patch",
"merged_at": 1635170538000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3155 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3155/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3155/comments | https://api.github.com/repos/huggingface/datasets/issues/3155/events | https://github.com/huggingface/datasets/issues/3155 | 1,034,468,757 | I_kwDODunzps49qL2V | 3,155 | Illegal instruction (core dumped) at datasets import | {
"login": "hacobe",
"id": 91226467,
"node_id": "MDQ6VXNlcjkxMjI2NDY3",
"avatar_url": "https://avatars.githubusercontent.com/u/91226467?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hacobe",
"html_url": "https://github.com/hacobe",
"followers_url": "https://api.github.com/users/hacobe/fo... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"It seems to be an issue with how conda-forge is building the binaries. It works on some machines, but not a machine with AMD Opteron 8384 processors."
] | 1,635,096,096,000 | 1,637,262,424,000 | 1,637,262,423,000 | CONTRIBUTOR | null | ## Describe the bug
I install datasets using conda and when I import datasets I get: "Illegal instruction (core dumped)"
## Steps to reproduce the bug
```
conda create --prefix path/to/env
conda activate path/to/env
conda install -c huggingface -c conda-forge datasets
# exits with output "Illegal instruction... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3155/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3155/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3154 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3154/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3154/comments | https://api.github.com/repos/huggingface/datasets/issues/3154/events | https://github.com/huggingface/datasets/issues/3154 | 1,034,361,806 | I_kwDODunzps49pxvO | 3,154 | Sacrebleu unexpected behaviour/requirement for data format | {
"login": "BramVanroy",
"id": 2779410,
"node_id": "MDQ6VXNlcjI3Nzk0MTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BramVanroy",
"html_url": "https://github.com/BramVanroy",
"followers_url": "https://api.github.com/users... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Hi @BramVanroy!\r\n\r\nGood question. This project relies on PyArrow (tables) to store data too big to fit in RAM. In the case of metrics, this means that the number of predictions and references has to match to form a table.\r\n\r\nThat's why your example throws an error even though it matches the schema:\r\n```p... | 1,635,065,733,000 | 1,635,671,312,000 | 1,635,671,311,000 | CONTRIBUTOR | null | ## Describe the bug
When comparing with the original `sacrebleu` implementation, the `datasets` implementation does some strange things that I do not quite understand. This issue was triggered when I was trying to implement TER and found the datasets implementation of BLEU [here](https://github.com/huggingface/dataset... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3154/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3154/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3153 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3153/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3153/comments | https://api.github.com/repos/huggingface/datasets/issues/3153/events | https://github.com/huggingface/datasets/pull/3153 | 1,034,179,198 | PR_kwDODunzps4tlEVE | 3,153 | Add TER (as implemented in sacrebleu) | {
"login": "BramVanroy",
"id": 2779410,
"node_id": "MDQ6VXNlcjI3Nzk0MTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BramVanroy",
"html_url": "https://github.com/BramVanroy",
"followers_url": "https://api.github.com/users... | [] | closed | false | null | [] | null | [
"The problem appears to stem from the omission of the lines that you mentioned. If you add them back and try examples from [this](https://huggingface.co/docs/datasets/using_metrics.html) tutorial (sacrebleu metric example) the code you implemented works fine.\r\n\r\nI think the purpose of these lines is follows:\r\... | 1,634,999,205,000 | 1,635,851,051,000 | 1,635,851,051,000 | CONTRIBUTOR | null | Implements TER (Translation Edit Rate) as per its implementation in sacrebleu. Sacrebleu for BLEU scores is already implemented in `datasets` so I thought this would be a nice addition.
I started from the sacrebleu implementation, as the two metrics have a lot in common.
Verified with sacrebleu's [testing suite](... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3153/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3153/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3153",
"html_url": "https://github.com/huggingface/datasets/pull/3153",
"diff_url": "https://github.com/huggingface/datasets/pull/3153.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3153.patch",
"merged_at": 1635851051000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3152 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3152/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3152/comments | https://api.github.com/repos/huggingface/datasets/issues/3152/events | https://github.com/huggingface/datasets/pull/3152 | 1,034,039,379 | PR_kwDODunzps4tkqi- | 3,152 | Fix some typos in the documentation | {
"login": "h4iku",
"id": 3812788,
"node_id": "MDQ6VXNlcjM4MTI3ODg=",
"avatar_url": "https://avatars.githubusercontent.com/u/3812788?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/h4iku",
"html_url": "https://github.com/h4iku",
"followers_url": "https://api.github.com/users/h4iku/follower... | [] | closed | false | null | [] | null | [] | 1,634,953,115,000 | 1,635,172,056,000 | 1,635,170,628,000 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3152/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3152/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3152",
"html_url": "https://github.com/huggingface/datasets/pull/3152",
"diff_url": "https://github.com/huggingface/datasets/pull/3152.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3152.patch",
"merged_at": 1635170628000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3151 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3151/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3151/comments | https://api.github.com/repos/huggingface/datasets/issues/3151/events | https://github.com/huggingface/datasets/pull/3151 | 1,033,890,501 | PR_kwDODunzps4tkL7t | 3,151 | Re-add faiss to windows testing suite | {
"login": "BramVanroy",
"id": 2779410,
"node_id": "MDQ6VXNlcjI3Nzk0MTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BramVanroy",
"html_url": "https://github.com/BramVanroy",
"followers_url": "https://api.github.com/users... | [] | closed | false | null | [] | null | [] | 1,634,931,269,000 | 1,635,850,054,000 | 1,635,847,563,000 | CONTRIBUTOR | null | In recent versions, `faiss-cpu` seems to be available for Windows as well. See the [PyPi page](https://pypi.org/project/faiss-cpu/#files) to confirm. We can therefore included it for Windows in the setup file.
At first tests didn't pass due to problems with permissions as caused by `NamedTemporaryFile` on Windows. T... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3151/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3151/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3151",
"html_url": "https://github.com/huggingface/datasets/pull/3151",
"diff_url": "https://github.com/huggingface/datasets/pull/3151.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3151.patch",
"merged_at": 1635847563000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3150 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3150/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3150/comments | https://api.github.com/repos/huggingface/datasets/issues/3150/events | https://github.com/huggingface/datasets/issues/3150 | 1,033,831,530 | I_kwDODunzps49nwRq | 3,150 | Faiss _is_ available on Windows | {
"login": "BramVanroy",
"id": 2779410,
"node_id": "MDQ6VXNlcjI3Nzk0MTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BramVanroy",
"html_url": "https://github.com/BramVanroy",
"followers_url": "https://api.github.com/users... | [] | closed | false | null | [] | null | [
"Sure, feel free to open a PR."
] | 1,634,926,036,000 | 1,635,847,563,000 | 1,635,847,563,000 | CONTRIBUTOR | null | In the setup file, I find the following:
https://github.com/huggingface/datasets/blob/87c71b9c29a40958973004910f97e4892559dfed/setup.py#L171
However, FAISS does install perfectly fine on Windows on my system. You can also confirm this on the [PyPi page](https://pypi.org/project/faiss-cpu/#files), where Windows wh... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3150/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3150/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3149 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3149/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3149/comments | https://api.github.com/repos/huggingface/datasets/issues/3149/events | https://github.com/huggingface/datasets/pull/3149 | 1,033,747,625 | PR_kwDODunzps4tjuUt | 3,149 | Add CMU Hinglish DoG Dataset for MT | {
"login": "Ishan-Kumar2",
"id": 46553104,
"node_id": "MDQ6VXNlcjQ2NTUzMTA0",
"avatar_url": "https://avatars.githubusercontent.com/u/46553104?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Ishan-Kumar2",
"html_url": "https://github.com/Ishan-Kumar2",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | [
"Hi @lhoestq, thanks a lot for the help. I have moved the part as suggested. \r\nAlthough still while running the dummy data script, I face this issue\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"/home/ishan/anaconda3/bin/datasets-cli\", line 8, in <module>\r\n sys.exit(main())\r\n File \"/home/... | 1,634,919,445,000 | 1,636,976,202,000 | 1,636,972,065,000 | CONTRIBUTOR | null | Address part of #2841
Added the CMU Hinglish DoG Dataset as in GLUECoS. Added it as a seperate dataset as unlike other tasks of GLUE CoS this can't be evaluated for a BERT like model.
Consists of parallel dataset between Hinglish (Hindi-English) and English, can be used for Machine Translation between the two.
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3149/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3149/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3149",
"html_url": "https://github.com/huggingface/datasets/pull/3149",
"diff_url": "https://github.com/huggingface/datasets/pull/3149.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3149.patch",
"merged_at": 1636972065000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3147 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3147/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3147/comments | https://api.github.com/repos/huggingface/datasets/issues/3147/events | https://github.com/huggingface/datasets/pull/3147 | 1,033,607,659 | PR_kwDODunzps4tjRHG | 3,147 | Fix CLI test to ignore verfications when saving infos | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,910,766,000 | 1,635,321,710,000 | 1,635,321,709,000 | MEMBER | null | Fix #3146. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3147/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3147/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3147",
"html_url": "https://github.com/huggingface/datasets/pull/3147",
"diff_url": "https://github.com/huggingface/datasets/pull/3147.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3147.patch",
"merged_at": 1635321709000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3146 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3146/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3146/comments | https://api.github.com/repos/huggingface/datasets/issues/3146/events | https://github.com/huggingface/datasets/issues/3146 | 1,033,605,947 | I_kwDODunzps49m5M7 | 3,146 | CLI test command throws NonMatchingSplitsSizesError when saving infos | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,634,910,653,000 | 1,635,321,709,000 | 1,635,321,709,000 | MEMBER | null | When trying to generate a datset JSON metadata, a `NonMatchingSplitsSizesError` is thrown:
```
$ datasets-cli test datasets/arabic_billion_words --save_infos --all_configs
Testing builder 'Alittihad' (1/10)
Downloading and preparing dataset arabic_billion_words/Alittihad (download: 332.13 MiB, generated: Unknown si... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3146/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3146/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3141 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3141/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3141/comments | https://api.github.com/repos/huggingface/datasets/issues/3141/events | https://github.com/huggingface/datasets/pull/3141 | 1,033,555,910 | PR_kwDODunzps4tjGYz | 3,141 | Fix caching bugs | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [] | 1,634,907,565,000 | 1,634,935,928,000 | 1,634,910,425,000 | CONTRIBUTOR | null | This PR fixes some caching bugs (most likely introduced in the latest refactor):
* remove ")" added by accident in the dataset dir name
* correctly pass the namespace kwargs in `CachedDatasetModuleFactory`
* improve the warning message if `HF_DATASETS_OFFLINE is `True`
| {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3141/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3141/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3141",
"html_url": "https://github.com/huggingface/datasets/pull/3141",
"diff_url": "https://github.com/huggingface/datasets/pull/3141.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3141.patch",
"merged_at": 1634910424000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3137 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3137/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3137/comments | https://api.github.com/repos/huggingface/datasets/issues/3137/events | https://github.com/huggingface/datasets/pull/3137 | 1,033,363,652 | PR_kwDODunzps4tievk | 3,137 | Fix numpy deprecation warning for ragged tensors | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"This'll be a really helpful fix, thank you!"
] | 1,634,894,266,000 | 1,634,918,655,000 | 1,634,918,654,000 | MEMBER | null | Numpy shows a deprecation warning when we call `np.array` on a list of ragged tensors without specifying the `dtype`. If their shapes match, the tensors can be collated together, otherwise the resulting array should have `dtype=np.object`.
Fix #3084
cc @Rocketknight1 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3137/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3137/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3137",
"html_url": "https://github.com/huggingface/datasets/pull/3137",
"diff_url": "https://github.com/huggingface/datasets/pull/3137.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3137.patch",
"merged_at": 1634918654000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3136 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3136/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3136/comments | https://api.github.com/repos/huggingface/datasets/issues/3136/events | https://github.com/huggingface/datasets/pull/3136 | 1,033,360,396 | PR_kwDODunzps4tieFi | 3,136 | Fix script of Arabic Billion Words dataset to return all data | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,894,064,000 | 1,634,909,321,000 | 1,634,909,320,000 | MEMBER | null | The script has a bug and only parses and generates a portion of the entire dataset.
This PR fixes the loading script so that is properly parses the entire dataset.
Current implementation generates the same number of examples as reported in the [original paper](https://arxiv.org/abs/1611.04033) for all configurat... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3136/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3136/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3136",
"html_url": "https://github.com/huggingface/datasets/pull/3136",
"diff_url": "https://github.com/huggingface/datasets/pull/3136.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3136.patch",
"merged_at": 1634909319000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3135 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3135/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3135/comments | https://api.github.com/repos/huggingface/datasets/issues/3135/events | https://github.com/huggingface/datasets/issues/3135 | 1,033,294,299 | I_kwDODunzps49ltHb | 3,135 | Make inspect.get_dataset_config_names always return a non-empty list of configs | {
"login": "severo",
"id": 1676121,
"node_id": "MDQ6VXNlcjE2NzYxMjE=",
"avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/severo",
"html_url": "https://github.com/severo",
"followers_url": "https://api.github.com/users/severo/foll... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
},
{
"id": 3470211881,
"node_id": "LA_k... | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Hi @severo, I guess this issue requests not only to be able to access the configuration name (by using `inspect.get_dataset_config_names`), but the configuration itself as well (I mean you use the name to get the configuration afterwards, maybe using `builder_cls.builder_configs`), is this right?",
"Yes, maybe t... | 1,634,889,770,000 | 1,635,399,889,000 | 1,635,399,889,000 | CONTRIBUTOR | null | **Is your feature request related to a problem? Please describe.**
Currently, some datasets have a configuration, while others don't. It would be simpler for the user to always have configuration names to refer to
**Describe the solution you'd like**
In that sense inspect.get_dataset_config_names should always... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3135/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3135/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3133 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3133/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3133/comments | https://api.github.com/repos/huggingface/datasets/issues/3133/events | https://github.com/huggingface/datasets/pull/3133 | 1,032,511,710 | PR_kwDODunzps4tftyZ | 3,133 | Support Audio feature in streaming mode | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,823,477,000 | 1,636,726,385,000 | 1,636,726,384,000 | MEMBER | null | Fix #3132. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3133/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3133/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3133",
"html_url": "https://github.com/huggingface/datasets/pull/3133",
"diff_url": "https://github.com/huggingface/datasets/pull/3133.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3133.patch",
"merged_at": 1636726384000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3132 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3132/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3132/comments | https://api.github.com/repos/huggingface/datasets/issues/3132/events | https://github.com/huggingface/datasets/issues/3132 | 1,032,505,430 | I_kwDODunzps49ishW | 3,132 | Support Audio feature in streaming mode | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,634,823,138,000 | 1,636,726,384,000 | 1,636,726,384,000 | MEMBER | null | Currently, Audio feature is only supported for non-streaming datasets.
Due to the large size of many speech datasets, we should also support Audio feature in streaming mode.
| {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3132/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3132/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3129 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3129/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3129/comments | https://api.github.com/repos/huggingface/datasets/issues/3129/events | https://github.com/huggingface/datasets/pull/3129 | 1,032,234,167 | PR_kwDODunzps4tezlA | 3,129 | Support Audio feature for TAR archives in sequential access | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"Also do you think we can adapt `cast_column` to keep the same value for this new parameter when the user only wants to change the sampling rate ?",
"Thanks for your comments, @lhoestq, I will address them afterwards.\r\n\r\nBut, I think it is more important/urgent first address the current blocking non-passing t... | 1,634,806,611,000 | 1,637,170,928,000 | 1,637,170,927,000 | MEMBER | null | Add Audio feature support for TAR archived files in sequential access.
Fix #3128. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3129/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3129/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3129",
"html_url": "https://github.com/huggingface/datasets/pull/3129",
"diff_url": "https://github.com/huggingface/datasets/pull/3129.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3129.patch",
"merged_at": 1637170927000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3128 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3128/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3128/comments | https://api.github.com/repos/huggingface/datasets/issues/3128/events | https://github.com/huggingface/datasets/issues/3128 | 1,032,201,870 | I_kwDODunzps49hiaO | 3,128 | Support Audio feature for TAR archives in sequential access | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892871,
"node_id": "MDU6TGFiZWwxOTM1ODkyODcx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement",
"name": "enhancement",
"color": "a2eeef",
"default": true,
"description": "New feature or request"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,634,804,581,000 | 1,637,170,927,000 | 1,637,170,927,000 | MEMBER | null | Currently, Audio feature accesses each audio file by their file path.
However, streamed TAR archive files do not allow random access to their archived files.
Therefore, we should enhance the Audio feature to support TAR archived files in sequential access. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3128/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3128/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3126 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3126/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3126/comments | https://api.github.com/repos/huggingface/datasets/issues/3126/events | https://github.com/huggingface/datasets/issues/3126 | 1,032,093,055 | I_kwDODunzps49hH1_ | 3,126 | "arabic_billion_words" dataset does not create the full dataset | {
"login": "vitalyshalumov",
"id": 33824221,
"node_id": "MDQ6VXNlcjMzODI0MjIx",
"avatar_url": "https://avatars.githubusercontent.com/u/33824221?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vitalyshalumov",
"html_url": "https://github.com/vitalyshalumov",
"followers_url": "https://api.gi... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Thanks for reporting, @vitalyshalumov.\r\n\r\nApparently the script to parse the data has a bug, and does not generate the entire dataset.\r\n\r\nI'm fixing it."
] | 1,634,796,158,000 | 1,634,909,320,000 | 1,634,909,320,000 | NONE | null | ## Describe the bug
When running:
raw_dataset = load_dataset('arabic_billion_words','Alittihad')
the correct dataset file is pulled from the url.
But, the generated dataset includes just a small portion of the data included in the file.
This is true for all other portions of the "arabic_billion_words" dataset ('A... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3126/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3126/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3125 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3125/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3125/comments | https://api.github.com/repos/huggingface/datasets/issues/3125/events | https://github.com/huggingface/datasets/pull/3125 | 1,032,046,666 | PR_kwDODunzps4teNPC | 3,125 | Add SLR83 to OpenSLR | {
"login": "tyrius02",
"id": 4561309,
"node_id": "MDQ6VXNlcjQ1NjEzMDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4561309?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tyrius02",
"html_url": "https://github.com/tyrius02",
"followers_url": "https://api.github.com/users/tyriu... | [] | closed | false | null | [] | null | [] | 1,634,790,360,000 | 1,634,933,405,000 | 1,634,891,422,000 | CONTRIBUTOR | null | The PR resolves #3119, adding SLR83 (UK and Ireland dialects) to the previously created OpenSLR dataset. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3125/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3125/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3125",
"html_url": "https://github.com/huggingface/datasets/pull/3125",
"diff_url": "https://github.com/huggingface/datasets/pull/3125.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3125.patch",
"merged_at": 1634891422000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3124 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3124/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3124/comments | https://api.github.com/repos/huggingface/datasets/issues/3124/events | https://github.com/huggingface/datasets/pull/3124 | 1,031,976,286 | PR_kwDODunzps4td-5w | 3,124 | More efficient nested features encoding | {
"login": "eladsegal",
"id": 13485709,
"node_id": "MDQ6VXNlcjEzNDg1NzA5",
"avatar_url": "https://avatars.githubusercontent.com/u/13485709?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eladsegal",
"html_url": "https://github.com/eladsegal",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | [
"@lhoestq @albertvillanova @mariosasko\r\nCan you please check this out?",
"Thanks, done!"
] | 1,634,781,331,000 | 1,635,865,633,000 | 1,635,851,044,000 | CONTRIBUTOR | null | Nested encoding of features wastes a lot of time on operations which are effectively doing nothing when lists are used.
For example, if in the input we have a list of integers, `encoded_nested_example` will iterate over it and apply `encoded_nested_example` on every element even though it just return the int as is.
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3124/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3124/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3124",
"html_url": "https://github.com/huggingface/datasets/pull/3124",
"diff_url": "https://github.com/huggingface/datasets/pull/3124.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3124.patch",
"merged_at": 1635851044000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3123 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3123/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3123/comments | https://api.github.com/repos/huggingface/datasets/issues/3123/events | https://github.com/huggingface/datasets/issues/3123 | 1,031,793,207 | I_kwDODunzps49f-o3 | 3,123 | Segmentation fault when loading datasets from file | {
"login": "TevenLeScao",
"id": 26709476,
"node_id": "MDQ6VXNlcjI2NzA5NDc2",
"avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/TevenLeScao",
"html_url": "https://github.com/TevenLeScao",
"followers_url": "https://api.github.com/... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"Hi ! I created an issue on Arrow's JIRA after making a minimum reproducible example\r\n\r\nhttps://issues.apache.org/jira/browse/ARROW-14439\r\n\r\n```python\r\nimport io\r\n\r\nimport pyarrow.json as paj\r\n\r\nbatch = b'{\"a\": [], \"b\": 1}\\n{\"b\": 1}'\r\nblock_size = 12\r\n\r\npaj.read_json(\r\n io.BytesI... | 1,634,760,971,000 | 1,635,865,027,000 | 1,635,865,027,000 | MEMBER | null | ## Describe the bug
Custom dataset loading sometimes segfaults and kills the process if chunks contain a variety of features/
## Steps to reproduce the bug
Download an example file:
```
wget https://gist.githubusercontent.com/TevenLeScao/11e2184394b3fa47d693de2550942c6b/raw/4232704d08fbfcaf93e5b51def9e50515076... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3123/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3123/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3122 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3122/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3122/comments | https://api.github.com/repos/huggingface/datasets/issues/3122/events | https://github.com/huggingface/datasets/issues/3122 | 1,031,787,509 | I_kwDODunzps49f9P1 | 3,122 | OSError with a custom dataset loading script | {
"login": "suzanab",
"id": 38602977,
"node_id": "MDQ6VXNlcjM4NjAyOTc3",
"avatar_url": "https://avatars.githubusercontent.com/u/38602977?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/suzanab",
"html_url": "https://github.com/suzanab",
"followers_url": "https://api.github.com/users/suzana... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Hi,\r\n\r\nthere is a difference in how the `data_dir` is zipped between the `classla/janes_tag` and the `classla/reldi_hr` dataset. After unzipping, for the former, the data files (`*.conllup`) are in the root directory (root -> data files), and for the latter, they are inside the `data` directory (root -> `data`... | 1,634,760,519,000 | 1,637,661,338,000 | 1,637,661,338,000 | NONE | null | ## Describe the bug
I am getting an OS error when trying to load the newly uploaded dataset classla/janes_tag. What puzzles me is that I have already uploaded a very similar dataset - classla/reldi_hr - with no issues. The loading scripts for the two datasets are almost identical and they have the same directory struc... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3122/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3122/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3121 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3121/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3121/comments | https://api.github.com/repos/huggingface/datasets/issues/3121/events | https://github.com/huggingface/datasets/pull/3121 | 1,031,673,115 | PR_kwDODunzps4tc_6q | 3,121 | Use huggingface_hub.HfApi to list datasets/metrics | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [] | 1,634,752,109,000 | 1,636,112,708,000 | 1,636,105,716,000 | CONTRIBUTOR | null | Delete `datasets.inspect.HfApi` and use `huggingface_hub.HfApi` instead.
WIP until https://github.com/huggingface/huggingface_hub/pull/429 is merged, then wait for the new release of `huggingface_hub`, update the `huggingface_hub` version in `setup.py` and merge this PR.
cc: @lhoestq | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3121/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3121/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3121",
"html_url": "https://github.com/huggingface/datasets/pull/3121",
"diff_url": "https://github.com/huggingface/datasets/pull/3121.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3121.patch",
"merged_at": 1636105715000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3120 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3120/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3120/comments | https://api.github.com/repos/huggingface/datasets/issues/3120/events | https://github.com/huggingface/datasets/pull/3120 | 1,031,574,511 | PR_kwDODunzps4tcril | 3,120 | Correctly update metadata to preserve features when concatenating datasets with axis=1 | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [] | 1,634,745,298,000 | 1,634,891,331,000 | 1,634,827,821,000 | CONTRIBUTOR | null | This PR correctly updates metadata to preserve higher-level feature types (e.g. `ClassLabel`) in `datasets.concatenate_datasets` when `axis=1`. Previously, we would delete the feature metadata in `datasets.concatenate_datasets` if `axis=1` and restore the feature types from the arrow table schema in `Dataset.__init__`.... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3120/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3120/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3120",
"html_url": "https://github.com/huggingface/datasets/pull/3120",
"diff_url": "https://github.com/huggingface/datasets/pull/3120.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3120.patch",
"merged_at": 1634827821000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3119 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3119/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3119/comments | https://api.github.com/repos/huggingface/datasets/issues/3119/events | https://github.com/huggingface/datasets/issues/3119 | 1,031,328,044 | I_kwDODunzps49eNEs | 3,119 | Add OpenSLR 83 - Crowdsourced high-quality UK and Ireland English Dialect speech | {
"login": "tyrius02",
"id": 4561309,
"node_id": "MDQ6VXNlcjQ1NjEzMDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4561309?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tyrius02",
"html_url": "https://github.com/tyrius02",
"followers_url": "https://api.github.com/users/tyriu... | [
{
"id": 2067376369,
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request",
"name": "dataset request",
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset"
}
] | closed | false | {
"login": "tyrius02",
"id": 4561309,
"node_id": "MDQ6VXNlcjQ1NjEzMDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4561309?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tyrius02",
"html_url": "https://github.com/tyrius02",
"followers_url": "https://api.github.com/users/tyriu... | [
{
"login": "tyrius02",
"id": 4561309,
"node_id": "MDQ6VXNlcjQ1NjEzMDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4561309?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tyrius02",
"html_url": "https://github.com/tyrius02",
"followers_url": "https://api.gi... | null | [
"Ugh. The index files for SLR83 are CSV, not TSV. I need to add logic to process these index files."
] | 1,634,731,507,000 | 1,634,929,252,000 | 1,634,891,422,000 | CONTRIBUTOR | null | ## Adding a Dataset
- **Name:** *openslr**
- **Description:** *Data set which contains male and female recordings of English from various dialects of the UK and Ireland.*
- **Paper:** *https://www.openslr.org/resources/83/about.html*
- **Data:** *Eleven separate data files can be found via https://www.openslr.org/r... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3119/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3119/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3118 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3118/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3118/comments | https://api.github.com/repos/huggingface/datasets/issues/3118/events | https://github.com/huggingface/datasets/pull/3118 | 1,031,309,549 | PR_kwDODunzps4tb0LY | 3,118 | Fix CI error at each release commit | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,730,278,000 | 1,634,734,956,000 | 1,634,734,956,000 | MEMBER | null | Fix test_load_dataset_canonical at release commit.
Fix #3117. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3118/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3118/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3118",
"html_url": "https://github.com/huggingface/datasets/pull/3118",
"diff_url": "https://github.com/huggingface/datasets/pull/3118.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3118.patch",
"merged_at": 1634734955000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3117 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3117/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3117/comments | https://api.github.com/repos/huggingface/datasets/issues/3117/events | https://github.com/huggingface/datasets/issues/3117 | 1,031,308,083 | I_kwDODunzps49eIMz | 3,117 | CI error at each release commit | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,634,730,173,000 | 1,634,734,955,000 | 1,634,734,955,000 | MEMBER | null | After 1.12.0, there is a recurrent CI error at each release commit: https://app.circleci.com/pipelines/github/huggingface/datasets/8289/workflows/665d954d-e409-4602-8202-e678594d2946/jobs/51110
```
____________________ LoadTest.test_load_dataset_canonical _____________________
[gw0] win32 -- Python 3.6.8 C:\tools\... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3117/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3117/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3116 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3116/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3116/comments | https://api.github.com/repos/huggingface/datasets/issues/3116/events | https://github.com/huggingface/datasets/pull/3116 | 1,031,270,611 | PR_kwDODunzps4tbr6g | 3,116 | Update doc links to point to new docs | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [
{
"id": 1935892861,
"node_id": "MDU6TGFiZWwxOTM1ODkyODYx",
"url": "https://api.github.com/repos/huggingface/datasets/labels/documentation",
"name": "documentation",
"color": "0075ca",
"default": true,
"description": "Improvements or additions to documentation"
}
] | closed | false | null | [] | null | [] | 1,634,727,647,000 | 1,634,891,368,000 | 1,634,891,205,000 | CONTRIBUTOR | null | This PR:
* updates the README links and the ADD_NEW_DATASET template to point to the new docs (the new docs don't have a section with the list of all the possible features, so I added that info to the `Features` docstring, which is then referenced in the ADD_NEW_DATASET template)
* fixes some broken links in the `.rs... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3116/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3116/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3116",
"html_url": "https://github.com/huggingface/datasets/pull/3116",
"diff_url": "https://github.com/huggingface/datasets/pull/3116.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3116.patch",
"merged_at": 1634891205000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3115 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3115/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3115/comments | https://api.github.com/repos/huggingface/datasets/issues/3115/events | https://github.com/huggingface/datasets/pull/3115 | 1,030,737,524 | PR_kwDODunzps4tZ-Vr | 3,115 | Fill in dataset card for NCBI disease dataset | {
"login": "edugp",
"id": 17855740,
"node_id": "MDQ6VXNlcjE3ODU1NzQw",
"avatar_url": "https://avatars.githubusercontent.com/u/17855740?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/edugp",
"html_url": "https://github.com/edugp",
"followers_url": "https://api.github.com/users/edugp/follow... | [] | closed | false | null | [] | null | [] | 1,634,677,025,000 | 1,634,891,107,000 | 1,634,891,107,000 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3115/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3115/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3115",
"html_url": "https://github.com/huggingface/datasets/pull/3115",
"diff_url": "https://github.com/huggingface/datasets/pull/3115.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3115.patch",
"merged_at": 1634891107000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3111 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3111/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3111/comments | https://api.github.com/repos/huggingface/datasets/issues/3111/events | https://github.com/huggingface/datasets/issues/3111 | 1,030,598,983 | I_kwDODunzps49bbFH | 3,111 | concatenate_datasets removes ClassLabel typing. | {
"login": "Dref360",
"id": 8976546,
"node_id": "MDQ6VXNlcjg5NzY1NDY=",
"avatar_url": "https://avatars.githubusercontent.com/u/8976546?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Dref360",
"html_url": "https://github.com/Dref360",
"followers_url": "https://api.github.com/users/Dref360/... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [
{
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https:... | null | [
"Something like this would fix it I think: https://github.com/huggingface/datasets/compare/master...Dref360:HF-3111/concatenate_types?expand=1"
] | 1,634,666,731,000 | 1,634,827,821,000 | 1,634,827,821,000 | CONTRIBUTOR | null | ## Describe the bug
When concatenating two datasets, we lose typing of ClassLabel columns.
I can work on this if this is a legitimate bug,
## Steps to reproduce the bug
```python
import datasets
from datasets import Dataset, ClassLabel, Value, concatenate_datasets
DS_LEN = 100
my_dataset = Dataset.from_... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3111/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3111/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3110 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3110/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3110/comments | https://api.github.com/repos/huggingface/datasets/issues/3110/events | https://github.com/huggingface/datasets/pull/3110 | 1,030,558,484 | PR_kwDODunzps4tZakS | 3,110 | Stream TAR-based dataset using iter_archive | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"I'm creating a new branch `stream-tar-audio` just for the audio datasets since they need https://github.com/huggingface/datasets/pull/3129 to be merged first",
"The CI fails are only related to missing sections or tags in the dataset cards - which is unrelated to this PR"
] | 1,634,663,784,000 | 1,636,134,529,000 | 1,636,134,528,000 | MEMBER | null | I converted all the dataset based on TAR archive to use iter_archive instead, so that they can be streamable.
It means that around 80 datasets become streamable :) | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3110/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3110/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3110",
"html_url": "https://github.com/huggingface/datasets/pull/3110",
"diff_url": "https://github.com/huggingface/datasets/pull/3110.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3110.patch",
"merged_at": 1636134528000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3109 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3109/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3109/comments | https://api.github.com/repos/huggingface/datasets/issues/3109/events | https://github.com/huggingface/datasets/pull/3109 | 1,030,543,284 | PR_kwDODunzps4tZXmC | 3,109 | Update BibTeX entry | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,662,771,000 | 1,634,663,608,000 | 1,634,663,607,000 | MEMBER | null | Update BibTeX entry. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3109/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3109/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3109",
"html_url": "https://github.com/huggingface/datasets/pull/3109",
"diff_url": "https://github.com/huggingface/datasets/pull/3109.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3109.patch",
"merged_at": 1634663607000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3108 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3108/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3108/comments | https://api.github.com/repos/huggingface/datasets/issues/3108/events | https://github.com/huggingface/datasets/pull/3108 | 1,030,405,618 | PR_kwDODunzps4tY8ID | 3,108 | Add Google BLEU (aka GLEU) metric | {
"login": "slowwavesleep",
"id": 44175589,
"node_id": "MDQ6VXNlcjQ0MTc1NTg5",
"avatar_url": "https://avatars.githubusercontent.com/u/44175589?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/slowwavesleep",
"html_url": "https://github.com/slowwavesleep",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | [] | 1,634,654,918,000 | 1,635,170,824,000 | 1,635,170,824,000 | CONTRIBUTOR | null | This PR adds the NLTK implementation of Google BLEU metric. This is also a part of an effort to resolve an unfortunate naming collision between GLEU for machine translation and GLEU for grammatical error correction.
I used [this page](https://huggingface.co/docs/datasets/add_metric.html) for reference. Please, point ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3108/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3108/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3108",
"html_url": "https://github.com/huggingface/datasets/pull/3108",
"diff_url": "https://github.com/huggingface/datasets/pull/3108.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3108.patch",
"merged_at": 1635170824000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3107 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3107/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3107/comments | https://api.github.com/repos/huggingface/datasets/issues/3107/events | https://github.com/huggingface/datasets/pull/3107 | 1,030,357,527 | PR_kwDODunzps4tYyhF | 3,107 | Add paper BibTeX citation | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,652,491,000 | 1,634,653,582,000 | 1,634,653,581,000 | MEMBER | null | Add paper BibTeX citation to README file. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3107/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3107/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3107",
"html_url": "https://github.com/huggingface/datasets/pull/3107",
"diff_url": "https://github.com/huggingface/datasets/pull/3107.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3107.patch",
"merged_at": 1634653581000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3106 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3106/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3106/comments | https://api.github.com/repos/huggingface/datasets/issues/3106/events | https://github.com/huggingface/datasets/pull/3106 | 1,030,112,473 | PR_kwDODunzps4tYA6i | 3,106 | Fix URLs in blog_authorship_corpus dataset | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,637,965,000 | 1,634,647,840,000 | 1,634,647,839,000 | MEMBER | null | After contacting the authors of the paper "Effects of Age and Gender on Blogging", they confirmed:
- the old URLs are no longer valid
- there are alternative host URLs
Fix #3091. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3106/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3106/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3106",
"html_url": "https://github.com/huggingface/datasets/pull/3106",
"diff_url": "https://github.com/huggingface/datasets/pull/3106.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3106.patch",
"merged_at": 1634647839000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3104 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3104/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3104/comments | https://api.github.com/repos/huggingface/datasets/issues/3104/events | https://github.com/huggingface/datasets/issues/3104 | 1,029,080,412 | I_kwDODunzps49VoVc | 3,104 | Missing Zenodo 1.13.3 release | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Zenodo has fixed on their side the 1.13.3 release: https://zenodo.org/record/5589150"
] | 1,634,561,838,000 | 1,634,908,945,000 | 1,634,908,944,000 | MEMBER | null | After `datasets` 1.13.3 release, this does not appear in Zenodo releases: https://zenodo.org/record/5570305
TODO:
- [x] Contact Zenodo support
- [x] Check it is fixed | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3104/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3104/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3103 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3103/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3103/comments | https://api.github.com/repos/huggingface/datasets/issues/3103/events | https://github.com/huggingface/datasets/pull/3103 | 1,029,069,310 | PR_kwDODunzps4tUzJQ | 3,103 | Fix project description in PyPI | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,561,249,000 | 1,634,561,997,000 | 1,634,561,996,000 | MEMBER | null | Fix project description appearing in PyPI, so that it contains the content of the README.md file (like transformers).
Currently, `datasets` project description appearing in PyPI shows the release instructions addressed to core maintainers: https://pypi.org/project/datasets/1.13.3/
Fix #3102. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3103/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3103/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3103",
"html_url": "https://github.com/huggingface/datasets/pull/3103",
"diff_url": "https://github.com/huggingface/datasets/pull/3103.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3103.patch",
"merged_at": 1634561996000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3102 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3102/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3102/comments | https://api.github.com/repos/huggingface/datasets/issues/3102/events | https://github.com/huggingface/datasets/issues/3102 | 1,029,067,062 | I_kwDODunzps49VlE2 | 3,102 | Unsuitable project description in PyPI | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,634,561,100,000 | 1,634,561,996,000 | 1,634,561,996,000 | MEMBER | null | Currently, `datasets` project description appearing in PyPI shows the release instructions addressed to core maintainers: https://pypi.org/project/datasets/1.13.3/ | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3102/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3102/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3101 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3101/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3101/comments | https://api.github.com/repos/huggingface/datasets/issues/3101/events | https://github.com/huggingface/datasets/pull/3101 | 1,028,966,968 | PR_kwDODunzps4tUelE | 3,101 | Update SUPERB to use Audio features | {
"login": "anton-l",
"id": 26864830,
"node_id": "MDQ6VXNlcjI2ODY0ODMw",
"avatar_url": "https://avatars.githubusercontent.com/u/26864830?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/anton-l",
"html_url": "https://github.com/anton-l",
"followers_url": "https://api.github.com/users/anton-... | [] | closed | false | null | [] | null | [
"Thank you! Sorry I forgot this one @albertvillanova"
] | 1,634,555,118,000 | 1,634,560,434,000 | 1,634,558,806,000 | CONTRIBUTOR | null | This is the same dataset refresh as the other Audio ones: https://github.com/huggingface/datasets/pull/3081
cc @patrickvonplaten | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3101/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3101/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3101",
"html_url": "https://github.com/huggingface/datasets/pull/3101",
"diff_url": "https://github.com/huggingface/datasets/pull/3101.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3101.patch",
"merged_at": 1634558806000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3100 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3100/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3100/comments | https://api.github.com/repos/huggingface/datasets/issues/3100/events | https://github.com/huggingface/datasets/pull/3100 | 1,028,738,180 | PR_kwDODunzps4tTwpn | 3,100 | Replace FSTimeoutError with parent TimeoutError | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,542,629,000 | 1,634,543,515,000 | 1,634,543,514,000 | MEMBER | null | PR #3050 introduced a dependency on `fsspec.FSTiemoutError`. Note that this error only exists from `fsspec` version `2021.06.0` (June 2021).
To fix #3097, there are 2 alternatives:
- Either pinning `fsspec` to versions newer or equal to `2021.06.0`
- Or replacing `fsspec.FSTimeoutError` wth its parent `asyncio.Tim... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3100/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3100/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3100",
"html_url": "https://github.com/huggingface/datasets/pull/3100",
"diff_url": "https://github.com/huggingface/datasets/pull/3100.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3100.patch",
"merged_at": 1634543514000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3099 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3099/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3099/comments | https://api.github.com/repos/huggingface/datasets/issues/3099/events | https://github.com/huggingface/datasets/issues/3099 | 1,028,338,078 | I_kwDODunzps49SzGe | 3,099 | AttributeError: module 'huggingface_hub.hf_api' has no attribute 'DatasetInfo' | {
"login": "JTWang2000",
"id": 49268567,
"node_id": "MDQ6VXNlcjQ5MjY4NTY3",
"avatar_url": "https://avatars.githubusercontent.com/u/49268567?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/JTWang2000",
"html_url": "https://github.com/JTWang2000",
"followers_url": "https://api.github.com/use... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Hi @JTWang2000, thanks for reporting.\r\n\r\nHowever, I cannot reproduce your reported bug:\r\n```python\r\n>>> from datasets import load_dataset\r\n\r\n>>> dataset = load_dataset(\"sst\", \"default\")\r\n>>> dataset\r\nDatasetDict({\r\n train: Dataset({\r\n features: ['sentence', 'label', 'tokens', 'tre... | 1,634,480,267,000 | 1,636,476,149,000 | 1,636,476,148,000 | NONE | null | ## Describe the bug
When using `pip install datasets`
or use `conda install -c huggingface -c conda-forge datasets`
cannot install datasets
## Steps to reproduce the bug
```python
from datasets import load_dataset
dataset = load_dataset("sst", "default")
```
## Actual results
---------------------------... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3099/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3099/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3098 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3098/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3098/comments | https://api.github.com/repos/huggingface/datasets/issues/3098/events | https://github.com/huggingface/datasets/pull/3098 | 1,028,210,790 | PR_kwDODunzps4tSRSZ | 3,098 | Push to hub capabilities for `Dataset` and `DatasetDict` | {
"login": "LysandreJik",
"id": 30755778,
"node_id": "MDQ6VXNlcjMwNzU1Nzc4",
"avatar_url": "https://avatars.githubusercontent.com/u/30755778?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/LysandreJik",
"html_url": "https://github.com/LysandreJik",
"followers_url": "https://api.github.com/... | [] | closed | false | null | [] | null | [
"Thank you for your reviews! I should have addressed all of your comments, and I added a test to ensure that `private` datasets work correctly too. I have merged the changes in `huggingface_hub`, so the `main` branch can be installed now; and I will release v0.1.0 soon.\r\n\r\nAs blockers for this PR:\r\n- It's sti... | 1,634,443,964,000 | 1,638,836,532,000 | 1,637,753,136,000 | MEMBER | null | This PR implements a `push_to_hub` method on `Dataset` and `DatasetDict`. This does not currently work in `IterableDatasetDict` nor `IterableDataset` as those are simple dicts and I would like your opinion on how you would like to implement this before going ahead and doing it.
This implementation needs to be used w... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3098/reactions",
"total_count": 4,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 3,
"eyes": 1
} | https://api.github.com/repos/huggingface/datasets/issues/3098/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3098",
"html_url": "https://github.com/huggingface/datasets/pull/3098",
"diff_url": "https://github.com/huggingface/datasets/pull/3098.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3098.patch",
"merged_at": 1637753136000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3097 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3097/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3097/comments | https://api.github.com/repos/huggingface/datasets/issues/3097/events | https://github.com/huggingface/datasets/issues/3097 | 1,027,750,811 | I_kwDODunzps49Qjub | 3,097 | `ModuleNotFoundError: No module named 'fsspec.exceptions'` | {
"login": "VictorSanh",
"id": 16107619,
"node_id": "MDQ6VXNlcjE2MTA3NjE5",
"avatar_url": "https://avatars.githubusercontent.com/u/16107619?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/VictorSanh",
"html_url": "https://github.com/VictorSanh",
"followers_url": "https://api.github.com/use... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Thanks for reporting, @VictorSanh.\r\n\r\nI'm fixing it."
] | 1,634,326,478,000 | 1,634,543,514,000 | 1,634,543,514,000 | MEMBER | null | ## Describe the bug
I keep runnig into a fsspec ModuleNotFound error
## Steps to reproduce the bug
```python
>>> from datasets import get_dataset_infos
2021-10-15 15:25:37.863206: W tensorflow/stream_executor/platform/default/dso_loader.cc:64] Could not load dynamic library 'libcudart.so.11.0'; dlerror: libcudar... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3097/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3097/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3096 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3096/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3096/comments | https://api.github.com/repos/huggingface/datasets/issues/3096/events | https://github.com/huggingface/datasets/pull/3096 | 1,027,535,685 | PR_kwDODunzps4tQblQ | 3,096 | Fix Audio feature mp3 resampling | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,310,319,000 | 1,634,312,310,000 | 1,634,312,310,000 | MEMBER | null | Issue #3095 is related to mp3 resampling, not to `cast_column`.
This PR fixes Audio feature mp3 resampling.
Fix #3095. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3096/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3096/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3096",
"html_url": "https://github.com/huggingface/datasets/pull/3096",
"diff_url": "https://github.com/huggingface/datasets/pull/3096.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3096.patch",
"merged_at": 1634312309000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3095 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3095/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3095/comments | https://api.github.com/repos/huggingface/datasets/issues/3095/events | https://github.com/huggingface/datasets/issues/3095 | 1,027,453,146 | I_kwDODunzps49PbDa | 3,095 | `cast_column` makes audio decoding fail | {
"login": "patrickvonplaten",
"id": 23423619,
"node_id": "MDQ6VXNlcjIzNDIzNjE5",
"avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/patrickvonplaten",
"html_url": "https://github.com/patrickvonplaten",
"followers_url": "https://... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"cc @anton-l @albertvillanova ",
"Thanks for reporting, @patrickvonplaten.\r\n\r\nI think the issue is related to mp3 resampling, not to `cast_column`.\r\n\r\nYou can check that `cast_column` works OK with non-mp3 audio files:\r\n```python\r\nfrom datasets import load_dataset\r\nimport datasets\r\nds = load_datas... | 1,634,305,018,000 | 1,634,312,310,000 | 1,634,312,310,000 | MEMBER | null | ## Describe the bug
After changing the sampling rate automatic decoding fails.
## Steps to reproduce the bug
```python
from datasets import load_dataset
import datasets
ds = load_dataset("common_voice", "ab", split="train")
ds = ds.cast_column("audio", datasets.features.Audio(sampling_rate=16_000))
pr... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3095/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3095/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3092 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3092/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3092/comments | https://api.github.com/repos/huggingface/datasets/issues/3092/events | https://github.com/huggingface/datasets/pull/3092 | 1,027,260,383 | PR_kwDODunzps4tPj6e | 3,092 | Fix JNLBA dataset | {
"login": "bhavitvyamalik",
"id": 19718818,
"node_id": "MDQ6VXNlcjE5NzE4ODE4",
"avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bhavitvyamalik",
"html_url": "https://github.com/bhavitvyamalik",
"followers_url": "https://api.gi... | [] | closed | false | null | [] | null | [
"Fix #3089.",
"@albertvillanova all tests are passing now. Either you or @lhoestq can review it!"
] | 1,634,290,274,000 | 1,634,891,037,000 | 1,634,891,037,000 | CONTRIBUTOR | null | As mentioned in #3089, I've added more tags and also updated the link for dataset which was earlier using a Google Drive link.
I'm having problem with generating dummy data as `datasets-cli dummy_data ./datasets/jnlpba --auto_generate --match_text_files "*.iob2"` is giving `datasets.keyhash.DuplicatedKeysError: FAIL... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3092/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3092/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3092",
"html_url": "https://github.com/huggingface/datasets/pull/3092",
"diff_url": "https://github.com/huggingface/datasets/pull/3092.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3092.patch",
"merged_at": 1634891037000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3091 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3091/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3091/comments | https://api.github.com/repos/huggingface/datasets/issues/3091/events | https://github.com/huggingface/datasets/issues/3091 | 1,027,251,530 | I_kwDODunzps49Op1K | 3,091 | `blog_authorship_corpus` is broken | {
"login": "fdtomasi",
"id": 12514317,
"node_id": "MDQ6VXNlcjEyNTE0MzE3",
"avatar_url": "https://avatars.githubusercontent.com/u/12514317?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fdtomasi",
"html_url": "https://github.com/fdtomasi",
"followers_url": "https://api.github.com/users/fdt... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Hi @fdtomasi, thanks for reporting.\r\n\r\nYou are right: the original host data URL does no longer exist.\r\n\r\nI've contacted the authors of the dataset to ask them if they host this dataset in another URL.",
"Hi, @fdtomasi, the URL is fixed.\r\n\r\nThe fix is already in our master branch and it will be acces... | 1,634,289,640,000 | 1,634,648,770,000 | 1,634,647,839,000 | NONE | null | ## Describe the bug
The dataset `blog_authorship_corpus` is broken.
By bypassing the checksum checks, the loading does not return any error but the resulting dataset is empty.
I suspect it is because the data download url is broken (http://www.cs.biu.ac.il/~koppel/blogs/blogs.zip).
## Steps to reproduce the bug
... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3091/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3091/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3090 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3090/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3090/comments | https://api.github.com/repos/huggingface/datasets/issues/3090/events | https://github.com/huggingface/datasets/pull/3090 | 1,027,100,371 | PR_kwDODunzps4tPEtH | 3,090 | Update BibTeX entry | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,276,367,000 | 1,634,283,357,000 | 1,634,283,357,000 | MEMBER | null | Update BibTeX entry. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3090/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3090/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3090",
"html_url": "https://github.com/huggingface/datasets/pull/3090",
"diff_url": "https://github.com/huggingface/datasets/pull/3090.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3090.patch",
"merged_at": 1634283357000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3089 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3089/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3089/comments | https://api.github.com/repos/huggingface/datasets/issues/3089/events | https://github.com/huggingface/datasets/issues/3089 | 1,026,973,360 | I_kwDODunzps49Nl6w | 3,089 | JNLPBA Dataset | {
"login": "sciarrilli",
"id": 10460111,
"node_id": "MDQ6VXNlcjEwNDYwMTEx",
"avatar_url": "https://avatars.githubusercontent.com/u/10460111?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sciarrilli",
"html_url": "https://github.com/sciarrilli",
"followers_url": "https://api.github.com/use... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"# Steps to reproduce\r\n\r\nTo reproduce:\r\n\r\n```python\r\nfrom datasets import load_dataset\r\ndataset = load_dataset('jnlpba')\r\n\r\ndataset['train'].features['ner_tags']\r\n```\r\nOutput:\r\n```python\r\nSequence(feature=ClassLabel(num_classes=3, names=['O', 'B', 'I'], names_file=None, id=None), length=-1, ... | 1,634,260,562,000 | 1,634,891,037,000 | 1,634,891,037,000 | NONE | null | ## Describe the bug
A clear and concise description of what the bug is.
## Steps to reproduce the bug
```python
# Sample code to reproduce the bug
```
## Expected results
The dataset loading script for this dataset is incorrect. This is a biomedical dataset used for named entity recognition. The entities in ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3089/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3089/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3088 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3088/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3088/comments | https://api.github.com/repos/huggingface/datasets/issues/3088/events | https://github.com/huggingface/datasets/pull/3088 | 1,026,920,369 | PR_kwDODunzps4tOhRx | 3,088 | Use template column_mapping to transmit_format instead of template features | {
"login": "mariosasko",
"id": 47462742,
"node_id": "MDQ6VXNlcjQ3NDYyNzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mariosasko",
"html_url": "https://github.com/mariosasko",
"followers_url": "https://api.github.com/use... | [] | closed | false | null | [] | null | [
"Thanks for fixing!"
] | 1,634,255,380,000 | 1,634,308,805,000 | 1,634,292,664,000 | CONTRIBUTOR | null | Use `template.column_mapping` to check for modified columns since `template.features` represent a generic template/column mapping.
Fix #3087
TODO:
- [x] Add a test | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3088/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3088/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3088",
"html_url": "https://github.com/huggingface/datasets/pull/3088",
"diff_url": "https://github.com/huggingface/datasets/pull/3088.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3088.patch",
"merged_at": 1634292664000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3087 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3087/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3087/comments | https://api.github.com/repos/huggingface/datasets/issues/3087/events | https://github.com/huggingface/datasets/issues/3087 | 1,026,780,469 | I_kwDODunzps49M201 | 3,087 | Removing label column in a text classification dataset yields to errors | {
"login": "sgugger",
"id": 35901082,
"node_id": "MDQ6VXNlcjM1OTAxMDgy",
"avatar_url": "https://avatars.githubusercontent.com/u/35901082?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sgugger",
"html_url": "https://github.com/sgugger",
"followers_url": "https://api.github.com/users/sgugge... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [] | 1,634,242,370,000 | 1,634,292,664,000 | 1,634,292,664,000 | MEMBER | null | ## Describe the bug
This looks like #3059 but it's not linked to the cache this time. Removing the `label` column from a text classification dataset and then performing any processing will result in an error.
To reproduce:
```py
from datasets import load_dataset
from transformers import AutoTokenizer
raw_da... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3087/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3087/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3086 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3086/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3086/comments | https://api.github.com/repos/huggingface/datasets/issues/3086/events | https://github.com/huggingface/datasets/pull/3086 | 1,026,481,905 | PR_kwDODunzps4tNIvp | 3,086 | Remove _resampler from Audio fields | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,222,330,000 | 1,634,224,421,000 | 1,634,224,420,000 | MEMBER | null | The `_resampler` Audio attribute was implemented to optimize audio resampling, but it should not be cached.
This PR removes `_resampler` from Audio fields, so that it is not returned by `fields()` or `asdict()`.
Fix #3083. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3086/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3086/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3086",
"html_url": "https://github.com/huggingface/datasets/pull/3086",
"diff_url": "https://github.com/huggingface/datasets/pull/3086.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3086.patch",
"merged_at": 1634224420000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3085 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3085/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3085/comments | https://api.github.com/repos/huggingface/datasets/issues/3085/events | https://github.com/huggingface/datasets/pull/3085 | 1,026,467,384 | PR_kwDODunzps4tNFza | 3,085 | Fixes to `to_tf_dataset` | {
"login": "Rocketknight1",
"id": 12866554,
"node_id": "MDQ6VXNlcjEyODY2NTU0",
"avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Rocketknight1",
"html_url": "https://github.com/Rocketknight1",
"followers_url": "https://api.githu... | [] | closed | false | null | [] | null | [
"Hi ! Can you give some details about why you need these changes ?",
"Hey, sorry, I should have explained! I've been getting a lot of `VisibleDeprecationWarning` from Numpy, due to an issue in the formatter, see #3084 . This is a temporary workaround (since I'm using these methods in the upcoming course) until I ... | 1,634,221,556,000 | 1,634,828,729,000 | 1,634,828,728,000 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3085/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3085/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3085",
"html_url": "https://github.com/huggingface/datasets/pull/3085",
"diff_url": "https://github.com/huggingface/datasets/pull/3085.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3085.patch",
"merged_at": 1634828728000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3084 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3084/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3084/comments | https://api.github.com/repos/huggingface/datasets/issues/3084/events | https://github.com/huggingface/datasets/issues/3084 | 1,026,428,992 | I_kwDODunzps49LhBA | 3,084 | VisibleDeprecationWarning when using `set_format("numpy")` | {
"login": "Rocketknight1",
"id": 12866554,
"node_id": "MDQ6VXNlcjEyODY2NTU0",
"avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Rocketknight1",
"html_url": "https://github.com/Rocketknight1",
"followers_url": "https://api.githu... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "Rocketknight1",
"id": 12866554,
"node_id": "MDQ6VXNlcjEyODY2NTU0",
"avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Rocketknight1",
"html_url": "https://github.com/Rocketknight1",
"followers_url": "https://api.githu... | [
{
"login": "Rocketknight1",
"id": 12866554,
"node_id": "MDQ6VXNlcjEyODY2NTU0",
"avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Rocketknight1",
"html_url": "https://github.com/Rocketknight1",
"followers_url"... | null | [
"I just opened a PR and I verified that the code you provided doesn't show any deprecation warning :)"
] | 1,634,219,581,000 | 1,634,918,654,000 | 1,634,918,654,000 | CONTRIBUTOR | null | Code to reproduce:
```
from datasets import load_dataset
dataset = load_dataset("glue", "mnli")
from transformers import AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained('distilbert-base-cased')
def tokenize_function(dataset):
return tokenizer(dataset['premise'])
tokenized_datasets = dataset.... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3084/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3084/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3083 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3083/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3083/comments | https://api.github.com/repos/huggingface/datasets/issues/3083/events | https://github.com/huggingface/datasets/issues/3083 | 1,026,397,062 | I_kwDODunzps49LZOG | 3,083 | Datasets with Audio feature raise error when loaded from cache due to _resampler parameter | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,634,217,833,000 | 1,634,224,420,000 | 1,634,224,420,000 | MEMBER | null | ## Describe the bug
As reported by @patrickvonplaten, when loaded from the cache, datasets containing the Audio feature raise TypeError.
## Steps to reproduce the bug
```python
from datasets import load_dataset
# load first time works
ds = load_dataset("patrickvonplaten/librispeech_asr_dummy", "clean")
# ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3083/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3083/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3082 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3082/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3082/comments | https://api.github.com/repos/huggingface/datasets/issues/3082/events | https://github.com/huggingface/datasets/pull/3082 | 1,026,388,994 | PR_kwDODunzps4tM2BV | 3,082 | Fix error related to huggingface_hub timeout parameter | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,217,467,000 | 1,634,222,392,000 | 1,634,222,391,000 | MEMBER | null | The `huggingface_hub` package added the parameter `timeout` from version 0.0.19.
This PR bumps this minimal version.
Fix #3080. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3082/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3082/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3082",
"html_url": "https://github.com/huggingface/datasets/pull/3082",
"diff_url": "https://github.com/huggingface/datasets/pull/3082.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3082.patch",
"merged_at": 1634222391000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3081 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3081/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3081/comments | https://api.github.com/repos/huggingface/datasets/issues/3081/events | https://github.com/huggingface/datasets/pull/3081 | 1,026,383,749 | PR_kwDODunzps4tM1Gy | 3,081 | [Audio datasets] Adapting all audio datasets | {
"login": "patrickvonplaten",
"id": 23423619,
"node_id": "MDQ6VXNlcjIzNDIzNjE5",
"avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/patrickvonplaten",
"html_url": "https://github.com/patrickvonplaten",
"followers_url": "https://... | [] | closed | false | null | [] | null | [
"@lhoestq - are there other important speech datasets that I'm forgetting here? \r\n\r\nThink PR is good to go otherwise",
"@lhoestq @albertvillanova - how can we make an exception for the AMI README so that the test doesn't fail? The dataset card definitely should have a data preprocessing section",
"Hi @patri... | 1,634,217,225,000 | 1,634,302,323,000 | 1,634,300,553,000 | MEMBER | null | This PR adds the new `Audio(...)` features - see: https://github.com/huggingface/datasets/pull/2324 to the most important audio datasets:
- Librispeech
- Timit
- Common Voice
- AMI
- ... (others I'm forgetting now)
The PR is curently blocked because the following leads to a problem:
```python
from dataset... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3081/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3081/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3081",
"html_url": "https://github.com/huggingface/datasets/pull/3081",
"diff_url": "https://github.com/huggingface/datasets/pull/3081.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3081.patch",
"merged_at": 1634300553000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3080 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3080/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3080/comments | https://api.github.com/repos/huggingface/datasets/issues/3080/events | https://github.com/huggingface/datasets/issues/3080 | 1,026,380,626 | I_kwDODunzps49LVNS | 3,080 | Error related to timeout keyword argument | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,634,217,058,000 | 1,634,222,391,000 | 1,634,222,391,000 | MEMBER | null | ## Describe the bug
As reported by @patrickvonplaten, a TypeError is raised when trying to load a dataset.
## Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_dataset("patrickvonplaten/librispeech_asr_dummy", "clean")
```
## Actual results
```
TypeError: dataset_info() got ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3080/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3080/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3077 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3077/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3077/comments | https://api.github.com/repos/huggingface/datasets/issues/3077/events | https://github.com/huggingface/datasets/pull/3077 | 1,026,150,362 | PR_kwDODunzps4tMFPG | 3,077 | Fix loading a metric with internal import | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,634,202,418,000 | 1,634,202,896,000 | 1,634,202,895,000 | MEMBER | null | After refactoring the module factory (#2986), a bug was introduced when loading metrics with internal imports.
This PR adds a new test case and fixes this bug.
Fix #3076.
CC: @sgugger @merveenoyan | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3077/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 2,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3077/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3077",
"html_url": "https://github.com/huggingface/datasets/pull/3077",
"diff_url": "https://github.com/huggingface/datasets/pull/3077.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3077.patch",
"merged_at": 1634202895000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3076 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3076/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3076/comments | https://api.github.com/repos/huggingface/datasets/issues/3076/events | https://github.com/huggingface/datasets/issues/3076 | 1,026,113,484 | I_kwDODunzps49KT_M | 3,076 | Error when loading a metric | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,634,200,167,000 | 1,634,202,895,000 | 1,634,202,895,000 | MEMBER | null | ## Describe the bug
As reported by @sgugger, after last release, exception is thrown when loading a metric.
## Steps to reproduce the bug
```python
from datasets import load_metric
metric = load_metric("squad_v2")
```
## Actual results
```
FileNotFoundError Traceback (most recent ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3076/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3076/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3075 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3075/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3075/comments | https://api.github.com/repos/huggingface/datasets/issues/3075/events | https://github.com/huggingface/datasets/pull/3075 | 1,026,103,388 | PR_kwDODunzps4tL75E | 3,075 | Updates LexGLUE and MultiEURLEX README.md files | {
"login": "iliaschalkidis",
"id": 1626984,
"node_id": "MDQ6VXNlcjE2MjY5ODQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/1626984?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/iliaschalkidis",
"html_url": "https://github.com/iliaschalkidis",
"followers_url": "https://api.gith... | [] | closed | false | null | [] | null | [] | 1,634,199,556,000 | 1,634,552,020,000 | 1,634,552,020,000 | CONTRIBUTOR | null | Updates LexGLUE and MultiEURLEX README.md files
- Fix leaderboard in LexGLUE.
- Fix an error in the CaseHOLD data example.
- Turn MultiEURLEX dataset statistics table into HTML to nicely render in HF website. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3075/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3075/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3075",
"html_url": "https://github.com/huggingface/datasets/pull/3075",
"diff_url": "https://github.com/huggingface/datasets/pull/3075.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3075.patch",
"merged_at": 1634552020000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3074 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3074/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3074/comments | https://api.github.com/repos/huggingface/datasets/issues/3074/events | https://github.com/huggingface/datasets/pull/3074 | 1,025,940,085 | PR_kwDODunzps4tLbe- | 3,074 | add XCSR dataset | {
"login": "yangxqiao",
"id": 42788901,
"node_id": "MDQ6VXNlcjQyNzg4OTAx",
"avatar_url": "https://avatars.githubusercontent.com/u/42788901?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yangxqiao",
"html_url": "https://github.com/yangxqiao",
"followers_url": "https://api.github.com/users/... | [] | closed | false | null | [] | null | [
"> Hi ! Thanks for adding this dataset :)\r\n> \r\n> Do you know how the translations were done ? Maybe we can mention that in the dataset card.\r\n> \r\n> The rest looks all good to me :) good job with the dataset script and the dataset card !\r\n> \r\n> Just one thing: we try to have dummy_data.zip files that are... | 1,634,186,399,000 | 1,636,379,556,000 | 1,636,379,556,000 | CONTRIBUTOR | null | Hi,
I wanted to add the [XCSR ](https://inklab.usc.edu//XCSR/xcsr_datasets) dataset to huggingface! :)
I followed the instructions of adding new dataset to huggingface and have all the required files ready now! It would be super helpful if you can take a look and review them. Thanks in advance for your time and ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3074/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3074/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3074",
"html_url": "https://github.com/huggingface/datasets/pull/3074",
"diff_url": "https://github.com/huggingface/datasets/pull/3074.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3074.patch",
"merged_at": 1636379556000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3073 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3073/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3073/comments | https://api.github.com/repos/huggingface/datasets/issues/3073/events | https://github.com/huggingface/datasets/issues/3073 | 1,025,718,469 | I_kwDODunzps49IzjF | 3,073 | Import error installing with ppc64le | {
"login": "gcervantes8",
"id": 21228908,
"node_id": "MDQ6VXNlcjIxMjI4OTA4",
"avatar_url": "https://avatars.githubusercontent.com/u/21228908?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gcervantes8",
"html_url": "https://github.com/gcervantes8",
"followers_url": "https://api.github.com/... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"This seems to be an issue with importing PyArrow so I posted the problem [here](https://issues.apache.org/jira/browse/ARROW-14323), and I'm closing this issue.\r\n"
] | 1,634,161,043,000 | 1,634,229,346,000 | 1,634,229,208,000 | NONE | null | ## Describe the bug
Installing the datasets library with a computer running with ppc64le seems to cause an issue when importing the datasets library.
```
python
Python 3.6.13 | packaged by conda-forge | (default, Sep 23 2021, 07:37:44)
[GCC 9.4.0] on linux
Type "help", "copyright", "credits" or "license" for... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3073/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3073/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3072 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3072/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3072/comments | https://api.github.com/repos/huggingface/datasets/issues/3072/events | https://github.com/huggingface/datasets/pull/3072 | 1,025,233,152 | PR_kwDODunzps4tJNnD | 3,072 | Fix pathlib patches for streaming | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [] | 1,634,130,675,000 | 1,634,131,865,000 | 1,634,131,865,000 | MEMBER | null | Fix issue https://github.com/huggingface/datasets/issues/2866 (for good this time)
`counter` now works in both streaming and non-streaming mode.
And the `AttributeError: 'str' object has no attribute 'as_posix'` related to the patch of Path.open is fixed as well
Note : the patches should only affect the datasets... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3072/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3072/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3072",
"html_url": "https://github.com/huggingface/datasets/pull/3072",
"diff_url": "https://github.com/huggingface/datasets/pull/3072.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3072.patch",
"merged_at": 1634131865000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3071 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3071/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3071/comments | https://api.github.com/repos/huggingface/datasets/issues/3071/events | https://github.com/huggingface/datasets/issues/3071 | 1,024,893,493 | I_kwDODunzps49FqI1 | 3,071 | Custom plain text dataset, plain json dataset and plain csv dataset are remove from datasets template folder | {
"login": "zixiliuUSC",
"id": 49173327,
"node_id": "MDQ6VXNlcjQ5MTczMzI3",
"avatar_url": "https://avatars.githubusercontent.com/u/49173327?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zixiliuUSC",
"html_url": "https://github.com/zixiliuUSC",
"followers_url": "https://api.github.com/use... | [
{
"id": 2067376369,
"node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5",
"url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request",
"name": "dataset request",
"color": "e99695",
"default": false,
"description": "Requesting to add a new dataset"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Hi @zixiliuUSC, \r\n\r\nAs explained in the documentation (https://huggingface.co/docs/datasets/loading.html#json), we support loading any dataset in JSON (as well as CSV, text, Parquet) format:\r\n```python\r\nds = load_dataset('json', data_files='my_file.json')\r\n```"
] | 1,634,110,330,000 | 1,634,113,624,000 | 1,634,113,623,000 | NONE | null | ## Adding a Dataset
- **Name:** text, json, csv
- **Description:** I am developing a customized dataset loading script. The problem is mainly about my custom dataset is seperate into many files and I only find a dataset loading template in [https://github.com/huggingface/datasets/blob/1.2.1/datasets/json/json.py](ht... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3071/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3071/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3070 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3070/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3070/comments | https://api.github.com/repos/huggingface/datasets/issues/3070/events | https://github.com/huggingface/datasets/pull/3070 | 1,024,856,745 | PR_kwDODunzps4tIBRp | 3,070 | Fix Windows CI with FileNotFoundError when stting up s3_base fixture | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [
"Thanks ! Sorry for the inconvenience ^^' "
] | 1,634,107,741,000 | 1,634,115,313,000 | 1,634,107,788,000 | MEMBER | null | Fix #3069. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3070/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3070/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3070",
"html_url": "https://github.com/huggingface/datasets/pull/3070",
"diff_url": "https://github.com/huggingface/datasets/pull/3070.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3070.patch",
"merged_at": 1634107788000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3069 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3069/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3069/comments | https://api.github.com/repos/huggingface/datasets/issues/3069/events | https://github.com/huggingface/datasets/issues/3069 | 1,024,818,680 | I_kwDODunzps49FX34 | 3,069 | CI fails on Windows with FileNotFoundError when stting up s3_base fixture | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,634,104,346,000 | 1,634,112,349,000 | 1,634,107,788,000 | MEMBER | null | ## Describe the bug
After commit 9353fc863d0c99ab0427f83cc5a4f04fcf52f1df, the CI fails on Windows with FileNotFoundError when stting up s3_base fixture. See: https://app.circleci.com/pipelines/github/huggingface/datasets/8151/workflows/5db8d154-badd-4d3d-b202-ca7a318997a2/jobs/50321
Error summary:
```
ERROR tes... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3069/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3069/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3068 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3068/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3068/comments | https://api.github.com/repos/huggingface/datasets/issues/3068/events | https://github.com/huggingface/datasets/pull/3068 | 1,024,681,264 | PR_kwDODunzps4tHhOC | 3,068 | feat: increase streaming retry config | {
"login": "borisdayma",
"id": 715491,
"node_id": "MDQ6VXNlcjcxNTQ5MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/715491?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/borisdayma",
"html_url": "https://github.com/borisdayma",
"followers_url": "https://api.github.com/users/b... | [] | closed | false | null | [] | null | [
"@lhoestq I had 2 runs for more than 2 days each, continuously streaming (they were failing before with 3 retries at 1 sec interval).\r\n\r\nThey are running on TPU's (so great internet connection) and only had connection errors a few times each (3 & 4). Each time it worked after only 1 retry.\r\nThe reason for a h... | 1,634,090,450,000 | 1,634,117,156,000 | 1,634,117,154,000 | CONTRIBUTOR | null | Increase streaming config parameters:
* retry interval set to 5 seconds
* max retries set to 20 (so 1mn 40s) | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3068/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3068/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3068",
"html_url": "https://github.com/huggingface/datasets/pull/3068",
"diff_url": "https://github.com/huggingface/datasets/pull/3068.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3068.patch",
"merged_at": 1634117154000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3067 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3067/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3067/comments | https://api.github.com/repos/huggingface/datasets/issues/3067/events | https://github.com/huggingface/datasets/pull/3067 | 1,024,023,185 | PR_kwDODunzps4tFSCy | 3,067 | add story_cloze | {
"login": "zaidalyafeai",
"id": 15667714,
"node_id": "MDQ6VXNlcjE1NjY3NzE0",
"avatar_url": "https://avatars.githubusercontent.com/u/15667714?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/zaidalyafeai",
"html_url": "https://github.com/zaidalyafeai",
"followers_url": "https://api.github.c... | [] | closed | false | null | [] | null | [
"Thanks for pushing this dataset :)\r\n\r\nAccording to the CI, the file `cloze_test_val__spring2016 - cloze_test_ALL_val.csv` is missing in the dummy data zip file (the zip files seem empty). Feel free to add this file with 4-5 lines and it should be good\r\n\r\nAnd you can fix the YAML tags with\r\n```yaml\r\npre... | 1,634,056,613,000 | 1,634,132,893,000 | 1,634,132,893,000 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3067/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3067/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3067",
"html_url": "https://github.com/huggingface/datasets/pull/3067",
"diff_url": "https://github.com/huggingface/datasets/pull/3067.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3067.patch",
"merged_at": 1634132893000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3066 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3066/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3066/comments | https://api.github.com/repos/huggingface/datasets/issues/3066/events | https://github.com/huggingface/datasets/pull/3066 | 1,024,005,311 | PR_kwDODunzps4tFObl | 3,066 | Add iter_archive | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [] | 1,634,055,436,000 | 1,634,548,367,000 | 1,634,548,366,000 | MEMBER | null | Added the `iter_archive` method for the StreamingDownloadManager.
It was already implemented in the regular DownloadManager.
Now it can be used to stream from TAR archives as mentioned in https://github.com/huggingface/datasets/issues/2829
I also updated the `food101` dataset as an example.
Any image/audio data... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3066/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3066/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3066",
"html_url": "https://github.com/huggingface/datasets/pull/3066",
"diff_url": "https://github.com/huggingface/datasets/pull/3066.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3066.patch",
"merged_at": 1634548366000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3065 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3065/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3065/comments | https://api.github.com/repos/huggingface/datasets/issues/3065/events | https://github.com/huggingface/datasets/pull/3065 | 1,023,951,322 | PR_kwDODunzps4tFDjk | 3,065 | Fix test command after refac | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [] | 1,634,052,210,000 | 1,634,052,527,000 | 1,634,052,526,000 | MEMBER | null | Fix the `datasets-cli` test command after the `prepare_module` change in #2986 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3065/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3065/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3065",
"html_url": "https://github.com/huggingface/datasets/pull/3065",
"diff_url": "https://github.com/huggingface/datasets/pull/3065.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3065.patch",
"merged_at": 1634052526000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3062 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3062/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3062/comments | https://api.github.com/repos/huggingface/datasets/issues/3062/events | https://github.com/huggingface/datasets/pull/3062 | 1,023,209,592 | PR_kwDODunzps4tCxfK | 3,062 | Update summary on PyPi beyond NLP | {
"login": "thomwolf",
"id": 7353373,
"node_id": "MDQ6VXNlcjczNTMzNzM=",
"avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/thomwolf",
"html_url": "https://github.com/thomwolf",
"followers_url": "https://api.github.com/users/thomw... | [] | closed | false | null | [] | null | [] | 1,633,994,866,000 | 1,634,115,354,000 | 1,634,115,354,000 | MEMBER | null | More than just NLP now | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3062/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3062/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3062",
"html_url": "https://github.com/huggingface/datasets/pull/3062",
"diff_url": "https://github.com/huggingface/datasets/pull/3062.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3062.patch",
"merged_at": 1634115353000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3060 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3060/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3060/comments | https://api.github.com/repos/huggingface/datasets/issues/3060/events | https://github.com/huggingface/datasets/issues/3060 | 1,022,936,396 | I_kwDODunzps48-MVM | 3,060 | load_dataset('openwebtext') yields "Compressed file ended before the end-of-stream marker was reached" | {
"login": "RylanSchaeffer",
"id": 8942987,
"node_id": "MDQ6VXNlcjg5NDI5ODc=",
"avatar_url": "https://avatars.githubusercontent.com/u/8942987?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RylanSchaeffer",
"html_url": "https://github.com/RylanSchaeffer",
"followers_url": "https://api.gith... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Hi @RylanSchaeffer, thanks for reporting.\r\n\r\nI'm sorry, but I was not able to reproduce your problem.\r\n\r\nNormally, the reason for this type of error is that, during your download of the data files, this was not fully complete.\r\n\r\nCould you please try to load the dataset again but forcing its redownload... | 1,633,971,927,000 | 1,635,400,341,000 | 1,635,400,341,000 | NONE | null | ## Describe the bug
When I try `load_dataset('openwebtext')`, I receive a "EOFError: Compressed file ended before the end-of-stream marker was reached" error.
## Steps to reproduce the bug
```
from datasets import load_dataset
dataset = load_dataset('openwebtext')
```
## Expected results
I expect the `datas... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3060/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3060/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3059 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3059/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3059/comments | https://api.github.com/repos/huggingface/datasets/issues/3059/events | https://github.com/huggingface/datasets/pull/3059 | 1,022,620,057 | PR_kwDODunzps4tA54w | 3,059 | Fix task reloading from cache | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [] | 1,633,953,784,000 | 1,633,955,019,000 | 1,633,955,019,000 | MEMBER | null | When reloading a dataset from the cache when doing `map`, the tasks templates were kept instead of being updated regarding the output of the `map` function. This is an issue because we drop the tasks templates that are not compatible anymore after `map`, for example if a column of the template was removed.
This PR f... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3059/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 2,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3059/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3059",
"html_url": "https://github.com/huggingface/datasets/pull/3059",
"diff_url": "https://github.com/huggingface/datasets/pull/3059.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3059.patch",
"merged_at": 1633955018000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3057 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3057/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3057/comments | https://api.github.com/repos/huggingface/datasets/issues/3057/events | https://github.com/huggingface/datasets/issues/3057 | 1,022,508,315 | I_kwDODunzps488j0b | 3,057 | Error in per class precision computation | {
"login": "tidhamecha2",
"id": 38906722,
"node_id": "MDQ6VXNlcjM4OTA2NzIy",
"avatar_url": "https://avatars.githubusercontent.com/u/38906722?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tidhamecha2",
"html_url": "https://github.com/tidhamecha2",
"followers_url": "https://api.github.com/... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Hi @tidhamecha2, thanks for reporting.\r\n\r\nIndeed, we fixed this issue just one week ago: #3008\r\n\r\nThe fix will be included in our next version release.\r\n\r\nIn the meantime, you can incorporate the fix by installing `datasets` from the master branch:\r\n```\r\npip install -U git+ssh://git@github.com/hugg... | 1,633,946,719,000 | 1,633,947,464,000 | 1,633,947,376,000 | NONE | null | ## Describe the bug
When trying to get the per class precision values by providing `average=None`, following error is thrown `ValueError: can only convert an array of size 1 to a Python scalar`
## Steps to reproduce the bug
```python
from datasets import load_dataset, load_metric
precision_metric = load_metric("... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3057/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3057/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3056 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3056/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3056/comments | https://api.github.com/repos/huggingface/datasets/issues/3056/events | https://github.com/huggingface/datasets/pull/3056 | 1,022,345,564 | PR_kwDODunzps4tAB9h | 3,056 | Fix meteor metric for version >= 3.6.4 | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,633,936,304,000 | 1,633,937,360,000 | 1,633,937,359,000 | MEMBER | null | After `nltk` update, the meteor metric expects pre-tokenized inputs (breaking change).
This PR fixes this issue, while maintaining compatibility with older versions. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3056/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3056/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3056",
"html_url": "https://github.com/huggingface/datasets/pull/3056",
"diff_url": "https://github.com/huggingface/datasets/pull/3056.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3056.patch",
"merged_at": 1633937359000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3055 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3055/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3055/comments | https://api.github.com/repos/huggingface/datasets/issues/3055/events | https://github.com/huggingface/datasets/issues/3055 | 1,022,319,238 | I_kwDODunzps4871qG | 3,055 | CI test suite fails after meteor metric update | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [] | 1,633,934,232,000 | 1,633,937,431,000 | 1,633,937,431,000 | MEMBER | null | ## Describe the bug
CI test suite fails: https://app.circleci.com/pipelines/github/huggingface/datasets/8110/workflows/f059ba43-9154-4632-bebb-82318447ddc9/jobs/50010
Stack trace:
```
___________________ LocalMetricTest.test_load_metric_meteor ____________________
[gw1] linux -- Python 3.6.15 /home/circleci/.pye... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3055/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3055/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3054 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3054/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3054/comments | https://api.github.com/repos/huggingface/datasets/issues/3054/events | https://github.com/huggingface/datasets/pull/3054 | 1,022,108,186 | PR_kwDODunzps4s_TmE | 3,054 | Update Biosses | {
"login": "bwang482",
"id": 6764450,
"node_id": "MDQ6VXNlcjY3NjQ0NTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/6764450?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bwang482",
"html_url": "https://github.com/bwang482",
"followers_url": "https://api.github.com/users/bwang... | [] | closed | false | null | [] | null | [] | 1,633,904,712,000 | 1,634,115,867,000 | 1,634,115,867,000 | CONTRIBUTOR | null | Fix variable naming | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3054/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3054/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3054",
"html_url": "https://github.com/huggingface/datasets/pull/3054",
"diff_url": "https://github.com/huggingface/datasets/pull/3054.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3054.patch",
"merged_at": 1634115867000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3052 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3052/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3052/comments | https://api.github.com/repos/huggingface/datasets/issues/3052/events | https://github.com/huggingface/datasets/issues/3052 | 1,021,944,435 | I_kwDODunzps486aJz | 3,052 | load_dataset cannot download the data and hangs on forever if cache dir specified | {
"login": "BenoitDalFerro",
"id": 69694610,
"node_id": "MDQ6VXNlcjY5Njk0NjEw",
"avatar_url": "https://avatars.githubusercontent.com/u/69694610?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BenoitDalFerro",
"html_url": "https://github.com/BenoitDalFerro",
"followers_url": "https://api.gi... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [
"Issue was environment inconsistency, updating packages did the trick\r\n\r\n`conda install -c huggingface -c conda-forge datasets`\r\n\r\n> Collecting package metadata (current_repodata.json): done\r\n> Solving environment: |\r\n> The environment is inconsistent, please check the package plan carefully\r\n> The fo... | 1,633,861,896,000 | 1,633,949,829,000 | 1,633,949,796,000 | NONE | null | ## Describe the bug
After updating datasets, a code that ran just fine for ages began to fail. Specifying _datasets.load_dataset_'s _cache_dir_ optional argument on Windows 10 machine results in data download to hang on forever. Same call without cache_dir works just fine. Surprisingly exact same code just runs perfec... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3052/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3052/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3050 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3050/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3050/comments | https://api.github.com/repos/huggingface/datasets/issues/3050/events | https://github.com/huggingface/datasets/pull/3050 | 1,021,772,622 | PR_kwDODunzps4s-anK | 3,050 | Fix streaming: catch Timeout error | {
"login": "borisdayma",
"id": 715491,
"node_id": "MDQ6VXNlcjcxNTQ5MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/715491?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/borisdayma",
"html_url": "https://github.com/borisdayma",
"followers_url": "https://api.github.com/users/b... | [] | closed | false | null | [] | null | [
"I'm running a large test.\r\nLet's see if I get any error within a few days.",
"This time it stopped after 8h but correctly raised `ConnectionError: Server Disconnected`.\r\n\r\nTraceback:\r\n```\r\nTraceback (most recent call last): ... | 1,633,803,560,000 | 1,634,052,498,000 | 1,633,944,938,000 | CONTRIBUTOR | null | Catches Timeout error during streaming.
fix #3049 | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3050/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3050/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3050",
"html_url": "https://github.com/huggingface/datasets/pull/3050",
"diff_url": "https://github.com/huggingface/datasets/pull/3050.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3050.patch",
"merged_at": 1633944938000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3049 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3049/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3049/comments | https://api.github.com/repos/huggingface/datasets/issues/3049/events | https://github.com/huggingface/datasets/issues/3049 | 1,021,770,008 | I_kwDODunzps485vkY | 3,049 | TimeoutError during streaming | {
"login": "borisdayma",
"id": 715491,
"node_id": "MDQ6VXNlcjcxNTQ5MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/715491?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/borisdayma",
"html_url": "https://github.com/borisdayma",
"followers_url": "https://api.github.com/users/b... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | null | [] | null | [] | 1,633,802,811,000 | 1,633,944,938,000 | 1,633,944,938,000 | CONTRIBUTOR | null | ## Describe the bug
I got a TimeoutError after streaming for about 10h.
## Steps to reproduce the bug
Very long code but we could do a test of streaming indefinitely data, though error may take a while to appear.
## Expected results
This error was not expected in the code which considers only `ClientError` but... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3049/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3049/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3047 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3047/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3047/comments | https://api.github.com/repos/huggingface/datasets/issues/3047/events | https://github.com/huggingface/datasets/issues/3047 | 1,021,360,616 | I_kwDODunzps484Lno | 3,047 | Loading from cache a dataset for LM built from a text classification dataset sometimes errors | {
"login": "sgugger",
"id": 35901082,
"node_id": "MDQ6VXNlcjM1OTAxMDgy",
"avatar_url": "https://avatars.githubusercontent.com/u/35901082?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sgugger",
"html_url": "https://github.com/sgugger",
"followers_url": "https://api.github.com/users/sgugge... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [
{
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.git... | null | [
"This has been fixed in 1.15, let me know if you still have this issue"
] | 1,633,717,391,000 | 1,635,959,588,000 | 1,635,959,588,000 | MEMBER | null | ## Describe the bug
Yes, I know, that description sucks. So the problem is arising in the course when we build a masked language modeling dataset using the IMDB dataset. To reproduce (or try since it's a bit fickle).
Create a dataset for masled-language modeling from the IMDB dataset.
```python
from datasets ... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3047/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3047/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3046 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3046/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3046/comments | https://api.github.com/repos/huggingface/datasets/issues/3046/events | https://github.com/huggingface/datasets/pull/3046 | 1,021,021,368 | PR_kwDODunzps4s8MjS | 3,046 | Fix MedDialog metadata JSON | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [] | closed | false | null | [] | null | [] | 1,633,694,680,000 | 1,633,938,403,000 | 1,633,938,402,000 | MEMBER | null | Fix #2969. | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3046/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3046/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3046",
"html_url": "https://github.com/huggingface/datasets/pull/3046",
"diff_url": "https://github.com/huggingface/datasets/pull/3046.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3046.patch",
"merged_at": 1633938402000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3045 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3045/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3045/comments | https://api.github.com/repos/huggingface/datasets/issues/3045/events | https://github.com/huggingface/datasets/pull/3045 | 1,020,968,704 | PR_kwDODunzps4s8B2b | 3,045 | Fix inconsistent caching behaviour in Dataset.map() with multiprocessing #3044 | {
"login": "vlievin",
"id": 9859840,
"node_id": "MDQ6VXNlcjk4NTk4NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/9859840?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vlievin",
"html_url": "https://github.com/vlievin",
"followers_url": "https://api.github.com/users/vlievin/... | [] | closed | false | null | [] | null | [
"Hi ! Thanks for noticing this inconsistence and suggesting a fix :)\r\n\r\nIf I understand correctly you try to pass the same fingerprint to each processed shard of the dataset. This can be an issue since each shard is actually a different dataset with different data: they shouldn't have the same fingerprint.\r\n\... | 1,633,690,761,000 | 1,634,835,512,000 | 1,634,826,164,000 | NONE | null | Fix #3044
1. A rough unit test that fails without the fix. It probably doesn't comply with your code standards, but that just to draft the idea.
2. A one liner fix | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3045/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3045/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3045",
"html_url": "https://github.com/huggingface/datasets/pull/3045",
"diff_url": "https://github.com/huggingface/datasets/pull/3045.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3045.patch",
"merged_at": null
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3041 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3041/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3041/comments | https://api.github.com/repos/huggingface/datasets/issues/3041/events | https://github.com/huggingface/datasets/pull/3041 | 1,018,911,385 | PR_kwDODunzps4s1ZAc | 3,041 | Load private data files + use glob on ZIP archives for json/csv/etc. module inference | {
"login": "lhoestq",
"id": 42851186,
"node_id": "MDQ6VXNlcjQyODUxMTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lhoestq",
"html_url": "https://github.com/lhoestq",
"followers_url": "https://api.github.com/users/lhoest... | [] | closed | false | null | [] | null | [
"I have an error on windows:\r\n```python\r\naiohttp.client_exceptions.ClientConnectorCertificateError: Cannot connect to host moon-staging.huggingface.co:443 ssl:True [SSLCertVerificationError: (1, '[SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed: certificate has expired (_ssl.c:1131)')]\r\n```\r\nat th... | 1,633,544,196,000 | 1,634,052,348,000 | 1,634,052,346,000 | MEMBER | null | As mentioned in https://github.com/huggingface/datasets/issues/3032 loading data files from private repository isn't working correctly because of the data files resolved.
#2986 did a refactor of the data files resolver. I added authentication to it.
I also improved it to glob inside ZIP archives to look for json/... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3041/reactions",
"total_count": 2,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3041/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3041",
"html_url": "https://github.com/huggingface/datasets/pull/3041",
"diff_url": "https://github.com/huggingface/datasets/pull/3041.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3041.patch",
"merged_at": 1634052346000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3040 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3040/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3040/comments | https://api.github.com/repos/huggingface/datasets/issues/3040/events | https://github.com/huggingface/datasets/issues/3040 | 1,018,782,475 | I_kwDODunzps48uWML | 3,040 | [save_to_disk] Using `select()` followed by `save_to_disk` saves complete dataset making it hard to create dummy dataset | {
"login": "patrickvonplaten",
"id": 23423619,
"node_id": "MDQ6VXNlcjIzNDIzNjE5",
"avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/patrickvonplaten",
"html_url": "https://github.com/patrickvonplaten",
"followers_url": "https://... | [
{
"id": 1935892857,
"node_id": "MDU6TGFiZWwxOTM1ODkyODU3",
"url": "https://api.github.com/repos/huggingface/datasets/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] | closed | false | {
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_url": "https://api.g... | [
{
"login": "albertvillanova",
"id": 8515462,
"node_id": "MDQ6VXNlcjg1MTU0NjI=",
"avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/albertvillanova",
"html_url": "https://github.com/albertvillanova",
"followers_... | null | [
"Hi,\r\n\r\nthe `save_to_disk` docstring explains that `flatten_indices` has to be called on a dataset before saving it to save only the shard/slice of the dataset.",
"That works! Thansk!\r\n\r\nMight be worth doing that automatically actually in case the `save_to_disk` is called on a dataset that has an indices ... | 1,633,540,127,000 | 1,635,867,668,000 | 1,635,867,668,000 | MEMBER | null | ## Describe the bug
When only keeping a dummy size of a dataset (say the first 100 samples), and then saving it to disk to upload it in the following to the hub for easy demo/use - not just the small dataset is saved but the whole dataset with an indices file. The problem with this is that the dataset is still very... | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3040/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3040/timeline | null | null | null | false |
https://api.github.com/repos/huggingface/datasets/issues/3039 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3039/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3039/comments | https://api.github.com/repos/huggingface/datasets/issues/3039/events | https://github.com/huggingface/datasets/pull/3039 | 1,018,219,800 | PR_kwDODunzps4sy_J- | 3,039 | Add sberquad dataset | {
"login": "Alenush",
"id": 13781234,
"node_id": "MDQ6VXNlcjEzNzgxMjM0",
"avatar_url": "https://avatars.githubusercontent.com/u/13781234?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Alenush",
"html_url": "https://github.com/Alenush",
"followers_url": "https://api.github.com/users/Alenus... | [] | closed | false | null | [] | null | [] | 1,633,523,522,000 | 1,634,120,351,000 | 1,634,120,164,000 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3039/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3039/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3039",
"html_url": "https://github.com/huggingface/datasets/pull/3039",
"diff_url": "https://github.com/huggingface/datasets/pull/3039.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3039.patch",
"merged_at": 1634120164000
} | true |
https://api.github.com/repos/huggingface/datasets/issues/3038 | https://api.github.com/repos/huggingface/datasets | https://api.github.com/repos/huggingface/datasets/issues/3038/labels{/name} | https://api.github.com/repos/huggingface/datasets/issues/3038/comments | https://api.github.com/repos/huggingface/datasets/issues/3038/events | https://github.com/huggingface/datasets/pull/3038 | 1,018,113,499 | PR_kwDODunzps4syno_ | 3,038 | add sberquad dataset | {
"login": "Alenush",
"id": 13781234,
"node_id": "MDQ6VXNlcjEzNzgxMjM0",
"avatar_url": "https://avatars.githubusercontent.com/u/13781234?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Alenush",
"html_url": "https://github.com/Alenush",
"followers_url": "https://api.github.com/users/Alenus... | [] | closed | false | null | [] | null | [] | 1,633,520,019,000 | 1,633,521,481,000 | 1,633,521,481,000 | CONTRIBUTOR | null | null | {
"url": "https://api.github.com/repos/huggingface/datasets/issues/3038/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
} | https://api.github.com/repos/huggingface/datasets/issues/3038/timeline | null | false | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/3038",
"html_url": "https://github.com/huggingface/datasets/pull/3038",
"diff_url": "https://github.com/huggingface/datasets/pull/3038.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/3038.patch",
"merged_at": null
} | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.