url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1.15B
node_id
stringlengths
18
32
number
int64
1
3.78k
title
stringlengths
1
276
user
dict
labels
listlengths
0
3
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
listlengths
0
30
created_at
int64
1,587B
1,646B
updated_at
int64
1,587B
1,646B
closed_at
int64
1,587B
1,646B
author_association
stringclasses
3 values
active_lock_reason
null
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
draft
bool
2 classes
pull_request
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/2153
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2153/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2153/comments
https://api.github.com/repos/huggingface/datasets/issues/2153/events
https://github.com/huggingface/datasets/issues/2153
846,181,502
MDU6SXNzdWU4NDYxODE1MDI=
2,153
load_dataset ignoring features
{ "login": "GuillemGSubies", "id": 37592763, "node_id": "MDQ6VXNlcjM3NTkyNzYz", "avatar_url": "https://avatars.githubusercontent.com/u/37592763?v=4", "gravatar_id": "", "url": "https://api.github.com/users/GuillemGSubies", "html_url": "https://github.com/GuillemGSubies", "followers_url": "https://api.gi...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! Thanks for reporting. I opened a PR to fix this issue: #2201", "Nice question which helped me a lot! I have wasted a lot of time to the `DatasetDict` creation from a csv file. Hope the document of this module add some simple examples.", "Hi :) We're indeed working on tutorials that we will add to the docs...
1,617,179,409,000
1,630,077,838,000
null
NONE
null
First of all, I'm sorry if it is a repeated issue or the changes are already in master, I searched and I didn't find anything. I'm using datasets 1.5.0 ![image](https://user-images.githubusercontent.com/37592763/113114369-8f376580-920b-11eb-900d-94365b59f04b.png) As you can see, when I load the dataset, the C...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2153/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2153/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2152
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2152/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2152/comments
https://api.github.com/repos/huggingface/datasets/issues/2152/events
https://github.com/huggingface/datasets/pull/2152
845,751,273
MDExOlB1bGxSZXF1ZXN0NjA0ODk0MDkz
2,152
Update README.md
{ "login": "JieyuZhao", "id": 22306304, "node_id": "MDQ6VXNlcjIyMzA2MzA0", "avatar_url": "https://avatars.githubusercontent.com/u/22306304?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JieyuZhao", "html_url": "https://github.com/JieyuZhao", "followers_url": "https://api.github.com/users/...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,617,160,879,000
1,617,272,437,000
1,617,272,436,000
CONTRIBUTOR
null
Updated some descriptions of Wino_Bias dataset.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2152/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2152/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2152", "html_url": "https://github.com/huggingface/datasets/pull/2152", "diff_url": "https://github.com/huggingface/datasets/pull/2152.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2152.patch", "merged_at": 1617272436000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2151
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2151/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2151/comments
https://api.github.com/repos/huggingface/datasets/issues/2151/events
https://github.com/huggingface/datasets/pull/2151
844,886,081
MDExOlB1bGxSZXF1ZXN0NjA0MDg5MDMw
2,151
Add support for axis in concatenate datasets
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/milestones/1", "html_url": "https://github.com/huggingface/datasets/milestone/1", "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/1/labels", "id": 6644198, "node_id": "MDk6TWlsZXN0b25lNjY0NDE5OA==", "number": 1, "title...
[ "@lhoestq I am going to implement the consolidation step you mentioned in #1870.", "@lhoestq I was thinking that the order of the TableBlocks is not relevant, isn't it?\r\n\r\nI mean, in order to consolidate _consecutive_ in-memory table blocks, in this case:\r\n```\r\nblocks = [in_memory_1, memory_mapped, in_mem...
1,617,123,524,000
1,624,470,062,000
1,618,848,438,000
MEMBER
null
Add support for `axis` (0 or 1) in `concatenate_datasets`. Close #853.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2151/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2151/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2151", "html_url": "https://github.com/huggingface/datasets/pull/2151", "diff_url": "https://github.com/huggingface/datasets/pull/2151.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2151.patch", "merged_at": 1618848438000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2150
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2150/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2150/comments
https://api.github.com/repos/huggingface/datasets/issues/2150/events
https://github.com/huggingface/datasets/pull/2150
844,776,448
MDExOlB1bGxSZXF1ZXN0NjAzOTg3OTcx
2,150
Allow pickling of big in-memory tables
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,617,119,516,000
1,617,187,035,000
1,617,187,034,000
MEMBER
null
This should fix issue #2134 Pickling is limited to <4GiB objects, it's not possible to pickle a big arrow table (for multiprocessing for example). For big tables, we have to write them on disk and only pickle the path to the table.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2150/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2150/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2150", "html_url": "https://github.com/huggingface/datasets/pull/2150", "diff_url": "https://github.com/huggingface/datasets/pull/2150.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2150.patch", "merged_at": 1617187034000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2149
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2149/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2149/comments
https://api.github.com/repos/huggingface/datasets/issues/2149/events
https://github.com/huggingface/datasets/issues/2149
844,734,076
MDU6SXNzdWU4NDQ3MzQwNzY=
2,149
Telugu subset missing for xtreme tatoeba dataset
{ "login": "jerryIsHere", "id": 50871412, "node_id": "MDQ6VXNlcjUwODcxNDEy", "avatar_url": "https://avatars.githubusercontent.com/u/50871412?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jerryIsHere", "html_url": "https://github.com/jerryIsHere", "followers_url": "https://api.github.com/...
[]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Good catch ! Thanks for reporting\r\n\r\nI just opened #2180 to fix this" ]
1,617,117,994,000
1,617,791,015,000
null
CONTRIBUTOR
null
from nlp import load_dataset train_dataset = load_dataset('xtreme', 'tatoeba.tel')['validation'] ValueError: BuilderConfig tatoeba.tel not found. but language tel is actually included in xtreme: https://github.com/google-research/xtreme/blob/master/utils_preprocess.py def tatoeba_preprocess(args): lang3_dict ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2149/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2149/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2148
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2148/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2148/comments
https://api.github.com/repos/huggingface/datasets/issues/2148/events
https://github.com/huggingface/datasets/issues/2148
844,700,910
MDU6SXNzdWU4NDQ3MDA5MTA=
2,148
Add configurable options to `seqeval` metric
{ "login": "marrodion", "id": 44571847, "node_id": "MDQ6VXNlcjQ0NTcxODQ3", "avatar_url": "https://avatars.githubusercontent.com/u/44571847?v=4", "gravatar_id": "", "url": "https://api.github.com/users/marrodion", "html_url": "https://github.com/marrodion", "followers_url": "https://api.github.com/users/...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi @marrodion. \r\n\r\nThanks for pointing this out. It would be great to incorporate this metric-specific enhancement.\r\n\r\nAnother possibility would be to require the user to input the scheme as a string `mode=\"strict\", scheme=\"IOB2\"` and then dynamically import the corresponding module using Python `impor...
1,617,116,646,000
1,618,494,586,000
1,618,494,586,000
CONTRIBUTOR
null
Right now `load_metric("seqeval")` only works in the default mode of evaluation (equivalent to conll evaluation). However, seqeval library [supports](https://github.com/chakki-works/seqeval#support-features) different evaluation schemes (IOB1, IOB2, etc.), which can be plugged in just by supporting additional kwargs...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2148/reactions", "total_count": 2, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2148/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2147
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2147/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2147/comments
https://api.github.com/repos/huggingface/datasets/issues/2147/events
https://github.com/huggingface/datasets/pull/2147
844,687,831
MDExOlB1bGxSZXF1ZXN0NjAzOTA3NjM4
2,147
Render docstring return type as inline
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,617,116,143,000
1,617,196,265,000
1,617,196,265,000
MEMBER
null
This documentation setting will avoid having the return type in a separate line under `Return type`. See e.g. current docs for `Dataset.to_csv`.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2147/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2147/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2147", "html_url": "https://github.com/huggingface/datasets/pull/2147", "diff_url": "https://github.com/huggingface/datasets/pull/2147.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2147.patch", "merged_at": 1617196265000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2146
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2146/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2146/comments
https://api.github.com/repos/huggingface/datasets/issues/2146/events
https://github.com/huggingface/datasets/issues/2146
844,673,244
MDU6SXNzdWU4NDQ2NzMyNDQ=
2,146
Dataset file size on disk is very large with 3D Array
{ "login": "jblemoine", "id": 22685854, "node_id": "MDQ6VXNlcjIyNjg1ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/22685854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jblemoine", "html_url": "https://github.com/jblemoine", "followers_url": "https://api.github.com/users/...
[]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! In the arrow file we store all the integers as uint8.\r\nSo your arrow file should weigh around `height x width x n_channels x n_images` bytes.\r\n\r\nWhat feature type do your TFDS dataset have ?\r\n\r\nIf it uses a `tfds.features.Image` type, then what is stored is the encoded data (as png or jpg for exampl...
1,617,115,569,000
1,618,578,422,000
null
NONE
null
Hi, I have created my own dataset using the provided dataset loading script. It is an image dataset where images are stored as 3D Array with dtype=uint8. The actual size on disk is surprisingly large. It takes 520 MB. Here is some info from `dataset_info.json`. `{ "description": "", "citation": ""...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2146/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2146/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2145
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2145/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2145/comments
https://api.github.com/repos/huggingface/datasets/issues/2145/events
https://github.com/huggingface/datasets/pull/2145
844,603,518
MDExOlB1bGxSZXF1ZXN0NjAzODMxOTE2
2,145
Implement Dataset add_column
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/milestones/3", "html_url": "https://github.com/huggingface/datasets/milestone/3", "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/3/labels", "id": 6644287, "node_id": "MDk6TWlsZXN0b25lNjY0NDI4Nw==", "number": 3, "title...
[ "#2274 has been merged. You can now merge master into this branch and use `assert_arrow_metadata_are_synced_with_dataset_features(dset)` to make sure that the metadata are good :)" ]
1,617,112,934,000
1,619,707,844,000
1,619,707,843,000
MEMBER
null
Implement `Dataset.add_column`. Close #1954.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2145/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2145/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2145", "html_url": "https://github.com/huggingface/datasets/pull/2145", "diff_url": "https://github.com/huggingface/datasets/pull/2145.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2145.patch", "merged_at": 1619707843000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2144
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2144/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2144/comments
https://api.github.com/repos/huggingface/datasets/issues/2144/events
https://github.com/huggingface/datasets/issues/2144
844,352,067
MDU6SXNzdWU4NDQzNTIwNjc=
2,144
Loading wikipedia 20200501.en throws pyarrow related error
{ "login": "TomPyonsuke", "id": 26637405, "node_id": "MDQ6VXNlcjI2NjM3NDA1", "avatar_url": "https://avatars.githubusercontent.com/u/26637405?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TomPyonsuke", "html_url": "https://github.com/TomPyonsuke", "followers_url": "https://api.github.com/...
[]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "That's how I loaded the dataset\r\n```python\r\nfrom datasets import load_dataset\r\nds = load_dataset('wikipedia', '20200501.en', cache_dir='/usr/local/workspace/NAS_NLP/cache')\r\n```", "Hi ! It looks like the arrow file in the folder\r\n`/usr/local/workspace/NAS_NLP/cache/wikipedia/20200501.en/1.0.0/50aa706aa...
1,617,100,711,000
1,617,268,877,000
null
NONE
null
**Problem description** I am getting the following error when trying to load wikipedia/20200501.en dataset. **Error log** Downloading and preparing dataset wikipedia/20200501.en (download: 16.99 GiB, generated: 17.07 GiB, post-processed: Unknown size, total: 34.06 GiB) to /usr/local/workspace/NAS_NLP/cache/wikiped...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2144/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2144/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2143
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2143/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2143/comments
https://api.github.com/repos/huggingface/datasets/issues/2143/events
https://github.com/huggingface/datasets/pull/2143
844,313,228
MDExOlB1bGxSZXF1ZXN0NjAzNTc0NjI0
2,143
task casting via load_dataset
{ "login": "theo-m", "id": 17948980, "node_id": "MDQ6VXNlcjE3OTQ4OTgw", "avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theo-m", "html_url": "https://github.com/theo-m", "followers_url": "https://api.github.com/users/theo-m/fo...
[]
closed
false
{ "login": "theo-m", "id": 17948980, "node_id": "MDQ6VXNlcjE3OTQ4OTgw", "avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theo-m", "html_url": "https://github.com/theo-m", "followers_url": "https://api.github.com/users/theo-m/fo...
[ { "login": "theo-m", "id": 17948980, "node_id": "MDQ6VXNlcjE3OTQ4OTgw", "avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theo-m", "html_url": "https://github.com/theo-m", "followers_url": "https://api.github...
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,617,098,442,000
1,623,417,641,000
1,623,417,636,000
CONTRIBUTOR
null
wip not satisfied with the API, it means as a dataset implementer I need to write a function with boilerplate and write classes for each `<dataset><task>` "facet".
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2143/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2143/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2143", "html_url": "https://github.com/huggingface/datasets/pull/2143", "diff_url": "https://github.com/huggingface/datasets/pull/2143.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2143.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/2142
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2142/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2142/comments
https://api.github.com/repos/huggingface/datasets/issues/2142/events
https://github.com/huggingface/datasets/pull/2142
843,919,420
MDExOlB1bGxSZXF1ZXN0NjAzMjQwMzUy
2,142
Gem V1.1
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,617,061,622,000
1,617,063,002,000
1,617,063,002,000
MEMBER
null
This branch updates the GEM benchmark to its 1.1 version which includes: - challenge sets for most tasks - detokenized TurkCorpus to match the rest of the text simplification subtasks - fixed inputs for TurkCorpus and ASSET test sets - 18 languages in WikiLingua cc @sebastianGehrmann
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2142/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2142/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2142", "html_url": "https://github.com/huggingface/datasets/pull/2142", "diff_url": "https://github.com/huggingface/datasets/pull/2142.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2142.patch", "merged_at": 1617063002000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2141
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2141/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2141/comments
https://api.github.com/repos/huggingface/datasets/issues/2141/events
https://github.com/huggingface/datasets/pull/2141
843,914,790
MDExOlB1bGxSZXF1ZXN0NjAzMjM2MjUw
2,141
added spans field for the wikiann datasets
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi @lhoestq \r\nThanks a lot for taking time checking it. I update \"dataset_infos.json\", I added description to the function of _generate_samples in wikiann.py but I was not sure about the format to write in README. thanks. ", "Thanks !\r\n\r\nFor the fields description in the dataset card, something like thi...
1,617,061,106,000
1,617,197,270,000
1,617,197,270,000
CONTRIBUTOR
null
Hi @lhoestq I tried to add spans to the wikiann datasets. Thanks a lot for kindly having a look. This addresses https://github.com/huggingface/datasets/issues/2130. Best regards Rabeeh
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2141/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2141/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2141", "html_url": "https://github.com/huggingface/datasets/pull/2141", "diff_url": "https://github.com/huggingface/datasets/pull/2141.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2141.patch", "merged_at": 1617197270000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2140
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2140/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2140/comments
https://api.github.com/repos/huggingface/datasets/issues/2140/events
https://github.com/huggingface/datasets/pull/2140
843,830,451
MDExOlB1bGxSZXF1ZXN0NjAzMTYxMjYx
2,140
add banking77 dataset
{ "login": "dkajtoch", "id": 32985207, "node_id": "MDQ6VXNlcjMyOTg1MjA3", "avatar_url": "https://avatars.githubusercontent.com/u/32985207?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dkajtoch", "html_url": "https://github.com/dkajtoch", "followers_url": "https://api.github.com/users/dka...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "@lhoestq I updated files" ]
1,617,053,543,000
1,617,960,738,000
1,617,960,738,000
CONTRIBUTOR
null
Intent classification/detection dataset from banking category with 77 unique intents.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2140/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2140/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2140", "html_url": "https://github.com/huggingface/datasets/pull/2140", "diff_url": "https://github.com/huggingface/datasets/pull/2140.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2140.patch", "merged_at": 1617960738000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2139
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2139/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2139/comments
https://api.github.com/repos/huggingface/datasets/issues/2139/events
https://github.com/huggingface/datasets/issues/2139
843,662,613
MDU6SXNzdWU4NDM2NjI2MTM=
2,139
TypeError when using save_to_disk in a dataset loaded with ReadInstruction split
{ "login": "PedroMLF", "id": 22480495, "node_id": "MDQ6VXNlcjIyNDgwNDk1", "avatar_url": "https://avatars.githubusercontent.com/u/22480495?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PedroMLF", "html_url": "https://github.com/PedroMLF", "followers_url": "https://api.github.com/users/Ped...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi !\r\nI think this has been fixed recently on `master`.\r\nCan you try again by installing `datasets` from `master` ?\r\n```\r\npip install git+https://github.com/huggingface/datasets.git\r\n```", "Hi!\r\n\r\nUsing that version of the code solves the issue. Thanks!" ]
1,617,042,234,000
1,617,095,573,000
1,617,095,573,000
NONE
null
Hi, Loading a dataset with `load_dataset` using a split defined via `ReadInstruction` and then saving it to disk results in the following error: `TypeError: Object of type ReadInstruction is not JSON serializable`. Here is the minimal reproducible example: ```python from datasets import load_dataset from dat...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2139/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2139/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2138
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2138/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2138/comments
https://api.github.com/repos/huggingface/datasets/issues/2138/events
https://github.com/huggingface/datasets/pull/2138
843,508,402
MDExOlB1bGxSZXF1ZXN0NjAyODc4NzU2
2,138
Add CER metric
{ "login": "chutaklee", "id": 6931004, "node_id": "MDQ6VXNlcjY5MzEwMDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/6931004?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chutaklee", "html_url": "https://github.com/chutaklee", "followers_url": "https://api.github.com/users/ch...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,617,033,147,000
1,617,725,771,000
1,617,693,278,000
CONTRIBUTOR
null
Add Character Error Rate (CER) metric that is used in evaluation in ASR. I also have written unittests (hopefully thorough enough) but I'm not sure how to integrate them into the existed codebase. ```python from cer import CER cer = CER() class TestCER(unittest.TestCase): def test_cer_case_senstive(self)...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2138/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2138/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2138", "html_url": "https://github.com/huggingface/datasets/pull/2138", "diff_url": "https://github.com/huggingface/datasets/pull/2138.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2138.patch", "merged_at": 1617693278000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2137
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2137/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2137/comments
https://api.github.com/repos/huggingface/datasets/issues/2137/events
https://github.com/huggingface/datasets/pull/2137
843,502,835
MDExOlB1bGxSZXF1ZXN0NjAyODc0MDYw
2,137
Fix missing infos from concurrent dataset loading
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,617,032,772,000
1,617,186,956,000
1,617,186,955,000
MEMBER
null
This should fix issue #2131 When calling `load_dataset` at the same time from 2 workers, one of the worker could have missing split infos when reloading the dataset from the cache.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2137/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2137/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2137", "html_url": "https://github.com/huggingface/datasets/pull/2137", "diff_url": "https://github.com/huggingface/datasets/pull/2137.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2137.patch", "merged_at": 1617186955000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2136
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2136/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2136/comments
https://api.github.com/repos/huggingface/datasets/issues/2136/events
https://github.com/huggingface/datasets/pull/2136
843,492,015
MDExOlB1bGxSZXF1ZXN0NjAyODY0ODY5
2,136
fix dialogue action slot name and value
{ "login": "adamlin120", "id": 31605305, "node_id": "MDQ6VXNlcjMxNjA1MzA1", "avatar_url": "https://avatars.githubusercontent.com/u/31605305?v=4", "gravatar_id": "", "url": "https://api.github.com/users/adamlin120", "html_url": "https://github.com/adamlin120", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,617,032,053,000
1,617,194,882,000
1,617,194,881,000
CONTRIBUTOR
null
fix #2128
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2136/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2136/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2136", "html_url": "https://github.com/huggingface/datasets/pull/2136", "diff_url": "https://github.com/huggingface/datasets/pull/2136.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2136.patch", "merged_at": 1617194881000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2135
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2135/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2135/comments
https://api.github.com/repos/huggingface/datasets/issues/2135/events
https://github.com/huggingface/datasets/issues/2135
843,246,344
MDU6SXNzdWU4NDMyNDYzNDQ=
2,135
en language data from MLQA dataset is missing
{ "login": "rabeehk", "id": 6278280, "node_id": "MDQ6VXNlcjYyNzgyODA=", "avatar_url": "https://avatars.githubusercontent.com/u/6278280?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rabeehk", "html_url": "https://github.com/rabeehk", "followers_url": "https://api.github.com/users/rabeehk/...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! Indeed only the languages of the `translate-train` data are included...\r\nI can't find a link to download the english train set on https://github.com/facebookresearch/MLQA though, do you know where we can download it ?", "Hi @lhoestq \r\nthank you very much for coming back to me, now I see, you are right, ...
1,617,014,870,000
1,617,099,623,000
1,617,099,623,000
CONTRIBUTOR
null
Hi I need mlqa-translate-train.en dataset, but it is missing from the MLQA dataset. could you have a look please? @lhoestq thank you for your help to fix this issue.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2135/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2135/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2134
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2134/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2134/comments
https://api.github.com/repos/huggingface/datasets/issues/2134/events
https://github.com/huggingface/datasets/issues/2134
843,242,849
MDU6SXNzdWU4NDMyNDI4NDk=
2,134
Saving large in-memory datasets with save_to_disk crashes because of pickling
{ "login": "prokopCerny", "id": 5815801, "node_id": "MDQ6VXNlcjU4MTU4MDE=", "avatar_url": "https://avatars.githubusercontent.com/u/5815801?v=4", "gravatar_id": "", "url": "https://api.github.com/users/prokopCerny", "html_url": "https://github.com/prokopCerny", "followers_url": "https://api.github.com/us...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi !\r\nIndeed `save_to_disk` doesn't call pickle anymore. Though the `OverflowError` can still appear for in-memory datasets bigger than 4GB. This happens when doing this for example:\r\n```python\r\nimport pyarrow as pa\r\nimport pickle\r\n\r\narr = pa.array([0] * ((4 * 8 << 30) // 64))\r\ntable = pa.Table.from_...
1,617,014,595,000
1,620,064,761,000
1,620,064,761,000
NONE
null
Using Datasets 1.5.0 on Python 3.7. Recently I've been working on medium to large size datasets (pretokenized raw text sizes from few gigabytes to low tens of gigabytes), and have found out that several preprocessing steps are massively faster when done in memory, and I have the ability to requisition a lot of RAM, so...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2134/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2134/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2133
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2133/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2133/comments
https://api.github.com/repos/huggingface/datasets/issues/2133/events
https://github.com/huggingface/datasets/issues/2133
843,149,680
MDU6SXNzdWU4NDMxNDk2ODA=
2,133
bug in mlqa dataset
{ "login": "dorost1234", "id": 79165106, "node_id": "MDQ6VXNlcjc5MTY1MTA2", "avatar_url": "https://avatars.githubusercontent.com/u/79165106?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dorost1234", "html_url": "https://github.com/dorost1234", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "If you print those questions, you get readable texts:\r\n```python\r\n>>> questions = [\r\n... \"\\u0645\\u062a\\u0649 \\u0628\\u062f\\u0627\\u062a \\u0627\\u0644\\u0645\\u062c\\u0644\\u0629 \\u0627\\u0644\\u0645\\u062f\\u0631\\u0633\\u064a\\u0629 \\u0641\\u064a \\u0646\\u0648\\u062a\\u0631\\u062f\\u0627\\u064...
1,617,008,589,000
1,617,126,057,000
1,617,126,057,000
NONE
null
Hi Looking into MLQA dataset for langauge "ar": ``` "question": [ "\u0645\u062a\u0649 \u0628\u062f\u0627\u062a \u0627\u0644\u0645\u062c\u0644\u0629 \u0627\u0644\u0645\u062f\u0631\u0633\u064a\u0629 \u0641\u064a \u0646\u0648\u062a\u0631\u062f\u0627\u0645 \u0628\u0627\u0644\u0646\u0634\u0631?", "\u0643\u0...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2133/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2133/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2132
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2132/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2132/comments
https://api.github.com/repos/huggingface/datasets/issues/2132/events
https://github.com/huggingface/datasets/issues/2132
843,142,822
MDU6SXNzdWU4NDMxNDI4MjI=
2,132
TydiQA dataset is mixed and is not split per language
{ "login": "dorost1234", "id": 79165106, "node_id": "MDQ6VXNlcjc5MTY1MTA2", "avatar_url": "https://avatars.githubusercontent.com/u/79165106?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dorost1234", "html_url": "https://github.com/dorost1234", "followers_url": "https://api.github.com/use...
[]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "You can filter the languages this way:\r\n```python\r\ntydiqa_en = tydiqa_dataset.filter(lambda x: x[\"language\"] == \"english\")\r\n```\r\n\r\nOtherwise maybe we can have one configuration per language ?\r\nWhat do you think of this for example ?\r\n\r\n```python\r\nload_dataset(\"tydiqa\", \"primary_task.en\")\...
1,617,008,181,000
1,617,530,235,000
null
NONE
null
Hi @lhoestq Currently TydiQA is mixed and user can only access the whole training set of all languages: https://www.tensorflow.org/datasets/catalog/tydi_qa for using this dataset, one need to train/evaluate in each separate language, and having them mixed, makes it hard to use this dataset. This is much convenien...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2132/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2132/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2131
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2131/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2131/comments
https://api.github.com/repos/huggingface/datasets/issues/2131/events
https://github.com/huggingface/datasets/issues/2131
843,133,112
MDU6SXNzdWU4NDMxMzMxMTI=
2,131
When training with Multi-Node Multi-GPU the worker 2 has TypeError: 'NoneType' object
{ "login": "andy-yangz", "id": 23011317, "node_id": "MDQ6VXNlcjIzMDExMzE3", "avatar_url": "https://avatars.githubusercontent.com/u/23011317?v=4", "gravatar_id": "", "url": "https://api.github.com/users/andy-yangz", "html_url": "https://github.com/andy-yangz", "followers_url": "https://api.github.com/use...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! Thanks for reporting\r\nI was able to reproduce this issue. This was caused by missing split infos if a worker reloads the cache of the other worker.\r\n\r\nI just opened https://github.com/huggingface/datasets/pull/2137 to fix this issue", "The PR got merged :)\r\nFeel free to try it out on the `master` br...
1,617,007,558,000
1,618,052,935,000
1,618,052,935,000
NONE
null
version: 1.5.0 met a very strange error, I am training large scale language model, and need train on 2 machines(workers). And sometimes I will get this error `TypeError: 'NoneType' object is not iterable` This is traceback ``` 71 |   | Traceback (most recent call last): -- | -- | -- 72 |   | File "run_gpt.py"...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2131/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 1, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2131/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2130
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2130/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2130/comments
https://api.github.com/repos/huggingface/datasets/issues/2130/events
https://github.com/huggingface/datasets/issues/2130
843,111,936
MDU6SXNzdWU4NDMxMTE5MzY=
2,130
wikiann dataset is missing columns
{ "login": "dorost1234", "id": 79165106, "node_id": "MDQ6VXNlcjc5MTY1MTA2", "avatar_url": "https://avatars.githubusercontent.com/u/79165106?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dorost1234", "html_url": "https://github.com/dorost1234", "followers_url": "https://api.github.com/use...
[ { "id": 1935892877, "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue", "name": "good first issue", "color": "7057ff", "default": true, "description": "Good for newcomers" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Here please find TFDS format of this dataset: https://www.tensorflow.org/datasets/catalog/wikiann\r\nwhere there is a span column, this is really necessary to be able to use the data, and I appreciate your help @lhoestq ", "Hi !\r\nApparently you can get the spans from the NER tags using `tags_to_spans` defined ...
1,617,006,180,000
1,630,075,458,000
1,630,075,458,000
NONE
null
Hi Wikiann dataset needs to have "spans" columns, which is necessary to be able to use this dataset, but this column is missing from huggingface datasets, could you please have a look? thank you @lhoestq
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2130/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2130/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2129
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2129/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2129/comments
https://api.github.com/repos/huggingface/datasets/issues/2129/events
https://github.com/huggingface/datasets/issues/2129
843,033,656
MDU6SXNzdWU4NDMwMzM2NTY=
2,129
How to train BERT model with next sentence prediction?
{ "login": "jnishi", "id": 836541, "node_id": "MDQ6VXNlcjgzNjU0MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/836541?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jnishi", "html_url": "https://github.com/jnishi", "followers_url": "https://api.github.com/users/jnishi/follow...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi !\r\nWe're not using `TextDatasetForNextSentencePrediction` in `datasets`.\r\nAlthough you can probably use the `TextDatasetForNextSentencePrediction.create_examples_from_document` on a dataset to prepare it for next sentence prediction.", "Thanks.\r\n\r\nDo you mean that `TextDatasetForNextSentencePrediction...
1,617,000,483,000
1,617,253,120,000
1,617,253,120,000
NONE
null
Hello. I'm trying to pretrain the BERT model with next sentence prediction. Is there any function that supports next sentence prediction like ` TextDatasetForNextSentencePrediction` of `huggingface/transformers` ?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2129/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2129/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2128
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2128/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2128/comments
https://api.github.com/repos/huggingface/datasets/issues/2128/events
https://github.com/huggingface/datasets/issues/2128
843,023,910
MDU6SXNzdWU4NDMwMjM5MTA=
2,128
Dialogue action slot name and value are reversed in MultiWoZ 2.2
{ "login": "adamlin120", "id": 31605305, "node_id": "MDQ6VXNlcjMxNjA1MzA1", "avatar_url": "https://avatars.githubusercontent.com/u/31605305?v=4", "gravatar_id": "", "url": "https://api.github.com/users/adamlin120", "html_url": "https://github.com/adamlin120", "followers_url": "https://api.github.com/use...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi\r\nGood catch ! Thanks for reporting\r\n\r\nIf you are interested in contributing, feel free to open a PR to fix this :) " ]
1,616,999,642,000
1,617,194,881,000
1,617,194,881,000
CONTRIBUTOR
null
Hi @yjernite, thank you for adding MultiWoZ 2.2 in the huggingface datasets platform. It is beneficial! I spot an error that the order of Dialogue action slot names and values are reversed. https://github.com/huggingface/datasets/blob/649b2c469779bc4221e1b6969aa2496d63eb5953/datasets/multi_woz_v22/multi_woz_v22.p...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2128/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/huggingface/datasets/issues/2128/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2127
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2127/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2127/comments
https://api.github.com/repos/huggingface/datasets/issues/2127/events
https://github.com/huggingface/datasets/pull/2127
843,017,199
MDExOlB1bGxSZXF1ZXN0NjAyNDYxMzc3
2,127
make documentation more clear to use different cloud storage
{ "login": "philschmid", "id": 32632186, "node_id": "MDQ6VXNlcjMyNjMyMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/32632186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/philschmid", "html_url": "https://github.com/philschmid", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,999,046,000
1,617,020,184,000
1,617,020,184,000
MEMBER
null
This PR extends the cloud storage documentation. To show you can use a different `fsspec` implementation.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2127/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2127/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2127", "html_url": "https://github.com/huggingface/datasets/pull/2127", "diff_url": "https://github.com/huggingface/datasets/pull/2127.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2127.patch", "merged_at": 1617020184000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2126
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2126/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2126/comments
https://api.github.com/repos/huggingface/datasets/issues/2126/events
https://github.com/huggingface/datasets/pull/2126
842,779,966
MDExOlB1bGxSZXF1ZXN0NjAyMjcyMjg4
2,126
Replace legacy torch.Tensor constructor with torch.tensor
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,950,650,000
1,617,010,034,000
1,617,010,033,000
CONTRIBUTOR
null
The title says it all (motivated by [this issue](https://github.com/pytorch/pytorch/issues/53146) in the pytorch repo).
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2126/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2126/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2126", "html_url": "https://github.com/huggingface/datasets/pull/2126", "diff_url": "https://github.com/huggingface/datasets/pull/2126.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2126.patch", "merged_at": 1617010033000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2125
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2125/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2125/comments
https://api.github.com/repos/huggingface/datasets/issues/2125/events
https://github.com/huggingface/datasets/issues/2125
842,690,570
MDU6SXNzdWU4NDI2OTA1NzA=
2,125
Is dataset timit_asr broken?
{ "login": "kosuke-kitahara", "id": 42398050, "node_id": "MDQ6VXNlcjQyMzk4MDUw", "avatar_url": "https://avatars.githubusercontent.com/u/42398050?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kosuke-kitahara", "html_url": "https://github.com/kosuke-kitahara", "followers_url": "https://api...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi,\r\n\r\nthanks for the report, but this is a duplicate of #2052. ", "@mariosasko \r\nThank you for your quick response! Following #2052, I've fixed the problem." ]
1,616,920,218,000
1,616,934,565,000
1,616,934,565,000
NONE
null
Using `timit_asr` dataset, I saw all records are the same. ``` python from datasets import load_dataset, load_metric timit = load_dataset("timit_asr") from datasets import ClassLabel import random import pandas as pd from IPython.display import display, HTML def show_random_elements(dataset, num_example...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2125/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2125/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2124
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2124/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2124/comments
https://api.github.com/repos/huggingface/datasets/issues/2124/events
https://github.com/huggingface/datasets/issues/2124
842,627,729
MDU6SXNzdWU4NDI2Mjc3Mjk=
2,124
Adding ScaNN library to do MIPS?
{ "login": "shamanez", "id": 16892570, "node_id": "MDQ6VXNlcjE2ODkyNTcw", "avatar_url": "https://avatars.githubusercontent.com/u/16892570?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shamanez", "html_url": "https://github.com/shamanez", "followers_url": "https://api.github.com/users/sha...
[]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "I haven't played with it (yet) but it sounds really cool !\r\n" ]
1,616,890,020,000
1,617,024,223,000
null
NONE
null
@lhoestq Hi I am thinking of adding this new google library to do the MIPS similar to **add_faiss_idex**. As the paper suggests, it is really fast when it comes to retrieving the nearest neighbors. https://github.com/google-research/google-research/tree/master/scann ![image](https://user-images.githubusercontent...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2124/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2124/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2123
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2123/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2123/comments
https://api.github.com/repos/huggingface/datasets/issues/2123/events
https://github.com/huggingface/datasets/issues/2123
842,577,285
MDU6SXNzdWU4NDI1NzcyODU=
2,123
Problem downloading GEM wiki_auto_asset_turk dataset
{ "login": "mille-s", "id": 29705940, "node_id": "MDQ6VXNlcjI5NzA1OTQw", "avatar_url": "https://avatars.githubusercontent.com/u/29705940?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mille-s", "html_url": "https://github.com/mille-s", "followers_url": "https://api.github.com/users/mille-...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi,\r\n\r\nsadly I can't replicate the problem on my Windows machine. Try to update the library to the newest version with:\r\n```bash\r\npip install git+https://github.com/huggingface/datasets\r\n``` ", "Thanks for the answer! I updated the library but unfortunately it didn't solve the problem.", "Is there an...
1,616,870,488,000
1,620,836,118,000
1,620,836,117,000
NONE
null
@yjernite ### Summary I am currently working on the GEM datasets and do not manage to download the wiki_auto_asset_turk data, whereas all other datasets download well with the same code. ### Steps to reproduce Code snippet: from datasets import load_dataset #dataset = load_dataset('gem', 'web_nlg_en') d...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2123/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2123/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2122
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2122/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2122/comments
https://api.github.com/repos/huggingface/datasets/issues/2122/events
https://github.com/huggingface/datasets/pull/2122
842,194,588
MDExOlB1bGxSZXF1ZXN0NjAxODE3MjI0
2,122
Fast table queries with interpolation search
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,782,160,000
1,628,100,719,000
1,617,719,581,000
MEMBER
null
## Intro This should fix issue #1803 Currently querying examples in a dataset is O(n) because of the underlying pyarrow ChunkedArrays implementation. To fix this I implemented interpolation search that is pretty effective since datasets usually verifies the condition of evenly distributed chunks (the default ch...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2122/reactions", "total_count": 4, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 4, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2122/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2122", "html_url": "https://github.com/huggingface/datasets/pull/2122", "diff_url": "https://github.com/huggingface/datasets/pull/2122.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2122.patch", "merged_at": 1617719581000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2121
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2121/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2121/comments
https://api.github.com/repos/huggingface/datasets/issues/2121/events
https://github.com/huggingface/datasets/pull/2121
842,148,633
MDExOlB1bGxSZXF1ZXN0NjAxNzc4NDc4
2,121
Add Validation For README
{ "login": "gchhablani", "id": 29076344, "node_id": "MDQ6VXNlcjI5MDc2MzQ0", "avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gchhablani", "html_url": "https://github.com/gchhablani", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Good start! Here are some proposed next steps:\r\n- We want the Class structure to reflect the template - so the parser know what section titles to expect and when something has gone wrong\r\n- As a result, we don't need to parse the table of contents, since it will always be the same\r\n- For each section/subsect...
1,616,778,137,000
1,620,652,638,000
1,620,639,701,000
CONTRIBUTOR
null
Hi @lhoestq, @yjernite This is a simple Readme parser. All classes specific to different sections can inherit `Section` class, and we can define more attributes in each. Let me know if this is going in the right direction :) Currently the output looks like this, for `to_dict()` on `FashionMNIST` `README.md`: ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2121/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2121/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2121", "html_url": "https://github.com/huggingface/datasets/pull/2121", "diff_url": "https://github.com/huggingface/datasets/pull/2121.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2121.patch", "merged_at": 1620639701000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2120
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2120/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2120/comments
https://api.github.com/repos/huggingface/datasets/issues/2120/events
https://github.com/huggingface/datasets/issues/2120
841,954,521
MDU6SXNzdWU4NDE5NTQ1MjE=
2,120
dataset viewer does not work anymore
{ "login": "dorost1234", "id": 79165106, "node_id": "MDQ6VXNlcjc5MTY1MTA2", "avatar_url": "https://avatars.githubusercontent.com/u/79165106?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dorost1234", "html_url": "https://github.com/dorost1234", "followers_url": "https://api.github.com/use...
[ { "id": 2107841032, "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer", "name": "nlp-viewer", "color": "94203D", "default": false, "description": "" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Thanks for reporting :) We're looking into it", "Back up. " ]
1,616,764,933,000
1,616,773,942,000
1,616,773,942,000
NONE
null
Hi I normally use this link to see all datasets and how I can load them https://huggingface.co/datasets/viewer/ Now I am getting 502 Bad Gateway nginx/1.18.0 (Ubuntu) could you bring this webpage back ? this was very helpful @lhoestq thanks for your help
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2120/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2120/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2119
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2119/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2119/comments
https://api.github.com/repos/huggingface/datasets/issues/2119/events
https://github.com/huggingface/datasets/pull/2119
841,567,199
MDExOlB1bGxSZXF1ZXN0NjAxMjg2MjIy
2,119
copy.deepcopy os.environ instead of copy
{ "login": "NihalHarish", "id": 5506053, "node_id": "MDQ6VXNlcjU1MDYwNTM=", "avatar_url": "https://avatars.githubusercontent.com/u/5506053?v=4", "gravatar_id": "", "url": "https://api.github.com/users/NihalHarish", "html_url": "https://github.com/NihalHarish", "followers_url": "https://api.github.com/us...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,731,118,000
1,616,771,632,000
1,616,771,632,000
CONTRIBUTOR
null
Fixes: https://github.com/huggingface/datasets/issues/2115 - bug fix: using envrion.copy() returns a dict. - using deepcopy(environ) returns an `_environ` object - Changing the datatype of the _environ object can break code, if subsequent libraries perform operations using apis exclusive to the environ object, lik...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2119/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2119/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2119", "html_url": "https://github.com/huggingface/datasets/pull/2119", "diff_url": "https://github.com/huggingface/datasets/pull/2119.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2119.patch", "merged_at": 1616771632000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2118
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2118/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2118/comments
https://api.github.com/repos/huggingface/datasets/issues/2118/events
https://github.com/huggingface/datasets/pull/2118
841,563,329
MDExOlB1bGxSZXF1ZXN0NjAxMjgzMDUx
2,118
Remove os.environ.copy in Dataset.map
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "I thought deepcopy on `os.environ` is unsafe (see [this](https://stackoverflow.com/questions/13142972/using-copy-deepcopy-on-os-environ-in-python-appears-broken)), but I can't replicate the behavior described in the linked SO thread.\r\n\r\nClosing this one because #2119 has a much cleaner approach." ]
1,616,730,497,000
1,616,760,203,000
1,616,760,005,000
CONTRIBUTOR
null
Replace `os.environ.copy` with in-place modification Fixes #2115
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2118/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2118/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2118", "html_url": "https://github.com/huggingface/datasets/pull/2118", "diff_url": "https://github.com/huggingface/datasets/pull/2118.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2118.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/2117
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2117/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2117/comments
https://api.github.com/repos/huggingface/datasets/issues/2117/events
https://github.com/huggingface/datasets/issues/2117
841,535,283
MDU6SXNzdWU4NDE1MzUyODM=
2,117
load_metric from local "glue.py" meet error 'NoneType' object is not callable
{ "login": "Frankie123421", "id": 54012361, "node_id": "MDQ6VXNlcjU0MDEyMzYx", "avatar_url": "https://avatars.githubusercontent.com/u/54012361?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Frankie123421", "html_url": "https://github.com/Frankie123421", "followers_url": "https://api.githu...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "@Frankie123421 what was the resolution to this?", "> @Frankie123421 what was the resolution to this?\r\n\r\nuse glue_metric.py instead of glue.py in load_metric", "thank you!" ]
1,616,726,122,000
1,629,927,845,000
1,616,726,426,000
NONE
null
actual_task = "mnli" if task == "mnli-mm" else task dataset = load_dataset(path='/home/glue.py', name=actual_task) metric = load_metric(path='/home/glue.py', name=actual_task) --------------------------------------------------------------------------- TypeError Traceback (most recent...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2117/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2117/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2116
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2116/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2116/comments
https://api.github.com/repos/huggingface/datasets/issues/2116/events
https://github.com/huggingface/datasets/issues/2116
841,481,292
MDU6SXNzdWU4NDE0ODEyOTI=
2,116
Creating custom dataset results in error while calling the map() function
{ "login": "GeetDsa", "id": 13940397, "node_id": "MDQ6VXNlcjEzOTQwMzk3", "avatar_url": "https://avatars.githubusercontent.com/u/13940397?v=4", "gravatar_id": "", "url": "https://api.github.com/users/GeetDsa", "html_url": "https://github.com/GeetDsa", "followers_url": "https://api.github.com/users/GeetDs...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi,\r\n\r\nthe `_data` attribute is missing due to `MyDataset.__init__` not calling the parent `__init__`. However, I don't think it's a good idea to subclass the `datasets.Dataset` class (e.g. it's kind of dangerous to override `datasets.Dataset.__getitem__`). Instead, it's better to follow the \"association over...
1,616,719,066,000
1,617,201,032,000
1,617,201,032,000
NONE
null
calling `map()` of `datasets` library results into an error while defining a Custom dataset. Reproducible example: ``` import datasets class MyDataset(datasets.Dataset): def __init__(self, sentences): "Initialization" self.samples = sentences def __len__(self): "Denotes the ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2116/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2116/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2115
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2115/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2115/comments
https://api.github.com/repos/huggingface/datasets/issues/2115/events
https://github.com/huggingface/datasets/issues/2115
841,283,974
MDU6SXNzdWU4NDEyODM5NzQ=
2,115
The datasets.map() implementation modifies the datatype of os.environ object
{ "login": "leleamol", "id": 19983848, "node_id": "MDQ6VXNlcjE5OTgzODQ4", "avatar_url": "https://avatars.githubusercontent.com/u/19983848?v=4", "gravatar_id": "", "url": "https://api.github.com/users/leleamol", "html_url": "https://github.com/leleamol", "followers_url": "https://api.github.com/users/lel...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,704,159,000
1,616,771,632,000
1,616,771,632,000
NONE
null
In our testing, we noticed that the datasets.map() implementation is modifying the datatype of python os.environ object from '_Environ' to 'dict'. This causes following function calls to fail as follows: ` x = os.environ.get("TEST_ENV_VARIABLE_AFTER_dataset_map", default=None) TypeError: get() takes...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2115/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2115/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2114
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2114/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2114/comments
https://api.github.com/repos/huggingface/datasets/issues/2114/events
https://github.com/huggingface/datasets/pull/2114
841,207,878
MDExOlB1bGxSZXF1ZXN0NjAwOTc1MTA3
2,114
Support for legal NLP datasets (EURLEX, ECtHR cases and EU-REG-IR)
{ "login": "iliaschalkidis", "id": 1626984, "node_id": "MDQ6VXNlcjE2MjY5ODQ=", "avatar_url": "https://avatars.githubusercontent.com/u/1626984?v=4", "gravatar_id": "", "url": "https://api.github.com/users/iliaschalkidis", "html_url": "https://github.com/iliaschalkidis", "followers_url": "https://api.gith...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "> Awesome thank you :)\r\n> This is really cool\r\n> \r\n> I left a few comments.\r\n> \r\n> Also it looks like the dummy data are quite big (100-200KB each). Can you try to reduce their sizes please ? For example I noticed that all the jsonl files inside the `dummy_data.zip` files have 20 lines. Can you only keep...
1,616,697,617,000
1,617,187,130,000
1,617,187,130,000
CONTRIBUTOR
null
Add support for two legal NLP datasets: - EURLEX (https://www.aclweb.org/anthology/P19-1636/) - ECtHR cases (https://arxiv.org/abs/2103.13084) - EU-REG-IR (https://arxiv.org/abs/2101.10726)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2114/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2114/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2114", "html_url": "https://github.com/huggingface/datasets/pull/2114", "diff_url": "https://github.com/huggingface/datasets/pull/2114.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2114.patch", "merged_at": 1617187130000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2113
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2113/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2113/comments
https://api.github.com/repos/huggingface/datasets/issues/2113/events
https://github.com/huggingface/datasets/pull/2113
841,191,303
MDExOlB1bGxSZXF1ZXN0NjAwOTYxMDEz
2,113
Implement Dataset as context manager
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,696,310,000
1,617,190,214,000
1,617,179,411,000
MEMBER
null
When used as context manager, it would be safely deleted if some exception is raised. This will avoid > During handling of the above exception, another exception occurred:
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2113/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2113/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2113", "html_url": "https://github.com/huggingface/datasets/pull/2113", "diff_url": "https://github.com/huggingface/datasets/pull/2113.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2113.patch", "merged_at": 1617179411000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2112
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2112/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2112/comments
https://api.github.com/repos/huggingface/datasets/issues/2112/events
https://github.com/huggingface/datasets/pull/2112
841,098,008
MDExOlB1bGxSZXF1ZXN0NjAwODgyMjA0
2,112
Support for legal NLP datasets (EURLEX and ECtHR cases)
{ "login": "iliaschalkidis", "id": 1626984, "node_id": "MDQ6VXNlcjE2MjY5ODQ=", "avatar_url": "https://avatars.githubusercontent.com/u/1626984?v=4", "gravatar_id": "", "url": "https://api.github.com/users/iliaschalkidis", "html_url": "https://github.com/iliaschalkidis", "followers_url": "https://api.gith...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,689,457,000
1,616,697,571,000
1,616,697,271,000
CONTRIBUTOR
null
Add support for two legal NLP datasets: - EURLEX (https://www.aclweb.org/anthology/P19-1636/) - ECtHR cases (https://arxiv.org/abs/2103.13084)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2112/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2112/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2112", "html_url": "https://github.com/huggingface/datasets/pull/2112", "diff_url": "https://github.com/huggingface/datasets/pull/2112.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2112.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/2111
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2111/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2111/comments
https://api.github.com/repos/huggingface/datasets/issues/2111/events
https://github.com/huggingface/datasets/pull/2111
841,082,087
MDExOlB1bGxSZXF1ZXN0NjAwODY4OTg5
2,111
Compute WER metric iteratively
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "I discussed with Patrick and I think we could have a nice addition: have a parameter `concatenate_texts` that, if `True`, uses the old implementation.\r\n\r\nBy default `concatenate_texts` would be `False`, so that sentences are evaluated independently, and to save resources (the WER computation has a quadratic co...
1,616,688,408,000
1,617,693,643,000
1,617,693,643,000
MEMBER
null
Compute WER metric iteratively to avoid MemoryError. Fix #2078.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2111/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2111/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2111", "html_url": "https://github.com/huggingface/datasets/pull/2111", "diff_url": "https://github.com/huggingface/datasets/pull/2111.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2111.patch", "merged_at": 1617693643000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2110
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2110/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2110/comments
https://api.github.com/repos/huggingface/datasets/issues/2110/events
https://github.com/huggingface/datasets/pull/2110
840,794,995
MDExOlB1bGxSZXF1ZXN0NjAwNjI1NDQ5
2,110
Fix incorrect assertion in builder.py
{ "login": "dreamgonfly", "id": 2340721, "node_id": "MDQ6VXNlcjIzNDA3MjE=", "avatar_url": "https://avatars.githubusercontent.com/u/2340721?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dreamgonfly", "html_url": "https://github.com/dreamgonfly", "followers_url": "https://api.github.com/us...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! The SplitInfo is not always available. By default you would get `split_info.num_examples == 0`\r\nSo unfortunately we can't use this assertion you suggested", "> Hi ! The SplitInfo is not always available. By default you would get `split_info.num_examples == 0`\r\n> So unfortunately we can't use this assert...
1,616,668,760,000
1,618,234,383,000
1,618,234,383,000
CONTRIBUTOR
null
Fix incorrect num_examples comparison assertion in builder.py
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2110/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2110/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2110", "html_url": "https://github.com/huggingface/datasets/pull/2110", "diff_url": "https://github.com/huggingface/datasets/pull/2110.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2110.patch", "merged_at": 1618234383000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2109
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2109/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2109/comments
https://api.github.com/repos/huggingface/datasets/issues/2109/events
https://github.com/huggingface/datasets/pull/2109
840,746,598
MDExOlB1bGxSZXF1ZXN0NjAwNTg1MzM5
2,109
Add more issue templates and customize issue template chooser
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "If you agree, I could also add a link to [Discussions](https://github.com/huggingface/datasets/discussions) in order to reinforce the use of Discussion to make Questions (instead of Issues).\r\n\r\nI could also add some other templates: Bug, Feature Request,...", "@theo-m we wrote our same comments at the same t...
1,616,665,313,000
1,618,813,211,000
1,618,813,211,000
MEMBER
null
When opening an issue, it is not evident for the users how to choose a blank issue template. There is a link at the bottom of all the other issue templates (`Don’t see your issue here? Open a blank issue.`), but this is not very visible for users. This is the reason why many users finally chose the `add-dataset` templa...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2109/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2109/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2109", "html_url": "https://github.com/huggingface/datasets/pull/2109", "diff_url": "https://github.com/huggingface/datasets/pull/2109.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2109.patch", "merged_at": 1618813210000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2108
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2108/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2108/comments
https://api.github.com/repos/huggingface/datasets/issues/2108/events
https://github.com/huggingface/datasets/issues/2108
840,181,055
MDU6SXNzdWU4NDAxODEwNTU=
2,108
Is there a way to use a GPU only when training an Index in the process of add_faisis_index?
{ "login": "shamanez", "id": 16892570, "node_id": "MDQ6VXNlcjE2ODkyNTcw", "avatar_url": "https://avatars.githubusercontent.com/u/16892570?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shamanez", "html_url": "https://github.com/shamanez", "followers_url": "https://api.github.com/users/sha...
[ { "id": 1935892912, "node_id": "MDU6TGFiZWwxOTM1ODkyOTEy", "url": "https://api.github.com/repos/huggingface/datasets/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "Further information is requested" } ]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,621,536,000
1,616,653,903,000
null
NONE
null
Motivation - Some FAISS indexes like IVF consist of the training step that clusters the dataset into a given number of indexes. It would be nice if we can use a GPU to do the training step and covert the index back to CPU as mention in [this faiss example](https://gist.github.com/mdouze/46d6bbbaabca0b9778fca37ed2bcccf6...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2108/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2108/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2107
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2107/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2107/comments
https://api.github.com/repos/huggingface/datasets/issues/2107/events
https://github.com/huggingface/datasets/pull/2107
839,495,825
MDExOlB1bGxSZXF1ZXN0NTk5NTAxODE5
2,107
Metadata validation
{ "login": "theo-m", "id": 17948980, "node_id": "MDQ6VXNlcjE3OTQ4OTgw", "avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theo-m", "html_url": "https://github.com/theo-m", "followers_url": "https://api.github.com/users/theo-m/fo...
[]
closed
false
{ "login": "SBrandeis", "id": 33657802, "node_id": "MDQ6VXNlcjMzNjU3ODAy", "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SBrandeis", "html_url": "https://github.com/SBrandeis", "followers_url": "https://api.github.com/users/...
[ { "login": "SBrandeis", "id": 33657802, "node_id": "MDQ6VXNlcjMzNjU3ODAy", "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SBrandeis", "html_url": "https://github.com/SBrandeis", "followers_url": "https://a...
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "> Also I was wondering this is really needed to have `utils.metadata` as a submodule of `datasets` ? This is only used by the CI so I'm not sure we should have this in the actual `datasets` package.\r\n\r\nI'm unclear on the suggestion, would you rather have a root-level `./metadata.py` file? I think it's well whe...
1,616,575,961,000
1,619,425,634,000
1,619,425,633,000
CONTRIBUTOR
null
- `pydantic` metadata schema with dedicated validators against our taxonomy - ci script to validate new changes against this schema and start a vertuous loop - soft validation on tasks ids since we expect the taxonomy to undergo some changes in the near future for reference with the current validation we have ~365...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2107/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2107/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2107", "html_url": "https://github.com/huggingface/datasets/pull/2107", "diff_url": "https://github.com/huggingface/datasets/pull/2107.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2107.patch", "merged_at": 1619425633000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2106
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2106/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2106/comments
https://api.github.com/repos/huggingface/datasets/issues/2106/events
https://github.com/huggingface/datasets/issues/2106
839,084,264
MDU6SXNzdWU4MzkwODQyNjQ=
2,106
WMT19 Dataset for Kazakh-English is not formatted correctly
{ "login": "trina731", "id": 22580542, "node_id": "MDQ6VXNlcjIyNTgwNTQy", "avatar_url": "https://avatars.githubusercontent.com/u/22580542?v=4", "gravatar_id": "", "url": "https://api.github.com/users/trina731", "html_url": "https://github.com/trina731", "followers_url": "https://api.github.com/users/tri...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! Thanks for reporting\r\n\r\nBy looking at the raw `news-commentary-v14.en-kk.tsv` file, it looks like there are at least 17 lines with this issue.\r\nMoreover these issues are not always the same:\r\n- L97 is only `kk` text and must be appended at the end of the `kk` text of the **next** line\r\n- L2897 is on...
1,616,530,487,000
1,616,708,180,000
null
NONE
null
In addition to the bug of languages being switched from Issue @415, there are incorrect translations in the dataset because the English-Kazakh translations have a one off formatting error. The News Commentary v14 parallel data set for kk-en from http://www.statmt.org/wmt19/translation-task.html has a bug here: > ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2106/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2106/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2105
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2105/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2105/comments
https://api.github.com/repos/huggingface/datasets/issues/2105/events
https://github.com/huggingface/datasets/issues/2105
839,059,226
MDU6SXNzdWU4MzkwNTkyMjY=
2,105
Request to remove S2ORC dataset
{ "login": "kyleclo", "id": 13603748, "node_id": "MDQ6VXNlcjEzNjAzNzQ4", "avatar_url": "https://avatars.githubusercontent.com/u/13603748?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kyleclo", "html_url": "https://github.com/kyleclo", "followers_url": "https://api.github.com/users/kylecl...
[]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hello @kyleclo! Currently, we are getting the data from your bucket, so if you remove it the HF script won't work anymore :) \r\n\r\nUntil you solve things on your end, @lhoestq suggested we just return a warning message when people try to load that dataset from HF. What would you like it to say?", "Hi @kyleclo,...
1,616,528,586,000
1,628,104,682,000
null
NONE
null
Hi! I was wondering if it's possible to remove [S2ORC](https://huggingface.co/datasets/s2orc) from hosting on Huggingface's platform? Unfortunately, there are some legal considerations about how we make this data available. Happy to add back to Huggingface's platform once we work out those hurdles! Thanks!
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2105/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/huggingface/datasets/issues/2105/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2104
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2104/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2104/comments
https://api.github.com/repos/huggingface/datasets/issues/2104/events
https://github.com/huggingface/datasets/issues/2104
839,027,834
MDU6SXNzdWU4MzkwMjc4MzQ=
2,104
Trouble loading wiki_movies
{ "login": "adityaarunsinghal", "id": 35391599, "node_id": "MDQ6VXNlcjM1MzkxNTk5", "avatar_url": "https://avatars.githubusercontent.com/u/35391599?v=4", "gravatar_id": "", "url": "https://api.github.com/users/adityaarunsinghal", "html_url": "https://github.com/adityaarunsinghal", "followers_url": "https...
[]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! `wiki_movies` was added in `datasets==1.2.0`. However it looks like you have `datasets==1.1.2`.\r\n\r\nTo use `wiki_movies`, please update `datasets` with\r\n```\r\npip install --upgrade datasets\r\n```", "Thanks a lot! That solved it and I was able to upload a model trained on it as well :)" ]
1,616,525,994,000
1,617,664,646,000
null
NONE
null
Hello, I am trying to load_dataset("wiki_movies") and it gives me this error - `FileNotFoundError: Couldn't find file locally at wiki_movies/wiki_movies.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.1.2/datasets/wiki_movies/wiki_movies.py or https://s3.amazonaws.com/datasets.huggingfa...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2104/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2104/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2103
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2103/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2103/comments
https://api.github.com/repos/huggingface/datasets/issues/2103/events
https://github.com/huggingface/datasets/issues/2103
838,946,916
MDU6SXNzdWU4Mzg5NDY5MTY=
2,103
citation, homepage, and license fields of `dataset_info.json` are duplicated many times
{ "login": "samsontmr", "id": 15007950, "node_id": "MDQ6VXNlcjE1MDA3OTUw", "avatar_url": "https://avatars.githubusercontent.com/u/15007950?v=4", "gravatar_id": "", "url": "https://api.github.com/users/samsontmr", "html_url": "https://github.com/samsontmr", "followers_url": "https://api.github.com/users/...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 1935892877, "node_id": "MDU6...
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Thanks for reporting :)\r\nMaybe we can concatenate fields only if they are different.\r\n\r\nCurrently this is done here:\r\n\r\nhttps://github.com/huggingface/nlp/blob/349ac4398a3bcae6356f14c5754483383a60e8a4/src/datasets/info.py#L180-L196\r\n\r\nThis can be a good first contribution to the library.\r\nPlease co...
1,616,519,889,000
1,617,719,999,000
1,617,719,999,000
NONE
null
This happens after a `map` operation when `num_proc` is set to `>1`. I tested this by cleaning up the json before running the `map` op on the dataset so it's unlikely it's coming from an earlier concatenation. Example result: ``` "citation": "@ONLINE {wikidump,\n author = {Wikimedia Foundation},\n title = {...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2103/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2103/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2102
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2102/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2102/comments
https://api.github.com/repos/huggingface/datasets/issues/2102/events
https://github.com/huggingface/datasets/pull/2102
838,794,090
MDExOlB1bGxSZXF1ZXN0NTk4OTEyNzUw
2,102
Move Dataset.to_csv to csv module
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 2851292821, "node_id": "MDU6TGFiZWwyODUxMjkyODIx", "url": "https://api.github.com/repos/huggingface/datasets/labels/refactoring", "name": "refactoring", "color": "B67A40", "default": false, "description": "Restructuring existing code without changing its external behavior" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,510,146,000
1,616,594,855,000
1,616,594,854,000
MEMBER
null
Move the implementation of `Dataset.to_csv` to module `datasets.io.csv`.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2102/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2102/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2102", "html_url": "https://github.com/huggingface/datasets/pull/2102", "diff_url": "https://github.com/huggingface/datasets/pull/2102.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2102.patch", "merged_at": 1616594854000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2101
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2101/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2101/comments
https://api.github.com/repos/huggingface/datasets/issues/2101/events
https://github.com/huggingface/datasets/pull/2101
838,586,184
MDExOlB1bGxSZXF1ZXN0NTk4NzQzMDM4
2,101
MIAM dataset - new citation details
{ "login": "eusip", "id": 1551356, "node_id": "MDQ6VXNlcjE1NTEzNTY=", "avatar_url": "https://avatars.githubusercontent.com/u/1551356?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eusip", "html_url": "https://github.com/eusip", "followers_url": "https://api.github.com/users/eusip/follower...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi !\r\nLooks like there's a unicode error in the new citation in the miam.py file.\r\nCould you try to fix it ? Not sure from which character it comes from though\r\n\r\nYou can test if it works on your side with\r\n```\r\nRUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_con...
1,616,496,083,000
1,616,522,890,000
1,616,522,890,000
CONTRIBUTOR
null
Hi @lhoestq, I have updated the citations to reference an OpenReview preprint.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2101/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2101/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2101", "html_url": "https://github.com/huggingface/datasets/pull/2101", "diff_url": "https://github.com/huggingface/datasets/pull/2101.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2101.patch", "merged_at": 1616522889000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2100
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2100/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2100/comments
https://api.github.com/repos/huggingface/datasets/issues/2100/events
https://github.com/huggingface/datasets/pull/2100
838,574,631
MDExOlB1bGxSZXF1ZXN0NTk4NzMzOTM0
2,100
Fix deprecated warning message and docstring
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "I have a question: what about `dictionary_encode_column_`?\r\n- It is deprecated in Dataset, but it recommends using a non-existing method instead: `Dataset.dictionary_encode_column` does not exist.\r\n- It is NOT deprecated in DatasetDict.", "`dictionary_encode_column_ ` should be deprecated since it never work...
1,616,495,272,000
1,616,573,981,000
1,616,522,629,000
MEMBER
null
Fix deprecated warnings: - Use deprecated Sphinx directive in docstring - Fix format of deprecated message - Raise FutureWarning
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2100/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2100/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2100", "html_url": "https://github.com/huggingface/datasets/pull/2100", "diff_url": "https://github.com/huggingface/datasets/pull/2100.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2100.patch", "merged_at": 1616522629000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2099
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2099/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2099/comments
https://api.github.com/repos/huggingface/datasets/issues/2099/events
https://github.com/huggingface/datasets/issues/2099
838,523,819
MDU6SXNzdWU4Mzg1MjM4MTk=
2,099
load_from_disk takes a long time to load local dataset
{ "login": "samsontmr", "id": 15007950, "node_id": "MDQ6VXNlcjE1MDA3OTUw", "avatar_url": "https://avatars.githubusercontent.com/u/15007950?v=4", "gravatar_id": "", "url": "https://api.github.com/users/samsontmr", "html_url": "https://github.com/samsontmr", "followers_url": "https://api.github.com/users/...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi !\r\nCan you share more information about the features of your dataset ? You can get them by printing `my_dataset.features`\r\nCan you also share the code of your `map` function ?", "It is actually just the tokenized `wikipedia` dataset with `input_ids`, `attention_mask`, etc, with one extra column which is a...
1,616,491,717,000
1,616,519,536,000
1,616,519,536,000
NONE
null
I have an extremely large tokenized dataset (24M examples) that loads in a few minutes. However, after adding a column similar to `input_ids` (basically a list of integers) and saving the dataset to disk, the load time goes to >1 hour. I've even tried using `np.uint8` after seeing #1985 but it doesn't seem to be helpin...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2099/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2099/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2098
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2098/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2098/comments
https://api.github.com/repos/huggingface/datasets/issues/2098/events
https://github.com/huggingface/datasets/issues/2098
838,447,959
MDU6SXNzdWU4Mzg0NDc5NTk=
2,098
SQuAD version
{ "login": "h-peng17", "id": 39556019, "node_id": "MDQ6VXNlcjM5NTU2MDE5", "avatar_url": "https://avatars.githubusercontent.com/u/39556019?v=4", "gravatar_id": "", "url": "https://api.github.com/users/h-peng17", "html_url": "https://github.com/h-peng17", "followers_url": "https://api.github.com/users/h-p...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! This is 1.1 as specified by the download urls here:\r\n\r\nhttps://github.com/huggingface/nlp/blob/349ac4398a3bcae6356f14c5754483383a60e8a4/datasets/squad/squad.py#L50-L55", "Got it. Thank you~" ]
1,616,485,674,000
1,616,752,134,000
1,616,752,134,000
NONE
null
Hi~ I want train on squad dataset. What's the version of the squad? Is it 1.1 or 1.0? I'm new in QA, I don't find some descriptions about it.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2098/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2098/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2097
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2097/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2097/comments
https://api.github.com/repos/huggingface/datasets/issues/2097/events
https://github.com/huggingface/datasets/pull/2097
838,105,289
MDExOlB1bGxSZXF1ZXN0NTk4MzM4MTA3
2,097
fixes issue #1110 by descending further if `obj["_type"]` is a dict
{ "login": "dcfidalgo", "id": 15979778, "node_id": "MDQ6VXNlcjE1OTc5Nzc4", "avatar_url": "https://avatars.githubusercontent.com/u/15979778?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dcfidalgo", "html_url": "https://github.com/dcfidalgo", "followers_url": "https://api.github.com/users/...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,446,855,000
1,616,446,871,000
1,616,446,871,000
CONTRIBUTOR
null
Check metrics
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2097/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2097/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2097", "html_url": "https://github.com/huggingface/datasets/pull/2097", "diff_url": "https://github.com/huggingface/datasets/pull/2097.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2097.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/2096
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2096/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2096/comments
https://api.github.com/repos/huggingface/datasets/issues/2096/events
https://github.com/huggingface/datasets/issues/2096
838,038,379
MDU6SXNzdWU4MzgwMzgzNzk=
2,096
CoNLL 2003 dataset not including German
{ "login": "rxian", "id": 8406802, "node_id": "MDQ6VXNlcjg0MDY4MDI=", "avatar_url": "https://avatars.githubusercontent.com/u/8406802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rxian", "html_url": "https://github.com/rxian", "followers_url": "https://api.github.com/users/rxian/follower...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,441,036,000
1,617,097,535,000
null
NONE
null
Hello, thanks for all the work on developing and maintaining this amazing platform, which I am enjoying working with! I was wondering if there is a reason why the German CoNLL 2003 dataset is not included in the [repository](https://github.com/huggingface/datasets/tree/master/datasets/conll2003), since a copy of it ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2096/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2096/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2093
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2093/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2093/comments
https://api.github.com/repos/huggingface/datasets/issues/2093/events
https://github.com/huggingface/datasets/pull/2093
837,209,211
MDExOlB1bGxSZXF1ZXN0NTk3NTgyNjUx
2,093
Fix: Allows a feature to be named "_type"
{ "login": "dcfidalgo", "id": 15979778, "node_id": "MDQ6VXNlcjE1OTc5Nzc4", "avatar_url": "https://avatars.githubusercontent.com/u/15979778?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dcfidalgo", "html_url": "https://github.com/dcfidalgo", "followers_url": "https://api.github.com/users/...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Nice thank you !\r\nThis looks like a pretty simple yet effective fix ;)\r\nCould you just add a test in `test_features.py` to make sure that you can create `features` with a `_type` field and that it is possible to convert it as a dict and reload it ?\r\n```python\r\nfrom datasets import Features, Value\r\n\r\n# ...
1,616,368,917,000
1,616,682,954,000
1,616,682,954,000
CONTRIBUTOR
null
This PR tries to fix issue #1110. Sorry for taking so long to come back to this. It's a simple fix, but i am not sure if it works for all possible types of `obj`. Let me know what you think @lhoestq
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2093/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2093/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2093", "html_url": "https://github.com/huggingface/datasets/pull/2093", "diff_url": "https://github.com/huggingface/datasets/pull/2093.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2093.patch", "merged_at": 1616682954000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2092
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2092/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2092/comments
https://api.github.com/repos/huggingface/datasets/issues/2092/events
https://github.com/huggingface/datasets/issues/2092
836,984,043
MDU6SXNzdWU4MzY5ODQwNDM=
2,092
How to disable making arrow tables in load_dataset ?
{ "login": "Jeevesh8", "id": 48825663, "node_id": "MDQ6VXNlcjQ4ODI1NjYz", "avatar_url": "https://avatars.githubusercontent.com/u/48825663?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Jeevesh8", "html_url": "https://github.com/Jeevesh8", "followers_url": "https://api.github.com/users/Jee...
[]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! We plan to add streaming features in the future.\r\n\r\nThis should allow to load a dataset instantaneously without generating the arrow table. The trade-off is that accessing examples from a streaming dataset must be done in an iterative way, and with an additional (but hopefully minor) overhead.\r\nWhat do ...
1,616,302,207,000
1,616,783,860,000
null
NONE
null
Is there a way to disable the construction of arrow tables, or to make them on the fly as the dataset is being used ?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2092/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2092/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2091
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2091/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2091/comments
https://api.github.com/repos/huggingface/datasets/issues/2091/events
https://github.com/huggingface/datasets/pull/2091
836,831,403
MDExOlB1bGxSZXF1ZXN0NTk3Mjk4ODI3
2,091
Fix copy snippet in docs
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[ { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,252,902,000
1,616,574,050,000
1,616,519,911,000
CONTRIBUTOR
null
With this change the lines starting with `...` in the code blocks can be properly copied to clipboard.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2091/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2091/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2091", "html_url": "https://github.com/huggingface/datasets/pull/2091", "diff_url": "https://github.com/huggingface/datasets/pull/2091.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2091.patch", "merged_at": 1616519911000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2090
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2090/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2090/comments
https://api.github.com/repos/huggingface/datasets/issues/2090/events
https://github.com/huggingface/datasets/pull/2090
836,807,498
MDExOlB1bGxSZXF1ZXN0NTk3MjgwNTEy
2,090
Add machine translated multilingual STS benchmark dataset
{ "login": "PhilipMay", "id": 229382, "node_id": "MDQ6VXNlcjIyOTM4Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/229382?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PhilipMay", "html_url": "https://github.com/PhilipMay", "followers_url": "https://api.github.com/users/Phil...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hello dear maintainer, are there any comments or questions about this PR?", "@iamollas thanks for the feedback. I did not see the template.\r\nI improved it...", "Should be clean for merge IMO.", "@lhoestq CI is green. ;-)", "Thanks again ! this is awesome :)", "Thanks for merging. :-)" ]
1,616,246,887,000
1,617,024,282,000
1,617,022,815,000
CONTRIBUTOR
null
also see here https://github.com/PhilipMay/stsb-multi-mt
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2090/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2090/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2090", "html_url": "https://github.com/huggingface/datasets/pull/2090", "diff_url": "https://github.com/huggingface/datasets/pull/2090.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2090.patch", "merged_at": 1617022815000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2089
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2089/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2089/comments
https://api.github.com/repos/huggingface/datasets/issues/2089/events
https://github.com/huggingface/datasets/issues/2089
836,788,019
MDU6SXNzdWU4MzY3ODgwMTk=
2,089
Add documentaton for dataset README.md files
{ "login": "PhilipMay", "id": 229382, "node_id": "MDQ6VXNlcjIyOTM4Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/229382?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PhilipMay", "html_url": "https://github.com/PhilipMay", "followers_url": "https://api.github.com/users/Phil...
[]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! We are using the [datasets-tagging app](https://github.com/huggingface/datasets-tagging) to select the tags to add.\r\n\r\nWe are also adding the full list of tags in #2107 \r\nThis covers multilinguality, language_creators, licenses, size_categories and task_categories.\r\n\r\nIn general if you want to add a...
1,616,240,678,000
1,626,111,700,000
null
CONTRIBUTOR
null
Hi, the dataset README files have special headers. Somehow a documenation of the allowed values and tags is missing. Could you add that? Just to give some concrete questions that should be answered imo: - which values can be passted to multilinguality? - what should be passed to language_creators? - which valu...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2089/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2089/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2088
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2088/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2088/comments
https://api.github.com/repos/huggingface/datasets/issues/2088/events
https://github.com/huggingface/datasets/pull/2088
836,763,733
MDExOlB1bGxSZXF1ZXN0NTk3MjQ4Mzk1
2,088
change bibtex template to author instead of authors
{ "login": "PhilipMay", "id": 229382, "node_id": "MDQ6VXNlcjIyOTM4Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/229382?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PhilipMay", "html_url": "https://github.com/PhilipMay", "followers_url": "https://api.github.com/users/Phil...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Trailing whitespace was removed. So more changes in diff than just this fix." ]
1,616,232,224,000
1,616,514,012,000
1,616,514,012,000
CONTRIBUTOR
null
Hi, IMO when using BibTex Author should be used instead of Authors. See here: http://www.bibtex.org/Using/de/ Thanks Philip
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2088/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2088/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2088", "html_url": "https://github.com/huggingface/datasets/pull/2088", "diff_url": "https://github.com/huggingface/datasets/pull/2088.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2088.patch", "merged_at": 1616514012000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2087
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2087/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2087/comments
https://api.github.com/repos/huggingface/datasets/issues/2087/events
https://github.com/huggingface/datasets/pull/2087
836,587,392
MDExOlB1bGxSZXF1ZXN0NTk3MDg4NTk2
2,087
Update metadata if dataset features are modified
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "@lhoestq I'll try to add a test later if you think this approach with the wrapper is good.", "Awesome thank you !\r\nYes this approach with a wrapper is good :)", "@lhoestq Added a test. To verify that this change fixes the problem, replace:\r\n```\r\n!pip install datasets==1.5\r\n```\r\nwith:\r\n```\r\n!pip i...
1,616,205,923,000
1,617,960,333,000
1,617,960,333,000
CONTRIBUTOR
null
This PR adds a decorator that updates the dataset metadata if a previously executed transform modifies its features. Fixes #2083
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2087/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 1, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2087/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2087", "html_url": "https://github.com/huggingface/datasets/pull/2087", "diff_url": "https://github.com/huggingface/datasets/pull/2087.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2087.patch", "merged_at": 1617960333000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2086
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2086/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2086/comments
https://api.github.com/repos/huggingface/datasets/issues/2086/events
https://github.com/huggingface/datasets/pull/2086
836,249,587
MDExOlB1bGxSZXF1ZXN0NTk2Nzg0Mjcz
2,086
change user permissions to -rw-r--r--
{ "login": "bhavitvyamalik", "id": 19718818, "node_id": "MDQ6VXNlcjE5NzE4ODE4", "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bhavitvyamalik", "html_url": "https://github.com/bhavitvyamalik", "followers_url": "https://api.gi...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "I tried this with `ade_corpus_v2` dataset. `ade_corpus_v2-train.arrow` (downloaded dataset) and `cache-25d41a4d3c2d8a25.arrow` (ran a mapping function on the dataset) both had file permission with octal value of `0644`. " ]
1,616,177,696,000
1,616,594,344,000
1,616,594,344,000
CONTRIBUTOR
null
Fix for #2065
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2086/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2086/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2086", "html_url": "https://github.com/huggingface/datasets/pull/2086", "diff_url": "https://github.com/huggingface/datasets/pull/2086.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2086.patch", "merged_at": 1616594344000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2085
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2085/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2085/comments
https://api.github.com/repos/huggingface/datasets/issues/2085/events
https://github.com/huggingface/datasets/pull/2085
835,870,994
MDExOlB1bGxSZXF1ZXN0NTk2NDYyOTc2
2,085
Fix max_wait_time in requests
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,152,946,000
1,616,513,798,000
1,616,513,797,000
MEMBER
null
it was handled as a min time, not max cc @SBrandeis
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2085/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2085/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2085", "html_url": "https://github.com/huggingface/datasets/pull/2085", "diff_url": "https://github.com/huggingface/datasets/pull/2085.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2085.patch", "merged_at": 1616513797000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2084
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2084/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2084/comments
https://api.github.com/repos/huggingface/datasets/issues/2084/events
https://github.com/huggingface/datasets/issues/2084
835,750,671
MDU6SXNzdWU4MzU3NTA2NzE=
2,084
CUAD - Contract Understanding Atticus Dataset
{ "login": "theo-m", "id": 17948980, "node_id": "MDQ6VXNlcjE3OTQ4OTgw", "avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theo-m", "html_url": "https://github.com/theo-m", "followers_url": "https://api.github.com/users/theo-m/fo...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "+1 on this request" ]
1,616,146,063,000
1,618,563,044,000
1,618,563,044,000
CONTRIBUTOR
null
## Adding a Dataset - **Name:** CUAD - Contract Understanding Atticus Dataset - **Description:** As one of the only large, specialized NLP benchmarks annotated by experts, CUAD can serve as a challenging research benchmark for the broader NLP community. - **Paper:** https://arxiv.org/abs/2103.06268 - **Data:** http...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2084/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2084/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2083
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2083/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2083/comments
https://api.github.com/repos/huggingface/datasets/issues/2083/events
https://github.com/huggingface/datasets/issues/2083
835,695,425
MDU6SXNzdWU4MzU2OTU0MjU=
2,083
`concatenate_datasets` throws error when changing the order of datasets to concatenate
{ "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "followers_url": "https://...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi,\r\n\r\nthis bug is related to `Dataset.{remove_columns, rename_column, flatten}` not propagating the change to the schema metadata when the info features are updated, so this line is the culprit:\r\n```python\r\ncommon_voice_train = common_voice_train.remove_columns(['client_id', 'up_votes', 'down_votes', 'age...
1,616,142,588,000
1,617,960,333,000
1,617,960,333,000
MEMBER
null
Hey, I played around with the `concatenate_datasets(...)` function: https://huggingface.co/docs/datasets/package_reference/main_classes.html?highlight=concatenate_datasets#datasets.concatenate_datasets and noticed that when the order in which the datasets are concatenated changes an error is thrown where it shou...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2083/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2083/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2082
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2082/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2082/comments
https://api.github.com/repos/huggingface/datasets/issues/2082/events
https://github.com/huggingface/datasets/pull/2082
835,401,555
MDExOlB1bGxSZXF1ZXN0NTk2MDY1NTM0
2,082
Updated card using information from data statement and datasheet
{ "login": "mcmillanmajora", "id": 26722925, "node_id": "MDQ6VXNlcjI2NzIyOTI1", "avatar_url": "https://avatars.githubusercontent.com/u/26722925?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mcmillanmajora", "html_url": "https://github.com/mcmillanmajora", "followers_url": "https://api.gi...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,114,378,000
1,616,164,149,000
1,616,164,149,000
CONTRIBUTOR
null
I updated and clarified the REFreSD [data card](https://github.com/mcmillanmajora/datasets/blob/refresd_card/datasets/refresd/README.md) with information from the Eleftheria's [website](https://elbria.github.io/post/refresd/). I added brief descriptions where the initial card referred to the paper, and I also recreated...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2082/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2082/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2082", "html_url": "https://github.com/huggingface/datasets/pull/2082", "diff_url": "https://github.com/huggingface/datasets/pull/2082.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2082.patch", "merged_at": 1616164148000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2081
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2081/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2081/comments
https://api.github.com/repos/huggingface/datasets/issues/2081/events
https://github.com/huggingface/datasets/pull/2081
835,112,968
MDExOlB1bGxSZXF1ZXN0NTk1ODE3OTM4
2,081
Fix docstrings issues
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,091,061,000
1,617,806,263,000
1,617,806,263,000
MEMBER
null
Fix docstring issues.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2081/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2081/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2081", "html_url": "https://github.com/huggingface/datasets/pull/2081", "diff_url": "https://github.com/huggingface/datasets/pull/2081.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2081.patch", "merged_at": 1617806263000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2080
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2080/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2080/comments
https://api.github.com/repos/huggingface/datasets/issues/2080/events
https://github.com/huggingface/datasets/issues/2080
835,023,000
MDU6SXNzdWU4MzUwMjMwMDA=
2,080
Multidimensional arrays in a Dataset
{ "login": "vermouthmjl", "id": 3142085, "node_id": "MDQ6VXNlcjMxNDIwODU=", "avatar_url": "https://avatars.githubusercontent.com/u/3142085?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vermouthmjl", "html_url": "https://github.com/vermouthmjl", "followers_url": "https://api.github.com/us...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi !\r\n\r\nThis is actually supported ! but not yet in `from_pandas`.\r\nYou can use `from_dict` for now instead:\r\n```python\r\nfrom datasets import Dataset, Array2D, Features, Value\r\nimport pandas as pd\r\nimport numpy as np\r\n\r\ndataset = {\r\n 'bbox': [\r\n np.array([[1,2,3,4],[1,2,3,4],[1,2,3,...
1,616,084,954,000
1,616,676,413,000
1,616,676,413,000
NONE
null
Hi, I'm trying to put together a `datasets.Dataset` to be used with LayoutLM which is available in `transformers`. This model requires as input the bounding boxes of each of the token of a sequence. This is when I realized that `Dataset` does not support multi-dimensional arrays as a value for a column in a row. ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2080/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2080/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2079
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2079/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2079/comments
https://api.github.com/repos/huggingface/datasets/issues/2079/events
https://github.com/huggingface/datasets/pull/2079
834,920,493
MDExOlB1bGxSZXF1ZXN0NTk1NjU2MDQ5
2,079
Refactorize Metric.compute signature to force keyword arguments only
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,079,950,000
1,616,513,504,000
1,616,513,504,000
MEMBER
null
Minor refactoring of Metric.compute signature to force the use of keyword arguments, by using the single star syntax.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2079/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2079/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2079", "html_url": "https://github.com/huggingface/datasets/pull/2079", "diff_url": "https://github.com/huggingface/datasets/pull/2079.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2079.patch", "merged_at": 1616513504000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2078
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2078/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2078/comments
https://api.github.com/repos/huggingface/datasets/issues/2078/events
https://github.com/huggingface/datasets/issues/2078
834,694,819
MDU6SXNzdWU4MzQ2OTQ4MTk=
2,078
MemoryError when computing WER metric
{ "login": "diego-fustes", "id": 5707233, "node_id": "MDQ6VXNlcjU3MDcyMzM=", "avatar_url": "https://avatars.githubusercontent.com/u/5707233?v=4", "gravatar_id": "", "url": "https://api.github.com/users/diego-fustes", "html_url": "https://github.com/diego-fustes", "followers_url": "https://api.github.com...
[ { "id": 2067393914, "node_id": "MDU6TGFiZWwyMDY3MzkzOTE0", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug", "name": "metric bug", "color": "25b21e", "default": false, "description": "A bug in a metric script" } ]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! Thanks for reporting.\r\nWe're indeed using `jiwer` to compute the WER.\r\n\r\nMaybe instead of calling `jiwer.wer` once for all the preditions/references we can compute the WER iteratively to avoid memory issues ? I'm not too familial with `jiwer` but this must be possible.\r\n\r\nCurrently the code to compu...
1,616,067,005,000
1,619,857,909,000
1,617,693,643,000
NONE
null
Hi, I'm trying to follow the ASR example to try Wav2Vec. This is the code that I use for WER calculation: ``` wer = load_metric("wer") print(wer.compute(predictions=result["predicted"], references=result["target"])) ``` However, I receive the following exception: `Traceback (most recent call last): File ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2078/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2078/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2077
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2077/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2077/comments
https://api.github.com/repos/huggingface/datasets/issues/2077/events
https://github.com/huggingface/datasets/pull/2077
834,649,536
MDExOlB1bGxSZXF1ZXN0NTk1NDI0MTYw
2,077
Bump huggingface_hub version
{ "login": "SBrandeis", "id": 33657802, "node_id": "MDQ6VXNlcjMzNjU3ODAy", "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SBrandeis", "html_url": "https://github.com/SBrandeis", "followers_url": "https://api.github.com/users/...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "🔥 " ]
1,616,064,874,000
1,616,067,206,000
1,616,067,206,000
CONTRIBUTOR
null
`0.0.2 => 0.0.6`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2077/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2077/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2077", "html_url": "https://github.com/huggingface/datasets/pull/2077", "diff_url": "https://github.com/huggingface/datasets/pull/2077.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2077.patch", "merged_at": 1616067206000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2076
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2076/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2076/comments
https://api.github.com/repos/huggingface/datasets/issues/2076/events
https://github.com/huggingface/datasets/issues/2076
834,445,296
MDU6SXNzdWU4MzQ0NDUyOTY=
2,076
Issue: Dataset download error
{ "login": "XuhuiZhou", "id": 20436061, "node_id": "MDQ6VXNlcjIwNDM2MDYx", "avatar_url": "https://avatars.githubusercontent.com/u/20436061?v=4", "gravatar_id": "", "url": "https://api.github.com/users/XuhuiZhou", "html_url": "https://github.com/XuhuiZhou", "followers_url": "https://api.github.com/users/...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi @XuhuiZhou, thanks for reporting this issue. \r\n\r\nIndeed, the old links are no longer valid (404 Not Found error), and the script must be updated with the new links to Google Drive.", "It would be nice to update the urls indeed !\r\n\r\nTo do this, you just need to replace the urls in `iwslt2017.py` and th...
1,616,049,366,000
1,616,413,951,000
null
NONE
null
The download link in `iwslt2017.py` file does not seem to work anymore. For example, `FileNotFoundError: Couldn't find file at https://wit3.fbk.eu/archive/2017-01-trnted/texts/zh/en/zh-en.tgz` Would be nice if we could modify it script and use the new downloadable link?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2076/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2076/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2075
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2075/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2075/comments
https://api.github.com/repos/huggingface/datasets/issues/2075/events
https://github.com/huggingface/datasets/issues/2075
834,301,246
MDU6SXNzdWU4MzQzMDEyNDY=
2,075
ConnectionError: Couldn't reach common_voice.py
{ "login": "LifaSun", "id": 6188893, "node_id": "MDQ6VXNlcjYxODg4OTM=", "avatar_url": "https://avatars.githubusercontent.com/u/6188893?v=4", "gravatar_id": "", "url": "https://api.github.com/users/LifaSun", "html_url": "https://github.com/LifaSun", "followers_url": "https://api.github.com/users/LifaSun/...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi @LifaSun, thanks for reporting this issue.\r\n\r\nSometimes, GitHub has some connectivity problems. Could you confirm that the problem persists?", "@albertvillanova Thanks! It works well now. " ]
1,616,030,346,000
1,616,236,181,000
1,616,236,181,000
NONE
null
When I run: from datasets import load_dataset, load_metric common_voice_train = load_dataset("common_voice", "zh-CN", split="train+validation") common_voice_test = load_dataset("common_voice", "zh-CN", split="test") Got: ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/ma...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2075/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2075/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2074
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2074/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2074/comments
https://api.github.com/repos/huggingface/datasets/issues/2074/events
https://github.com/huggingface/datasets/pull/2074
834,268,463
MDExOlB1bGxSZXF1ZXN0NTk1MTIzMjYw
2,074
Fix size categories in YAML Tags
{ "login": "gchhablani", "id": 29076344, "node_id": "MDQ6VXNlcjI5MDc2MzQ0", "avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gchhablani", "html_url": "https://github.com/gchhablani", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "> It would be great if there was a way to make the task categories consistent too. For this, the streamlit app can look into all the datasets and check for existing categories and show them in the list. This may add some consistency.\r\n\r\nWe can also update the task lists here: https://github.com/huggingface/dat...
1,616,025,756,000
1,616,519,470,000
1,616,519,470,000
CONTRIBUTOR
null
This PR fixes several `size_categories` in YAML tags and makes them consistent. Additionally, I have added a few more categories after `1M`, up to `1T`. I would like to add that to the streamlit app also. This PR also adds a couple of infos that I found missing. The code for generating this: ```python for datas...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2074/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2074/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2074", "html_url": "https://github.com/huggingface/datasets/pull/2074", "diff_url": "https://github.com/huggingface/datasets/pull/2074.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2074.patch", "merged_at": 1616519469000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2073
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2073/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2073/comments
https://api.github.com/repos/huggingface/datasets/issues/2073/events
https://github.com/huggingface/datasets/pull/2073
834,192,501
MDExOlB1bGxSZXF1ZXN0NTk1MDYyMzQ2
2,073
Fixes check of TF_AVAILABLE and TORCH_AVAILABLE
{ "login": "philschmid", "id": 32632186, "node_id": "MDQ6VXNlcjMyNjMyMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/32632186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/philschmid", "html_url": "https://github.com/philschmid", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,616,016,533,000
1,616,058,565,000
1,616,058,564,000
MEMBER
null
# What is this PR doing This PR implements the checks if `Tensorflow` and `Pytorch` are available the same way as `transformers` does it. I added the additional checks for the different `Tensorflow` and `torch` versions. #2068
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2073/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2073/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2073", "html_url": "https://github.com/huggingface/datasets/pull/2073", "diff_url": "https://github.com/huggingface/datasets/pull/2073.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2073.patch", "merged_at": 1616058564000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2072
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2072/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2072/comments
https://api.github.com/repos/huggingface/datasets/issues/2072/events
https://github.com/huggingface/datasets/pull/2072
834,054,837
MDExOlB1bGxSZXF1ZXN0NTk0OTQ5NjA4
2,072
Fix docstring issues
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "id": 1935892861, "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "I think I will stop pushing to this PR, so that it can me merged for today release. \r\n\r\nI will open another PR for further fixing docs.\r\n\r\nDo you agree, @lhoestq ?", "Sounds good thanks !" ]
1,616,004,824,000
1,616,574,057,000
1,616,071,281,000
MEMBER
null
Fix docstring issues.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2072/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2072/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2072", "html_url": "https://github.com/huggingface/datasets/pull/2072", "diff_url": "https://github.com/huggingface/datasets/pull/2072.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2072.patch", "merged_at": 1616071281000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2071
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2071/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2071/comments
https://api.github.com/repos/huggingface/datasets/issues/2071/events
https://github.com/huggingface/datasets/issues/2071
833,950,824
MDU6SXNzdWU4MzM5NTA4MjQ=
2,071
Multiprocessing is slower than single process
{ "login": "theo-m", "id": 17948980, "node_id": "MDQ6VXNlcjE3OTQ4OTgw", "avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theo-m", "html_url": "https://github.com/theo-m", "followers_url": "https://api.github.com/users/theo-m/fo...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "dupe of #1992" ]
1,615,997,338,000
1,616,058,623,000
1,616,058,623,000
CONTRIBUTOR
null
```python # benchmark_filter.py import logging import sys import time from datasets import load_dataset, set_caching_enabled if __name__ == "__main__": set_caching_enabled(False) logging.basicConfig(level=logging.DEBUG) bc = load_dataset("bookcorpus") now = time.time() try: ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2071/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2071/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2070
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2070/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2070/comments
https://api.github.com/repos/huggingface/datasets/issues/2070/events
https://github.com/huggingface/datasets/issues/2070
833,799,035
MDU6SXNzdWU4MzM3OTkwMzU=
2,070
ArrowInvalid issue for squad v2 dataset
{ "login": "MichaelYxWang", "id": 29818977, "node_id": "MDQ6VXNlcjI5ODE4OTc3", "avatar_url": "https://avatars.githubusercontent.com/u/29818977?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MichaelYxWang", "html_url": "https://github.com/MichaelYxWang", "followers_url": "https://api.githu...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! This error happens when you use `map` in batched mode and then your function doesn't return the same number of values per column.\r\n\r\nIndeed since you're using `map` in batched mode, `prepare_validation_features` must take a batch as input (i.e. a dictionary of multiple rows of the dataset), and return a b...
1,615,989,109,000
1,628,099,836,000
1,628,099,836,000
NONE
null
Hello, I am using the huggingface official question answering example notebook (https://colab.research.google.com/github/huggingface/notebooks/blob/master/examples/question_answering.ipynb). In the prepare_validation_features function, I made some modifications to tokenize a new set of quesions with the original co...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2070/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2070/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2069
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2069/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2069/comments
https://api.github.com/repos/huggingface/datasets/issues/2069/events
https://github.com/huggingface/datasets/pull/2069
833,768,926
MDExOlB1bGxSZXF1ZXN0NTk0NzA5ODYw
2,069
Add and fix docstring for NamedSplit
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Maybe we should add some other split classes?" ]
1,615,987,168,000
1,616,063,260,000
1,616,063,260,000
MEMBER
null
Add and fix docstring for `NamedSplit`, which was missing.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2069/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2069/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2069", "html_url": "https://github.com/huggingface/datasets/pull/2069", "diff_url": "https://github.com/huggingface/datasets/pull/2069.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2069.patch", "merged_at": 1616063260000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2068
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2068/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2068/comments
https://api.github.com/repos/huggingface/datasets/issues/2068/events
https://github.com/huggingface/datasets/issues/2068
833,602,832
MDU6SXNzdWU4MzM2MDI4MzI=
2,068
PyTorch not available error on SageMaker GPU docker though it is installed
{ "login": "sivakhno", "id": 1651457, "node_id": "MDQ6VXNlcjE2NTE0NTc=", "avatar_url": "https://avatars.githubusercontent.com/u/1651457?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sivakhno", "html_url": "https://github.com/sivakhno", "followers_url": "https://api.github.com/users/sivak...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "cc @philschmid ", "Hey @sivakhno,\r\n\r\nhow does your `requirements.txt` look like to install the `datasets` library and which version of it are you running? Can you try to install `datasets>=1.4.0`", "Hi @philschmid - thanks for suggestion. I am using `datasets==1.4.1`. \r\nI have also tried using `torch=1.6...
1,615,975,467,000
1,623,646,050,000
1,623,646,050,000
NONE
null
I get en error when running data loading using SageMaker SDK ``` File "main.py", line 34, in <module> run_training() File "main.py", line 25, in run_training dm.setup('fit') File "/opt/conda/lib/python3.6/site-packages/pytorch_lightning/core/datamodule.py", line 92, in wrapped_fn return fn(*a...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2068/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2068/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2067
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2067/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2067/comments
https://api.github.com/repos/huggingface/datasets/issues/2067/events
https://github.com/huggingface/datasets/issues/2067
833,559,940
MDU6SXNzdWU4MzM1NTk5NDA=
2,067
Multiprocessing windows error
{ "login": "flozi00", "id": 47894090, "node_id": "MDQ6VXNlcjQ3ODk0MDkw", "avatar_url": "https://avatars.githubusercontent.com/u/47894090?v=4", "gravatar_id": "", "url": "https://api.github.com/users/flozi00", "html_url": "https://github.com/flozi00", "followers_url": "https://api.github.com/users/flozi0...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! Thanks for reporting.\r\nThis looks like a bug, could you try to provide a minimal code example that reproduces the issue ? This would be very helpful !\r\n\r\nOtherwise I can try to run the wav2vec2 code above on my side but probably not this week..", "```\r\nfrom datasets import load_dataset\r\n\r\ndatase...
1,615,972,348,000
1,628,099,948,000
1,628,099,948,000
NONE
null
As described here https://huggingface.co/blog/fine-tune-xlsr-wav2vec2 When using the num_proc argument on windows the whole Python environment crashes and hanging in loop. For example at the map_to_array part. An error occures because the cache file already exists and windows throws and error. After this the log c...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2067/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2067/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2066
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2066/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2066/comments
https://api.github.com/repos/huggingface/datasets/issues/2066/events
https://github.com/huggingface/datasets/pull/2066
833,480,551
MDExOlB1bGxSZXF1ZXN0NTk0NDcwMjEz
2,066
Fix docstring rendering of Dataset/DatasetDict.from_csv args
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,615,965,790,000
1,615,972,881,000
1,615,972,881,000
MEMBER
null
Fix the docstring rendering of Dataset/DatasetDict.from_csv args.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2066/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2066/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2066", "html_url": "https://github.com/huggingface/datasets/pull/2066", "diff_url": "https://github.com/huggingface/datasets/pull/2066.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2066.patch", "merged_at": 1615972881000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2065
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2065/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2065/comments
https://api.github.com/repos/huggingface/datasets/issues/2065/events
https://github.com/huggingface/datasets/issues/2065
833,291,432
MDU6SXNzdWU4MzMyOTE0MzI=
2,065
Only user permission of saved cache files, not group
{ "login": "lorr1", "id": 57237365, "node_id": "MDQ6VXNlcjU3MjM3MzY1", "avatar_url": "https://avatars.githubusercontent.com/u/57237365?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lorr1", "html_url": "https://github.com/lorr1", "followers_url": "https://api.github.com/users/lorr1/follow...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" }, { "id": 1935892877, "node_id": "MDU6...
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi ! Thanks for reporting.\r\n\r\nCurrently there's no way to specify this.\r\n\r\nWhen loading/processing a dataset, the arrow file is written using a temporary file. Then once writing is finished, it's moved to the cache directory (using `shutil.move` [here](https://github.com/huggingface/datasets/blob/f6b8251eb...
1,615,940,422,000
1,620,629,129,000
1,620,629,129,000
NONE
null
Hello, It seems when a cached file is saved from calling `dataset.map` for preprocessing, it gets the user permissions and none of the user's group permissions. As we share data files across members of our team, this is causing a bit of an issue as we have to continually reset the permission of the files. Do you kno...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2065/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2065/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2064
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2064/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2064/comments
https://api.github.com/repos/huggingface/datasets/issues/2064/events
https://github.com/huggingface/datasets/pull/2064
833,002,360
MDExOlB1bGxSZXF1ZXN0NTk0MDczOTQ1
2,064
Fix ted_talks_iwslt version error
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,615,913,025,000
1,615,917,608,000
1,615,917,608,000
CONTRIBUTOR
null
This PR fixes the bug where the version argument would be passed twice if the dataset configuration was created on the fly. Fixes #2059
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2064/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2064/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2064", "html_url": "https://github.com/huggingface/datasets/pull/2064", "diff_url": "https://github.com/huggingface/datasets/pull/2064.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2064.patch", "merged_at": 1615917607000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2063
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2063/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2063/comments
https://api.github.com/repos/huggingface/datasets/issues/2063/events
https://github.com/huggingface/datasets/pull/2063
832,993,705
MDExOlB1bGxSZXF1ZXN0NTk0MDY2NzI5
2,063
[Common Voice] Adapt dataset script so that no manual data download is actually needed
{ "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "followers_url": "https://...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,615,912,424,000
1,615,974,172,000
1,615,974,157,000
MEMBER
null
This PR changes the dataset script so that no manual data dir is needed anymore.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2063/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2063/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2063", "html_url": "https://github.com/huggingface/datasets/pull/2063", "diff_url": "https://github.com/huggingface/datasets/pull/2063.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2063.patch", "merged_at": 1615974157000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2062
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2062/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2062/comments
https://api.github.com/repos/huggingface/datasets/issues/2062/events
https://github.com/huggingface/datasets/pull/2062
832,625,483
MDExOlB1bGxSZXF1ZXN0NTkzNzUyNTMz
2,062
docs: fix missing quotation
{ "login": "neal2018", "id": 46561493, "node_id": "MDQ6VXNlcjQ2NTYxNDkz", "avatar_url": "https://avatars.githubusercontent.com/u/46561493?v=4", "gravatar_id": "", "url": "https://api.github.com/users/neal2018", "html_url": "https://github.com/neal2018", "followers_url": "https://api.github.com/users/nea...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,615,889,274,000
1,615,972,917,000
1,615,972,917,000
CONTRIBUTOR
null
The json code misses a quote
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2062/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2062/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2062", "html_url": "https://github.com/huggingface/datasets/pull/2062", "diff_url": "https://github.com/huggingface/datasets/pull/2062.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2062.patch", "merged_at": 1615972916000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2061
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2061/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2061/comments
https://api.github.com/repos/huggingface/datasets/issues/2061/events
https://github.com/huggingface/datasets/issues/2061
832,596,228
MDU6SXNzdWU4MzI1OTYyMjg=
2,061
Cannot load udpos subsets from xtreme dataset using load_dataset()
{ "login": "adzcodez", "id": 55791365, "node_id": "MDQ6VXNlcjU1NzkxMzY1", "avatar_url": "https://avatars.githubusercontent.com/u/55791365?v=4", "gravatar_id": "", "url": "https://api.github.com/users/adzcodez", "html_url": "https://github.com/adzcodez", "followers_url": "https://api.github.com/users/adz...
[ { "id": 1935892877, "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue", "name": "good first issue", "color": "7057ff", "default": true, "description": "Good for newcomers" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "@lhoestq Adding \"_\" to the class labels in the dataset script will fix the issue.\r\n\r\nThe bigger issue IMO is that the data files are in conll format, but the examples are tokens, not sentences.", "Hi ! Thanks for reporting @adzcodez \r\n\r\n\r\n> @lhoestq Adding \"_\" to the class labels in the dataset scr...
1,615,887,133,000
1,624,017,251,000
1,624,017,250,000
NONE
null
Hello, I am trying to load the udpos English subset from xtreme dataset, but this faces an error during loading. I am using datasets v1.4.1, pip install. I have tried with other udpos languages which also fail, though loading a different subset altogether (such as XNLI) has no issue. I have also tried on Colab and ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2061/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2061/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2060
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2060/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2060/comments
https://api.github.com/repos/huggingface/datasets/issues/2060/events
https://github.com/huggingface/datasets/pull/2060
832,588,591
MDExOlB1bGxSZXF1ZXN0NTkzNzIxNzcx
2,060
Filtering refactor
{ "login": "theo-m", "id": 17948980, "node_id": "MDQ6VXNlcjE3OTQ4OTgw", "avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theo-m", "html_url": "https://github.com/theo-m", "followers_url": "https://api.github.com/users/theo-m/fo...
[]
closed
false
{ "login": "theo-m", "id": 17948980, "node_id": "MDQ6VXNlcjE3OTQ4OTgw", "avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theo-m", "html_url": "https://github.com/theo-m", "followers_url": "https://api.github.com/users/theo-m/fo...
[ { "login": "theo-m", "id": 17948980, "node_id": "MDQ6VXNlcjE3OTQ4OTgw", "avatar_url": "https://avatars.githubusercontent.com/u/17948980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theo-m", "html_url": "https://github.com/theo-m", "followers_url": "https://api.github...
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "I thought at first that the multiproc test was not relevant now that we do stuff only in memory, but I think there's something that's actually broken, my tiny benchmark on bookcorpus runs forever (2hrs+) when I add `num_proc=4` as a kwarg, will investigate 👀 \r\n\r\nI'm not familiar with the caching you describe ...
1,615,886,610,000
1,634,116,144,000
1,634,116,143,000
CONTRIBUTOR
null
fix https://github.com/huggingface/datasets/issues/2032 benchmarking is somewhat inconclusive, currently running on `book_corpus` with: ```python bc = load_dataset("bookcorpus") now = time.time() bc.filter(lambda x: len(x["text"]) < 64) elapsed = time.time() - now print(elapsed) ``` t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2060/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2060/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2060", "html_url": "https://github.com/huggingface/datasets/pull/2060", "diff_url": "https://github.com/huggingface/datasets/pull/2060.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2060.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/2059
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2059/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2059/comments
https://api.github.com/repos/huggingface/datasets/issues/2059/events
https://github.com/huggingface/datasets/issues/2059
832,579,156
MDU6SXNzdWU4MzI1NzkxNTY=
2,059
Error while following docs to load the `ted_talks_iwslt` dataset
{ "login": "ekdnam", "id": 40426312, "node_id": "MDQ6VXNlcjQwNDI2MzEy", "avatar_url": "https://avatars.githubusercontent.com/u/40426312?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ekdnam", "html_url": "https://github.com/ekdnam", "followers_url": "https://api.github.com/users/ekdnam/fo...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "@skyprince999 as you authored the PR for this dataset, any comments?", "This has been fixed in #2064 by @mariosasko (thanks again !)\r\n\r\nThe fix is available on the master branch and we'll do a new release very soon :)" ]
1,615,885,939,000
1,615,917,631,000
1,615,917,607,000
NONE
null
I am currently trying to load the `ted_talks_iwslt` dataset into google colab. The [docs](https://huggingface.co/datasets/ted_talks_iwslt) mention the following way of doing so. ```python dataset = load_dataset("ted_talks_iwslt", language_pair=("it", "pl"), year="2014") ``` Executing it results in the error ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2059/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2059/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2058
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2058/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2058/comments
https://api.github.com/repos/huggingface/datasets/issues/2058/events
https://github.com/huggingface/datasets/issues/2058
832,159,844
MDU6SXNzdWU4MzIxNTk4NDQ=
2,058
Is it possible to convert a `tfds` to HuggingFace `dataset`?
{ "login": "abarbosa94", "id": 6608232, "node_id": "MDQ6VXNlcjY2MDgyMzI=", "avatar_url": "https://avatars.githubusercontent.com/u/6608232?v=4", "gravatar_id": "", "url": "https://api.github.com/users/abarbosa94", "html_url": "https://github.com/abarbosa94", "followers_url": "https://api.github.com/users...
[]
open
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,615,839,527,000
1,615,839,527,000
null
CONTRIBUTOR
null
I was having some weird bugs with `C4`dataset version of HuggingFace, so I decided to try to download `C4`from `tfds`. I would like to know if it is possible to convert a tfds dataset to HuggingFace dataset format :) I can also open a new issue reporting the bug I'm receiving with `datasets.load_dataset('c4','en')` ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2058/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2058/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2057
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2057/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2057/comments
https://api.github.com/repos/huggingface/datasets/issues/2057/events
https://github.com/huggingface/datasets/pull/2057
832,120,522
MDExOlB1bGxSZXF1ZXN0NTkzMzMzMjM0
2,057
update link to ZEST dataset
{ "login": "matt-peters", "id": 619844, "node_id": "MDQ6VXNlcjYxOTg0NA==", "avatar_url": "https://avatars.githubusercontent.com/u/619844?v=4", "gravatar_id": "", "url": "https://api.github.com/users/matt-peters", "html_url": "https://github.com/matt-peters", "followers_url": "https://api.github.com/user...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[]
1,615,836,177,000
1,615,914,388,000
1,615,914,388,000
CONTRIBUTOR
null
Updating the link as the original one is no longer working.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2057/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2057/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2057", "html_url": "https://github.com/huggingface/datasets/pull/2057", "diff_url": "https://github.com/huggingface/datasets/pull/2057.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2057.patch", "merged_at": 1615914388000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2056
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2056/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2056/comments
https://api.github.com/repos/huggingface/datasets/issues/2056/events
https://github.com/huggingface/datasets/issues/2056
831,718,397
MDU6SXNzdWU4MzE3MTgzOTc=
2,056
issue with opus100/en-fr dataset
{ "login": "dorost1234", "id": 79165106, "node_id": "MDQ6VXNlcjc5MTY1MTA2", "avatar_url": "https://avatars.githubusercontent.com/u/79165106?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dorost1234", "html_url": "https://github.com/dorost1234", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "@lhoestq I also deleted the cache and redownload the file and still the same issue, I appreciate any help on this. thanks ", "Here please find the minimal code to reproduce the issue @lhoestq note this only happens with MT5TokenizerFast\r\n\r\n```\r\nfrom datasets import load_dataset\r\nfrom transformers impor...
1,615,807,962,000
1,615,909,740,000
1,615,909,739,000
NONE
null
Hi I am running run_mlm.py code of huggingface repo with opus100/fr-en pair, I am getting this error, note that this error occurs for only this pairs and not the other pairs. Any idea why this is occurring? and how I can solve this? Thanks a lot @lhoestq for your help in advance. ` thread '<unnamed>' panicked...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2056/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2056/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2055
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2055/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2055/comments
https://api.github.com/repos/huggingface/datasets/issues/2055/events
https://github.com/huggingface/datasets/issues/2055
831,684,312
MDU6SXNzdWU4MzE2ODQzMTI=
2,055
is there a way to override a dataset object saved with save_to_disk?
{ "login": "shamanez", "id": 16892570, "node_id": "MDQ6VXNlcjE2ODkyNTcw", "avatar_url": "https://avatars.githubusercontent.com/u/16892570?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shamanez", "html_url": "https://github.com/shamanez", "followers_url": "https://api.github.com/users/sha...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi\r\nYou can rename the arrow file and update the name in `state.json`", "I tried this way, but when there is a mapping process to the dataset, it again uses a random cache name. atm, I am trying to use the following method by setting an exact cache file,\r\n\r\n```\r\n dataset_with_embedding =csv_da...
1,615,805,453,000
1,616,385,977,000
1,616,385,977,000
NONE
null
At the moment when I use save_to_disk, it uses the arbitrary name for the arrow file. Is there a way to override such an object?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2055/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2055/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2054
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2054/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2054/comments
https://api.github.com/repos/huggingface/datasets/issues/2054/events
https://github.com/huggingface/datasets/issues/2054
831,597,665
MDU6SXNzdWU4MzE1OTc2NjU=
2,054
Could not find file for ZEST dataset
{ "login": "bhadreshpsavani", "id": 26653468, "node_id": "MDQ6VXNlcjI2NjUzNDY4", "avatar_url": "https://avatars.githubusercontent.com/u/26653468?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bhadreshpsavani", "html_url": "https://github.com/bhadreshpsavani", "followers_url": "https://api...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "The zest dataset url was changed (allenai/zest#3) and #2057 should resolve this.", "This has been fixed in #2057 by @matt-peters (thanks again !)\r\n\r\nThe fix is available on the master branch and we'll do a new release very soon :)", "Thanks @lhoestq and @matt-peters ", "I am closing this issue since its ...
1,615,799,518,000
1,620,034,224,000
1,620,034,224,000
CONTRIBUTOR
null
I am trying to use zest dataset from Allen AI using below code in colab, ``` !pip install -q datasets from datasets import load_dataset dataset = load_dataset("zest") ``` I am getting the following error, ``` Using custom data configuration default Downloading and preparing dataset zest/default (download: ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2054/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2054/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false
https://api.github.com/repos/huggingface/datasets/issues/2053
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2053/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2053/comments
https://api.github.com/repos/huggingface/datasets/issues/2053/events
https://github.com/huggingface/datasets/pull/2053
831,151,728
MDExOlB1bGxSZXF1ZXN0NTkyNTM4ODY2
2,053
Add bAbI QA tasks
{ "login": "gchhablani", "id": 29076344, "node_id": "MDQ6VXNlcjI5MDc2MzQ0", "avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gchhablani", "html_url": "https://github.com/gchhablani", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi @lhoestq,\r\n\r\nShould I remove the 160 configurations? Is it too much?\r\n\r\nEDIT:\r\nCan you also check the task category? I'm not sure if there is an appropriate tag for the same.", "Thanks for the changes !\r\n\r\n> Should I remove the 160 configurations? Is it too much?\r\n\r\nYea 160 configuration is ...
1,615,727,079,000
1,617,021,708,000
1,617,021,708,000
CONTRIBUTOR
null
- **Name:** *The (20) QA bAbI tasks* - **Description:** *The (20) QA bAbI tasks are a set of proxy tasks that evaluate reading comprehension via question answering. Our tasks measure understanding in several ways: whether a system is able to answer questions via chaining facts, simple induction, deduction and many mor...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2053/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2053/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/2053", "html_url": "https://github.com/huggingface/datasets/pull/2053", "diff_url": "https://github.com/huggingface/datasets/pull/2053.diff", "patch_url": "https://github.com/huggingface/datasets/pull/2053.patch", "merged_at": 1617021708000 }
true
https://api.github.com/repos/huggingface/datasets/issues/2052
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/2052/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/2052/comments
https://api.github.com/repos/huggingface/datasets/issues/2052/events
https://github.com/huggingface/datasets/issues/2052
831,135,704
MDU6SXNzdWU4MzExMzU3MDQ=
2,052
Timit_asr dataset repeats examples
{ "login": "fermaat", "id": 7583522, "node_id": "MDQ6VXNlcjc1ODM1MjI=", "avatar_url": "https://avatars.githubusercontent.com/u/7583522?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fermaat", "html_url": "https://github.com/fermaat", "followers_url": "https://api.github.com/users/fermaat/...
[]
closed
false
{ "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url": "", "gists_url": "", "starred_url": "", "subscriptions_url": "", "organizations_url": "", "repos_url": "", "events_url": "", "received_events_url":...
[]
{ "url": "", "html_url": "", "labels_url": "", "id": 0, "node_id": "", "number": 0, "title": "", "description": "", "creator": { "login": "", "id": 0, "node_id": "", "avatar_url": "", "gravatar_id": "", "url": "", "html_url": "", "followers_url": "", "following_url"...
[ "Hi,\r\n\r\nthis was fixed by #1995, so you can wait for the next release or install the package directly from the master branch with the following command: \r\n```bash\r\npip install git+https://github.com/huggingface/datasets\r\n```", "Ty!" ]
1,615,722,223,000
1,615,804,636,000
1,615,804,636,000
NONE
null
Summary When loading timit_asr dataset on datasets 1.4+, every row in the dataset is the same Steps to reproduce As an example, on this code there is the text from the training part: Code snippet: ``` from datasets import load_dataset, load_metric timit = load_dataset("timit_asr") timit['train']['text']...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/2052/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/2052/timeline
null
null
{ "url": "", "html_url": "", "diff_url": "", "patch_url": "", "merged_at": 0 }
false