comments_url
stringlengths
70
70
timeline_url
stringlengths
70
70
closed_at
stringlengths
20
20
βŒ€
performed_via_github_app
null
state_reason
stringclasses
3 values
node_id
stringlengths
18
32
state
stringclasses
2 values
assignees
listlengths
0
4
draft
bool
2 classes
number
int64
1.61k
6.73k
user
dict
title
stringlengths
1
290
events_url
stringlengths
68
68
milestone
dict
labels_url
stringlengths
75
75
created_at
stringlengths
20
20
active_lock_reason
null
locked
bool
1 class
assignee
dict
pull_request
dict
id
int64
771M
2.18B
labels
listlengths
0
4
url
stringlengths
61
61
comments
listlengths
0
30
repository_url
stringclasses
1 value
author_association
stringclasses
3 values
body
stringlengths
0
228k
βŒ€
updated_at
stringlengths
20
20
html_url
stringlengths
49
51
reactions
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/2324/comments
https://api.github.com/repos/huggingface/datasets/issues/2324/timeline
2021-10-13T10:26:33Z
null
null
MDExOlB1bGxSZXF1ZXN0NjMwNzE1NTQz
closed
[]
false
2,324
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Create Audio feature
https://api.github.com/repos/huggingface/datasets/issues/2324/events
{ "closed_at": null, "closed_issues": 2, "created_at": "2021-07-21T15:34:56Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-08-30T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/8", "id": 6968069, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/8/labels", "node_id": "MI_kwDODunzps4AalMF", "number": 8, "open_issues": 4, "state": "open", "title": "1.12", "updated_at": "2021-10-13T10:26:33Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/8" }
https://api.github.com/repos/huggingface/datasets/issues/2324/labels{/name}
2021-05-05T15:55:22Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2324.diff", "html_url": "https://github.com/huggingface/datasets/pull/2324", "merged_at": "2021-10-13T10:26:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/2324.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2324" }
876,602,064
[]
https://api.github.com/repos/huggingface/datasets/issues/2324
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Create `Audio` feature to handle raw audio files. Some decisions to be further discussed: - I have chosen `soundfile` as the audio library; another interesting library is `librosa`, but this requires `soundfile` (see [here](https://github.com/librosa/librosa/blob/main/setup.cfg#L53)). If we require some more advanced functionalities, we could eventually switch the library. - I have implemented the audio feature as an extra: `pip install datasets[audio]`. For the moment, the typical datasets user uses only text datasets, and there is no need for them for additional package requirements for audio/image if they do not need them. - For tests, I require audio dependencies (so that all audio functionalities are checked with our CI test suite); I exclude Linux platforms, which require an additional library to be installed with the distribution package manager - I also require `pytest-datadir`, which allow to have (audio) data files for tests - The audio data contain: array and sample_rate. - The array is reshaped as 1D array (expected input for `Wav2Vec2`). Note that to install `soundfile` on Linux, you need to install `libsndfile` using your distribution’s package manager, for example `sudo apt-get install libsndfile1`. ## Requirements Specification - Access example with audio loading and resampling: ```python ds[0]["audio"] ``` - Map with audio loading & resampling: ```python def preprocess(batch): batch["input_values"] = processor(batch["audio"]).input_values return batch ds = ds.map(preprocess) ``` - Map without audio loading and resampling: ```python def preprocess(batch): batch["labels"] = processor(batch["target_text"]).input_values return batch ds = ds.map(preprocess) ``` - Additional requirement specification (see https://github.com/huggingface/datasets/pull/2324#pullrequestreview-768864998): Cast audio column to change sampling sate: ```python ds = ds.cast_column("audio", Audio(sampling_rate=16_000)) ```
2021-10-13T10:26:33Z
https://github.com/huggingface/datasets/pull/2324
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 3, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/2324/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2323/comments
https://api.github.com/repos/huggingface/datasets/issues/2323/timeline
2021-05-07T10:32:30Z
null
completed
MDU6SXNzdWU4NzY0Mzg1MDc=
closed
[]
null
2,323
{ "avatar_url": "https://avatars.githubusercontent.com/u/33647474?v=4", "events_url": "https://api.github.com/users/ekeleshian/events{/privacy}", "followers_url": "https://api.github.com/users/ekeleshian/followers", "following_url": "https://api.github.com/users/ekeleshian/following{/other_user}", "gists_url": "https://api.github.com/users/ekeleshian/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ekeleshian", "id": 33647474, "login": "ekeleshian", "node_id": "MDQ6VXNlcjMzNjQ3NDc0", "organizations_url": "https://api.github.com/users/ekeleshian/orgs", "received_events_url": "https://api.github.com/users/ekeleshian/received_events", "repos_url": "https://api.github.com/users/ekeleshian/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ekeleshian/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ekeleshian/subscriptions", "type": "User", "url": "https://api.github.com/users/ekeleshian" }
load_dataset("timit_asr") gives back duplicates of just one sample text
https://api.github.com/repos/huggingface/datasets/issues/2323/events
null
https://api.github.com/repos/huggingface/datasets/issues/2323/labels{/name}
2021-05-05T13:14:48Z
null
false
null
null
876,438,507
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2323
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug When you look up on key ["train"] and then ['text'], you get back a list with just one sentence duplicated 4620 times. Namely, the sentence "Would such an act of refusal be useful?". Similarly when you look up ['test'] and then ['text'], the list is one sentence repeated "The bungalow was pleasantly situated near the shore." 1680 times. I tried to work around the issue by downgrading to datasets version 1.3.0, inspired by [this post](https://www.gitmemory.com/issue/huggingface/datasets/2052/798904836) and removing the entire huggingface directory from ~/.cache, but I still get the same issue. ## Steps to reproduce the bug ```python from datasets import load_dataset timit = load_dataset("timit_asr") print(timit['train']['text']) print(timit['test']['text']) ``` ## Expected Result Rows of diverse text, like how it is shown in the [wav2vec2.0 tutorial](https://colab.research.google.com/github/patrickvonplaten/notebooks/blob/master/Fine_tuning_Wav2Vec2_for_English_ASR.ipynb) <img width="485" alt="Screen Shot 2021-05-05 at 9 09 57 AM" src="https://user-images.githubusercontent.com/33647474/117146094-d9b77f00-ad81-11eb-8306-f281850c127a.png"> ## Actual results Rows of repeated text. <img width="319" alt="Screen Shot 2021-05-05 at 9 11 53 AM" src="https://user-images.githubusercontent.com/33647474/117146231-f8b61100-ad81-11eb-834a-fc10410b0c9c.png"> ## Versions - Datasets: 1.3.0 - Python: 3.9.1 - Platform: macOS-11.2.1-x86_64-i386-64bit}
2021-05-07T10:32:30Z
https://github.com/huggingface/datasets/issues/2323
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2323/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2322/comments
https://api.github.com/repos/huggingface/datasets/issues/2322/timeline
2021-06-08T19:08:21Z
null
completed
MDU6SXNzdWU4NzYzODM4NTM=
closed
[]
null
2,322
{ "avatar_url": "https://avatars.githubusercontent.com/u/2743060?v=4", "events_url": "https://api.github.com/users/villmow/events{/privacy}", "followers_url": "https://api.github.com/users/villmow/followers", "following_url": "https://api.github.com/users/villmow/following{/other_user}", "gists_url": "https://api.github.com/users/villmow/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/villmow", "id": 2743060, "login": "villmow", "node_id": "MDQ6VXNlcjI3NDMwNjA=", "organizations_url": "https://api.github.com/users/villmow/orgs", "received_events_url": "https://api.github.com/users/villmow/received_events", "repos_url": "https://api.github.com/users/villmow/repos", "site_admin": false, "starred_url": "https://api.github.com/users/villmow/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/villmow/subscriptions", "type": "User", "url": "https://api.github.com/users/villmow" }
Calls to map are not cached.
https://api.github.com/repos/huggingface/datasets/issues/2322/events
null
https://api.github.com/repos/huggingface/datasets/issues/2322/labels{/name}
2021-05-05T12:11:27Z
null
false
null
null
876,383,853
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2322
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Somehow caching does not work for me anymore. Am I doing something wrong, or is there anything that I missed? ## Steps to reproduce the bug ```python import datasets datasets.set_caching_enabled(True) sst = datasets.load_dataset("sst") def foo(samples, i): print("executed", i[:10]) return samples # first call x = sst.map(foo, batched=True, with_indices=True, num_proc=2) print('\n'*3, "#" * 30, '\n'*3) # second call y = sst.map(foo, batched=True, with_indices=True, num_proc=2) # print version import sys import platform print(f""" - Datasets: {datasets.__version__} - Python: {sys.version} - Platform: {platform.platform()} """) ``` ## Actual results This code prints the following output for me: ```bash No config specified, defaulting to: sst/default Reusing dataset sst (/home/johannes/.cache/huggingface/datasets/sst/default/1.0.0/b8a7889ef01c5d3ae8c379b84cc4080f8aad3ac2bc538701cbe0ac6416fb76ff) #0: 0%| | 0/5 [00:00<?, ?ba/s] #1: 0%| | 0/5 [00:00<?, ?ba/s] executed [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] executed [4272, 4273, 4274, 4275, 4276, 4277, 4278, 4279, 4280, 4281] executed [1000, 1001, 1002, 1003, 1004, 1005, 1006, 1007, 1008, 1009] executed [5272, 5273, 5274, 5275, 5276, 5277, 5278, 5279, 5280, 5281] executed [2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009] executed [6272, 6273, 6274, 6275, 6276, 6277, 6278, 6279, 6280, 6281] executed [3000, 3001, 3002, 3003, 3004, 3005, 3006, 3007, 3008, 3009] executed [7272, 7273, 7274, 7275, 7276, 7277, 7278, 7279, 7280, 7281] executed [4000, 4001, 4002, 4003, 4004, 4005, 4006, 4007, 4008, 4009] #0: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 5/5 [00:00<00:00, 59.85ba/s] executed [8272, 8273, 8274, 8275, 8276, 8277, 8278, 8279, 8280, 8281] #1: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 5/5 [00:00<00:00, 60.85ba/s] #0: 0%| | 0/1 [00:00<?, ?ba/s] #1: 0%| | 0/1 [00:00<?, ?ba/s]executed [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] #0: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:00<00:00, 69.32ba/s] executed [551, 552, 553, 554, 555, 556, 557, 558, 559, 560] #1: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:00<00:00, 70.93ba/s] #0: 0%| | 0/2 [00:00<?, ?ba/s] #1: 0%| | 0/2 [00:00<?, ?ba/s]executed [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] executed [1000, 1001, 1002, 1003, 1004, 1005, 1006, 1007, 1008, 1009] #0: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2/2 [00:00<00:00, 63.25ba/s] executed [1105, 1106, 1107, 1108, 1109, 1110, 1111, 1112, 1113, 1114] executed [2105, 2106, 2107, 2108, 2109, 2110, 2111, 2112, 2113, 2114] #1: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2/2 [00:00<00:00, 57.69ba/s] ############################## #0: 0%| | 0/5 [00:00<?, ?ba/s] #1: 0%| | 0/5 [00:00<?, ?ba/s] executed [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] executed [4272, 4273, 4274, 4275, 4276, 4277, 4278, 4279, 4280, 4281] executed [1000, 1001, 1002, 1003, 1004, 1005, 1006, 1007, 1008, 1009] executed [5272, 5273, 5274, 5275, 5276, 5277, 5278, 5279, 5280, 5281] executed [2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009] executed [6272, 6273, 6274, 6275, 6276, 6277, 6278, 6279, 6280, 6281] executed [3000, 3001, 3002, 3003, 3004, 3005, 3006, 3007, 3008, 3009] executed [4000, 4001, 4002, 4003, 4004, 4005, 4006, 4007, 4008, 4009] #0: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 5/5 [00:00<00:00, 58.10ba/s] executed [7272, 7273, 7274, 7275, 7276, 7277, 7278, 7279, 7280, 7281] executed [8272, 8273, 8274, 8275, 8276, 8277, 8278, 8279, 8280, 8281] #1: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 5/5 [00:00<00:00, 57.19ba/s] #0: 0%| | 0/1 [00:00<?, ?ba/s] #1: 0%| | 0/1 [00:00<?, ?ba/s] executed [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] #0: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:00<00:00, 60.10ba/s] executed [551, 552, 553, 554, 555, 556, 557, 558, 559, 560] #1: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:00<00:00, 53.82ba/s] #0: 0%| | 0/2 [00:00<?, ?ba/s] #1: 0%| | 0/2 [00:00<?, ?ba/s] executed [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] executed [1000, 1001, 1002, 1003, 1004, 1005, 1006, 1007, 1008, 1009] executed [1105, 1106, 1107, 1108, 1109, 1110, 1111, 1112, 1113, 1114] #0: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2/2 [00:00<00:00, 72.76ba/s] executed [2105, 2106, 2107, 2108, 2109, 2110, 2111, 2112, 2113, 2114] #1: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2/2 [00:00<00:00, 71.55ba/s] - Datasets: 1.6.1 - Python: 3.8.3 (default, May 19 2020, 18:47:26) [GCC 7.3.0] - Platform: Linux-5.4.0-72-generic-x86_64-with-glibc2.10 ``` ## Expected results Caching should work.
2021-06-08T19:10:02Z
https://github.com/huggingface/datasets/issues/2322
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2322/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2321/comments
https://api.github.com/repos/huggingface/datasets/issues/2321/timeline
2021-05-05T10:50:55Z
null
null
MDExOlB1bGxSZXF1ZXN0NjMwNDc3NDUy
closed
[]
false
2,321
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Set encoding in OSCAR dataset
https://api.github.com/repos/huggingface/datasets/issues/2321/events
null
https://api.github.com/repos/huggingface/datasets/issues/2321/labels{/name}
2021-05-05T10:27:03Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2321.diff", "html_url": "https://github.com/huggingface/datasets/pull/2321", "merged_at": "2021-05-05T10:50:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/2321.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2321" }
876,304,364
[]
https://api.github.com/repos/huggingface/datasets/issues/2321
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Set explicit `utf-8` encoding in OSCAR dataset, to avoid using the system default `cp1252` on Windows platforms. Fix #2319.
2021-05-05T10:50:55Z
https://github.com/huggingface/datasets/pull/2321
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2321/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2320/comments
https://api.github.com/repos/huggingface/datasets/issues/2320/timeline
2021-05-06T07:57:54Z
null
null
MDExOlB1bGxSZXF1ZXN0NjMwNDM5NjI5
closed
[]
false
2,320
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Set default name in init_dynamic_modules
https://api.github.com/repos/huggingface/datasets/issues/2320/events
null
https://api.github.com/repos/huggingface/datasets/issues/2320/labels{/name}
2021-05-05T09:30:03Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2320.diff", "html_url": "https://github.com/huggingface/datasets/pull/2320", "merged_at": "2021-05-06T07:57:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/2320.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2320" }
876,257,026
[]
https://api.github.com/repos/huggingface/datasets/issues/2320
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Set default value for the name of dynamic modules. Close #2318.
2021-05-06T07:57:54Z
https://github.com/huggingface/datasets/pull/2320
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2320/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2319/comments
https://api.github.com/repos/huggingface/datasets/issues/2319/timeline
2021-05-05T10:50:55Z
null
completed
MDU6SXNzdWU4NzYyNTEzNzY=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,319
{ "avatar_url": "https://avatars.githubusercontent.com/u/8904453?v=4", "events_url": "https://api.github.com/users/sgraaf/events{/privacy}", "followers_url": "https://api.github.com/users/sgraaf/followers", "following_url": "https://api.github.com/users/sgraaf/following{/other_user}", "gists_url": "https://api.github.com/users/sgraaf/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sgraaf", "id": 8904453, "login": "sgraaf", "node_id": "MDQ6VXNlcjg5MDQ0NTM=", "organizations_url": "https://api.github.com/users/sgraaf/orgs", "received_events_url": "https://api.github.com/users/sgraaf/received_events", "repos_url": "https://api.github.com/users/sgraaf/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sgraaf/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sgraaf/subscriptions", "type": "User", "url": "https://api.github.com/users/sgraaf" }
UnicodeDecodeError for OSCAR (Afrikaans)
https://api.github.com/repos/huggingface/datasets/issues/2319/events
null
https://api.github.com/repos/huggingface/datasets/issues/2319/labels{/name}
2021-05-05T09:22:52Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
876,251,376
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2319
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug When loading the [OSCAR dataset](https://huggingface.co/datasets/oscar) (specifically `unshuffled_deduplicated_af`), I encounter a `UnicodeDecodeError`. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("oscar", "unshuffled_deduplicated_af") ``` ## Expected results Anything but an error, really. ## Actual results ```python >>> from datasets import load_dataset >>> dataset = load_dataset("oscar", "unshuffled_deduplicated_af") Downloading: 14.7kB [00:00, 4.91MB/s] Downloading: 3.07MB [00:00, 32.6MB/s] Downloading and preparing dataset oscar/unshuffled_deduplicated_af (download: 62.93 MiB, generated: 163.38 MiB, post-processed: Unknown size, total: 226.32 MiB) to C:\Users\sgraaf\.cache\huggingface\datasets\oscar\unshuffled_deduplicated_af\1.0.0\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464... Downloading: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 81.0/81.0 [00:00<00:00, 40.5kB/s] Downloading: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 66.0M/66.0M [00:18<00:00, 3.50MB/s] Traceback (most recent call last): File "<stdin>", line 1, in <module> File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\site-packages\datasets\load.py", line 745, in load_dataset builder_instance.download_and_prepare( File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\site-packages\datasets\builder.py", line 574, in download_and_prepare self._download_and_prepare( File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\site-packages\datasets\builder.py", line 652, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\site-packages\datasets\builder.py", line 979, in _prepare_split for key, record in utils.tqdm( File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\site-packages\tqdm\std.py", line 1133, in __iter__ for obj in iterable: File "C:\Users\sgraaf\.cache\huggingface\modules\datasets_modules\datasets\oscar\bd4f96df5b4512007ef9fd17bbc1ecde459fa53d2fc0049cf99392ba2efcc464\oscar.py", line 359, in _generate_examples for line in f: File "C:\Users\sgraaf\AppData\Local\Programs\Python\Python39\lib\encodings\cp1252.py", line 23, in decode return codecs.charmap_decode(input,self.errors,decoding_table)[0] UnicodeDecodeError: 'charmap' codec can't decode byte 0x9d in position 7454: character maps to <undefined> ``` ## Versions Paste the output of the following code: ```python import datasets import sys import platform print(f""" - Datasets: {datasets.__version__} - Python: {sys.version} - Platform: {platform.platform()} """) ``` - Datasets: 1.6.2 - Python: 3.9.4 (tags/v3.9.4:1f2e308, Apr 6 2021, 13:40:21) [MSC v.1928 64 bit (AMD64)] - Platform: Windows-10-10.0.19041-SP0
2021-05-05T10:57:31Z
https://github.com/huggingface/datasets/issues/2319
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2319/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2318/comments
https://api.github.com/repos/huggingface/datasets/issues/2318/timeline
2021-05-06T07:57:54Z
null
completed
MDU6SXNzdWU4NzYyMTI0NjA=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" } ]
null
2,318
{ "avatar_url": "https://avatars.githubusercontent.com/u/4529381?v=4", "events_url": "https://api.github.com/users/richardliaw/events{/privacy}", "followers_url": "https://api.github.com/users/richardliaw/followers", "following_url": "https://api.github.com/users/richardliaw/following{/other_user}", "gists_url": "https://api.github.com/users/richardliaw/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/richardliaw", "id": 4529381, "login": "richardliaw", "node_id": "MDQ6VXNlcjQ1MjkzODE=", "organizations_url": "https://api.github.com/users/richardliaw/orgs", "received_events_url": "https://api.github.com/users/richardliaw/received_events", "repos_url": "https://api.github.com/users/richardliaw/repos", "site_admin": false, "starred_url": "https://api.github.com/users/richardliaw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/richardliaw/subscriptions", "type": "User", "url": "https://api.github.com/users/richardliaw" }
[api request] API to obtain "dataset_module" dynamic path?
https://api.github.com/repos/huggingface/datasets/issues/2318/events
null
https://api.github.com/repos/huggingface/datasets/issues/2318/labels{/name}
2021-05-05T08:40:48Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
null
876,212,460
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2318
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. This is an awesome library. It seems like the dynamic module path in this library has broken some of hyperparameter tuning functionality: https://discuss.huggingface.co/t/using-hyperparameter-search-in-trainer/785/34 This is because Ray will spawn new processes, and each process will load modules by path. However, we need to explicitly inform Ray to load the right modules, or else it will error upon import. I'd like an API to obtain the dynamic paths. This will allow us to support this functionality in this awesome library while being future proof. **Describe the solution you'd like** A clear and concise description of what you want to happen. `datasets.get_dynamic_paths -> List[str]` will be sufficient for my use case. By offering this API, we will be able to address the following issues (by patching the ray integration sufficiently): https://github.com/huggingface/blog/issues/106 https://github.com/huggingface/transformers/issues/11565 https://discuss.huggingface.co/t/using-hyperparameter-search-in-trainer/785/34 https://discuss.huggingface.co/t/using-hyperparameter-search-in-trainer/785/35
2021-05-06T08:45:45Z
https://github.com/huggingface/datasets/issues/2318
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2318/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2317/comments
https://api.github.com/repos/huggingface/datasets/issues/2317/timeline
2021-05-05T09:21:58Z
null
null
MDExOlB1bGxSZXF1ZXN0NjMwMDQxNzc4
closed
[]
false
2,317
{ "avatar_url": "https://avatars.githubusercontent.com/u/32267027?v=4", "events_url": "https://api.github.com/users/cemilcengiz/events{/privacy}", "followers_url": "https://api.github.com/users/cemilcengiz/followers", "following_url": "https://api.github.com/users/cemilcengiz/following{/other_user}", "gists_url": "https://api.github.com/users/cemilcengiz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cemilcengiz", "id": 32267027, "login": "cemilcengiz", "node_id": "MDQ6VXNlcjMyMjY3MDI3", "organizations_url": "https://api.github.com/users/cemilcengiz/orgs", "received_events_url": "https://api.github.com/users/cemilcengiz/received_events", "repos_url": "https://api.github.com/users/cemilcengiz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cemilcengiz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cemilcengiz/subscriptions", "type": "User", "url": "https://api.github.com/users/cemilcengiz" }
Fix incorrect version specification for the pyarrow package
https://api.github.com/repos/huggingface/datasets/issues/2317/events
null
https://api.github.com/repos/huggingface/datasets/issues/2317/labels{/name}
2021-05-04T19:30:20Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2317.diff", "html_url": "https://github.com/huggingface/datasets/pull/2317", "merged_at": "2021-05-05T09:21:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/2317.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2317" }
875,767,318
[]
https://api.github.com/repos/huggingface/datasets/issues/2317
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This PR addresses the bug in the pyarrow version specification, which is detailed in #2316 . Simply, I put a comma between the version bounds. Fix #2316.
2021-05-05T10:09:16Z
https://github.com/huggingface/datasets/pull/2317
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2317/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2316/comments
https://api.github.com/repos/huggingface/datasets/issues/2316/timeline
2021-05-05T10:10:03Z
null
completed
MDU6SXNzdWU4NzU3NTYzNTM=
closed
[]
null
2,316
{ "avatar_url": "https://avatars.githubusercontent.com/u/32267027?v=4", "events_url": "https://api.github.com/users/cemilcengiz/events{/privacy}", "followers_url": "https://api.github.com/users/cemilcengiz/followers", "following_url": "https://api.github.com/users/cemilcengiz/following{/other_user}", "gists_url": "https://api.github.com/users/cemilcengiz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cemilcengiz", "id": 32267027, "login": "cemilcengiz", "node_id": "MDQ6VXNlcjMyMjY3MDI3", "organizations_url": "https://api.github.com/users/cemilcengiz/orgs", "received_events_url": "https://api.github.com/users/cemilcengiz/received_events", "repos_url": "https://api.github.com/users/cemilcengiz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cemilcengiz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cemilcengiz/subscriptions", "type": "User", "url": "https://api.github.com/users/cemilcengiz" }
Incorrect version specification for pyarrow
https://api.github.com/repos/huggingface/datasets/issues/2316/events
null
https://api.github.com/repos/huggingface/datasets/issues/2316/labels{/name}
2021-05-04T19:15:11Z
null
false
null
null
875,756,353
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2316
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug The pyarrow dependency is incorrectly specified in setup.py file, in [this line](https://github.com/huggingface/datasets/blob/3a3e5a4da20bfcd75f8b6a6869b240af8feccc12/setup.py#L77). Also as a snippet: ```python "pyarrow>=1.0.0<4.0.0", ``` ## Steps to reproduce the bug ```bash pip install "pyarrow>=1.0.0<4.0.0" ``` ## Expected results It is expected to get a pyarrow version between 1.0.0 (inclusive) and 4.0.0 (exclusive). ## Actual results pip ignores the specified versions since there is a missing comma between the lower and upper limits. Therefore, pip installs the latest pyarrow version from PYPI, which is 4.0.0. This is especially problematic since "conda env export" fails due to incorrect version specification. Here is the conda error as well: ```bash conda env export InvalidVersionSpec: Invalid version '1.0.0<4.0.0': invalid character(s) ``` ## Fix suggestion Put a comma between the version limits which means replacing the line in setup.py file with the following: ```python "pyarrow>=1.0.0,<4.0.0", ``` ## Versions Paste the output of the following code: ```python - Datasets: 1.6.2 - Python: 3.7.10 (default, Feb 26 2021, 18:47:35) [GCC 7.3.0] - Platform: Linux-5.4.0-42-generic-x86_64-with-debian-buster-sid ```
2021-05-05T10:10:03Z
https://github.com/huggingface/datasets/issues/2316
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2316/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2315/comments
https://api.github.com/repos/huggingface/datasets/issues/2315/timeline
2021-05-10T16:36:50Z
null
null
MDExOlB1bGxSZXF1ZXN0NjMwMDIyMDYy
closed
[]
false
2,315
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Datasets cli improvements
https://api.github.com/repos/huggingface/datasets/issues/2315/events
null
https://api.github.com/repos/huggingface/datasets/issues/2315/labels{/name}
2021-05-04T18:55:11Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2315.diff", "html_url": "https://github.com/huggingface/datasets/pull/2315", "merged_at": "2021-05-10T16:36:50Z", "patch_url": "https://github.com/huggingface/datasets/pull/2315.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2315" }
875,742,200
[]
https://api.github.com/repos/huggingface/datasets/issues/2315
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This PR: * replaces the code from the `bug_report.md` that was used to get relevant system info with a dedicated command (a more elegant approach than copy-pasting the code IMO) * removes the `download` command (copied from the transformers repo?) * adds missing help messages to the cli commands
2021-05-10T16:36:51Z
https://github.com/huggingface/datasets/pull/2315
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2315/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2314/comments
https://api.github.com/repos/huggingface/datasets/issues/2314/timeline
2021-10-13T09:07:34Z
null
null
MDExOlB1bGxSZXF1ZXN0NjMwMDExODc4
closed
[]
false
2,314
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Minor refactor prepare_module
https://api.github.com/repos/huggingface/datasets/issues/2314/events
null
https://api.github.com/repos/huggingface/datasets/issues/2314/labels{/name}
2021-05-04T18:37:26Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2314.diff", "html_url": "https://github.com/huggingface/datasets/pull/2314", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2314.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2314" }
875,729,271
[]
https://api.github.com/repos/huggingface/datasets/issues/2314
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Start to refactor `prepare_module` to try to decouple functionality. This PR does: - extract function `_initialize_dynamic_modules_namespace_package` - extract function `_find_module_in_github_or_s3` - some renaming of variables - use of f-strings
2021-10-13T09:07:34Z
https://github.com/huggingface/datasets/pull/2314
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2314/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2313/comments
https://api.github.com/repos/huggingface/datasets/issues/2313/timeline
2021-05-07T09:31:42Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI5ODEwNTc4
closed
[]
false
2,313
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Remove unused head_hf_s3 function
https://api.github.com/repos/huggingface/datasets/issues/2313/events
null
https://api.github.com/repos/huggingface/datasets/issues/2313/labels{/name}
2021-05-04T13:42:06Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2313.diff", "html_url": "https://github.com/huggingface/datasets/pull/2313", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2313.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2313" }
875,475,367
[]
https://api.github.com/repos/huggingface/datasets/issues/2313
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Currently, the function `head_hf_s3` is not used: - neither its returned result is used - nor it raises any exception, as exceptions are catched and returned (not raised) This PR removes it.
2021-05-07T09:31:42Z
https://github.com/huggingface/datasets/pull/2313
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2313/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2312/comments
https://api.github.com/repos/huggingface/datasets/issues/2312/timeline
2021-05-04T13:43:12Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI5Nzc4NjUz
closed
[]
false
2,312
{ "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "events_url": "https://api.github.com/users/SBrandeis/events{/privacy}", "followers_url": "https://api.github.com/users/SBrandeis/followers", "following_url": "https://api.github.com/users/SBrandeis/following{/other_user}", "gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SBrandeis", "id": 33657802, "login": "SBrandeis", "node_id": "MDQ6VXNlcjMzNjU3ODAy", "organizations_url": "https://api.github.com/users/SBrandeis/orgs", "received_events_url": "https://api.github.com/users/SBrandeis/received_events", "repos_url": "https://api.github.com/users/SBrandeis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions", "type": "User", "url": "https://api.github.com/users/SBrandeis" }
Add rename_columnS method
https://api.github.com/repos/huggingface/datasets/issues/2312/events
null
https://api.github.com/repos/huggingface/datasets/issues/2312/labels{/name}
2021-05-04T12:57:53Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2312.diff", "html_url": "https://github.com/huggingface/datasets/pull/2312", "merged_at": "2021-05-04T13:43:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/2312.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2312" }
875,435,726
[]
https://api.github.com/repos/huggingface/datasets/issues/2312
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Cherry-picked from #2255
2021-05-04T13:43:13Z
https://github.com/huggingface/datasets/pull/2312
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2312/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2311/comments
https://api.github.com/repos/huggingface/datasets/issues/2311/timeline
2021-05-07T09:50:55Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI5NjQwNTMx
closed
[]
false
2,311
{ "avatar_url": "https://avatars.githubusercontent.com/u/7669893?v=4", "events_url": "https://api.github.com/users/cahya-wirawan/events{/privacy}", "followers_url": "https://api.github.com/users/cahya-wirawan/followers", "following_url": "https://api.github.com/users/cahya-wirawan/following{/other_user}", "gists_url": "https://api.github.com/users/cahya-wirawan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cahya-wirawan", "id": 7669893, "login": "cahya-wirawan", "node_id": "MDQ6VXNlcjc2Njk4OTM=", "organizations_url": "https://api.github.com/users/cahya-wirawan/orgs", "received_events_url": "https://api.github.com/users/cahya-wirawan/received_events", "repos_url": "https://api.github.com/users/cahya-wirawan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cahya-wirawan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cahya-wirawan/subscriptions", "type": "User", "url": "https://api.github.com/users/cahya-wirawan" }
Add SLR52, SLR53 and SLR54 to OpenSLR
https://api.github.com/repos/huggingface/datasets/issues/2311/events
null
https://api.github.com/repos/huggingface/datasets/issues/2311/labels{/name}
2021-05-04T09:08:03Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2311.diff", "html_url": "https://github.com/huggingface/datasets/pull/2311", "merged_at": "2021-05-07T09:50:55Z", "patch_url": "https://github.com/huggingface/datasets/pull/2311.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2311" }
875,262,208
[]
https://api.github.com/repos/huggingface/datasets/issues/2311
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Add large speech datasets for Sinhala, Bengali and Nepali.
2021-05-07T09:50:55Z
https://github.com/huggingface/datasets/pull/2311
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2311/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2310/comments
https://api.github.com/repos/huggingface/datasets/issues/2310/timeline
2022-07-06T15:19:58Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI5NTEwNTg5
closed
[]
false
2,310
{ "avatar_url": "https://avatars.githubusercontent.com/u/15029054?v=4", "events_url": "https://api.github.com/users/cryoff/events{/privacy}", "followers_url": "https://api.github.com/users/cryoff/followers", "following_url": "https://api.github.com/users/cryoff/following{/other_user}", "gists_url": "https://api.github.com/users/cryoff/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cryoff", "id": 15029054, "login": "cryoff", "node_id": "MDQ6VXNlcjE1MDI5MDU0", "organizations_url": "https://api.github.com/users/cryoff/orgs", "received_events_url": "https://api.github.com/users/cryoff/received_events", "repos_url": "https://api.github.com/users/cryoff/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cryoff/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cryoff/subscriptions", "type": "User", "url": "https://api.github.com/users/cryoff" }
Update README.md
https://api.github.com/repos/huggingface/datasets/issues/2310/events
null
https://api.github.com/repos/huggingface/datasets/issues/2310/labels{/name}
2021-05-04T04:38:01Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2310.diff", "html_url": "https://github.com/huggingface/datasets/pull/2310", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2310.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2310" }
875,096,051
[]
https://api.github.com/repos/huggingface/datasets/issues/2310
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Provides description of data instances and dataset features
2022-07-06T15:19:58Z
https://github.com/huggingface/datasets/pull/2310
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2310/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2309/comments
https://api.github.com/repos/huggingface/datasets/issues/2309/timeline
2021-05-03T16:01:17Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI5MTU4NjQx
closed
[]
false
2,309
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix conda release
https://api.github.com/repos/huggingface/datasets/issues/2309/events
null
https://api.github.com/repos/huggingface/datasets/issues/2309/labels{/name}
2021-05-03T14:52:59Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2309.diff", "html_url": "https://github.com/huggingface/datasets/pull/2309", "merged_at": "2021-05-03T16:01:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/2309.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2309" }
874,644,990
[]
https://api.github.com/repos/huggingface/datasets/issues/2309
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
There were a few issues with conda releases (they've been failing for a while now). To fix this I had to: - add the --single-version-externally-managed tag to the build stage (suggestion from [here](https://stackoverflow.com/a/64825075)) - set the python version of the conda build stage to 3.8 since 3.9 isn't supported - sync the evrsion requirement of `huggingface_hub` With these changes I'm working on uploading all missing versions until 1.6.2 to conda EDIT: I managed to build and upload all missing versions until 1.6.2 to conda :)
2021-05-03T16:01:17Z
https://github.com/huggingface/datasets/pull/2309
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2309/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2302/comments
https://api.github.com/repos/huggingface/datasets/issues/2302/timeline
2021-05-10T09:21:19Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI4NjIzMDQ3
closed
[]
false
2,302
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
Add SubjQA dataset
https://api.github.com/repos/huggingface/datasets/issues/2302/events
null
https://api.github.com/repos/huggingface/datasets/issues/2302/labels{/name}
2021-05-02T14:51:20Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2302.diff", "html_url": "https://github.com/huggingface/datasets/pull/2302", "merged_at": "2021-05-10T09:21:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/2302.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2302" }
873,961,435
[]
https://api.github.com/repos/huggingface/datasets/issues/2302
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Hello datasetters πŸ™‚! Here's an interesting dataset about extractive question-answering on _subjective_ product / restaurant reviews. It's quite challenging for models fine-tuned on SQuAD and provides a nice example of domain adaptation (i.e. fine-tuning a SQuAD model on this domain gives better performance). I found a bug in the start/end indices that I've proposed a fix for here: https://github.com/megagonlabs/SubjQA/pull/2 Unfortunately, the dataset creators are unresponsive, so for now I am using my fork as the source. Will update the URL if/when the creators respond.
2021-05-10T09:21:19Z
https://github.com/huggingface/datasets/pull/2302
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2302/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2301/comments
https://api.github.com/repos/huggingface/datasets/issues/2301/timeline
2021-05-03T15:17:34Z
null
completed
MDU6SXNzdWU4NzM5NDEyNjY=
closed
[]
null
2,301
{ "avatar_url": "https://avatars.githubusercontent.com/u/29076344?v=4", "events_url": "https://api.github.com/users/gchhablani/events{/privacy}", "followers_url": "https://api.github.com/users/gchhablani/followers", "following_url": "https://api.github.com/users/gchhablani/following{/other_user}", "gists_url": "https://api.github.com/users/gchhablani/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/gchhablani", "id": 29076344, "login": "gchhablani", "node_id": "MDQ6VXNlcjI5MDc2MzQ0", "organizations_url": "https://api.github.com/users/gchhablani/orgs", "received_events_url": "https://api.github.com/users/gchhablani/received_events", "repos_url": "https://api.github.com/users/gchhablani/repos", "site_admin": false, "starred_url": "https://api.github.com/users/gchhablani/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gchhablani/subscriptions", "type": "User", "url": "https://api.github.com/users/gchhablani" }
Unable to setup dev env on Windows
https://api.github.com/repos/huggingface/datasets/issues/2301/events
null
https://api.github.com/repos/huggingface/datasets/issues/2301/labels{/name}
2021-05-02T13:20:42Z
null
false
null
null
873,941,266
[]
https://api.github.com/repos/huggingface/datasets/issues/2301
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Hi I tried installing the `".[dev]"` version on Windows 10 after cloning. Here is the error I'm facing: ```bat (env) C:\testing\datasets>pip install -e ".[dev]" Obtaining file:///C:/testing/datasets Requirement already satisfied: numpy>=1.17 in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (1.19.5) Collecting pyarrow>=0.17.1 Using cached pyarrow-4.0.0-cp37-cp37m-win_amd64.whl (13.3 MB) Requirement already satisfied: dill in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (0.3.1.1) Collecting pandas Using cached pandas-1.2.4-cp37-cp37m-win_amd64.whl (9.1 MB) Requirement already satisfied: requests>=2.19.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (2.25.1) Requirement already satisfied: tqdm<4.50.0,>=4.27 in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (4.49.0) Requirement already satisfied: xxhash in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (2.0.2) Collecting multiprocess Using cached multiprocess-0.70.11.1-py37-none-any.whl (108 kB) Requirement already satisfied: fsspec in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (2021.4.0) Collecting huggingface_hub<0.1.0 Using cached huggingface_hub-0.0.8-py3-none-any.whl (34 kB) Requirement already satisfied: importlib_metadata in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (4.0.1) Requirement already satisfied: absl-py in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (0.12.0) Requirement already satisfied: pytest in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (6.2.3) Collecting pytest-xdist Using cached pytest_xdist-2.2.1-py3-none-any.whl (37 kB) Collecting apache-beam>=2.24.0 Using cached apache_beam-2.29.0-cp37-cp37m-win_amd64.whl (3.7 MB) Collecting elasticsearch Using cached elasticsearch-7.12.1-py2.py3-none-any.whl (339 kB) Requirement already satisfied: boto3==1.16.43 in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (1.16.43) Requirement already satisfied: botocore==1.19.43 in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (1.19.43) Collecting moto[s3]==1.3.16 Using cached moto-1.3.16-py2.py3-none-any.whl (879 kB) Collecting rarfile>=4.0 Using cached rarfile-4.0-py3-none-any.whl (28 kB) Collecting tensorflow>=2.3 Using cached tensorflow-2.4.1-cp37-cp37m-win_amd64.whl (370.7 MB) Requirement already satisfied: torch in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (1.8.1) Requirement already satisfied: transformers in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (4.5.1) Collecting bs4 Using cached bs4-0.0.1-py3-none-any.whl Collecting conllu Using cached conllu-4.4-py2.py3-none-any.whl (15 kB) Collecting langdetect Using cached langdetect-1.0.8-py3-none-any.whl Collecting lxml Using cached lxml-4.6.3-cp37-cp37m-win_amd64.whl (3.5 MB) Collecting mwparserfromhell Using cached mwparserfromhell-0.6-cp37-cp37m-win_amd64.whl (101 kB) Collecting nltk Using cached nltk-3.6.2-py3-none-any.whl (1.5 MB) Collecting openpyxl Using cached openpyxl-3.0.7-py2.py3-none-any.whl (243 kB) Collecting py7zr Using cached py7zr-0.15.2-py3-none-any.whl (66 kB) Collecting tldextract Using cached tldextract-3.1.0-py2.py3-none-any.whl (87 kB) Collecting zstandard Using cached zstandard-0.15.2-cp37-cp37m-win_amd64.whl (582 kB) Collecting bert_score>=0.3.6 Using cached bert_score-0.3.9-py3-none-any.whl (59 kB) Collecting rouge_score Using cached rouge_score-0.0.4-py2.py3-none-any.whl (22 kB) Collecting sacrebleu Using cached sacrebleu-1.5.1-py3-none-any.whl (54 kB) Requirement already satisfied: scipy in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (1.6.3) Collecting seqeval Using cached seqeval-1.2.2-py3-none-any.whl Collecting sklearn Using cached sklearn-0.0-py2.py3-none-any.whl Collecting jiwer Using cached jiwer-2.2.0-py3-none-any.whl (13 kB) Requirement already satisfied: toml>=0.10.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (0.10.2) Requirement already satisfied: requests_file>=1.5.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (1.5.1) Requirement already satisfied: texttable>=1.6.3 in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (1.6.3) Requirement already satisfied: s3fs>=0.4.2 in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (0.4.2) Requirement already satisfied: Werkzeug>=1.0.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from datasets==1.5.0.dev0) (1.0.1) Collecting black Using cached black-21.4b2-py3-none-any.whl (130 kB) Collecting isort Using cached isort-5.8.0-py3-none-any.whl (103 kB) Collecting flake8==3.7.9 Using cached flake8-3.7.9-py2.py3-none-any.whl (69 kB) Requirement already satisfied: jmespath<1.0.0,>=0.7.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from boto3==1.16.43->datasets==1.5.0.dev0) (0.10.0) Requirement already satisfied: s3transfer<0.4.0,>=0.3.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from boto3==1.16.43->datasets==1.5.0.dev0) (0.3.7) Requirement already satisfied: urllib3<1.27,>=1.25.4 in c:\programdata\anaconda3\envs\env\lib\site-packages (from botocore==1.19.43->datasets==1.5.0.dev0) (1.26.4) Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from botocore==1.19.43->datasets==1.5.0.dev0) (2.8.1) Collecting entrypoints<0.4.0,>=0.3.0 Using cached entrypoints-0.3-py2.py3-none-any.whl (11 kB) Collecting pyflakes<2.2.0,>=2.1.0 Using cached pyflakes-2.1.1-py2.py3-none-any.whl (59 kB) Collecting pycodestyle<2.6.0,>=2.5.0 Using cached pycodestyle-2.5.0-py2.py3-none-any.whl (51 kB) Collecting mccabe<0.7.0,>=0.6.0 Using cached mccabe-0.6.1-py2.py3-none-any.whl (8.6 kB) Requirement already satisfied: jsondiff>=1.1.2 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.3.0) Requirement already satisfied: pytz in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2021.1) Requirement already satisfied: mock in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (4.0.3) Requirement already satisfied: MarkupSafe<2.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.1.1) Requirement already satisfied: python-jose[cryptography]<4.0.0,>=3.1.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.2.0) Requirement already satisfied: aws-xray-sdk!=0.96,>=0.93 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.8.0) Requirement already satisfied: cryptography>=2.3.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.4.7) Requirement already satisfied: more-itertools in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (8.7.0) Requirement already satisfied: PyYAML>=5.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (5.4.1) Requirement already satisfied: boto>=2.36.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.49.0) Requirement already satisfied: idna<3,>=2.5 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.10) Requirement already satisfied: sshpubkeys>=3.1.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.3.1) Requirement already satisfied: responses>=0.9.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.13.3) Requirement already satisfied: xmltodict in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.12.0) Requirement already satisfied: setuptools in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (52.0.0.post20210125) Requirement already satisfied: Jinja2>=2.10.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.11.3) Requirement already satisfied: zipp in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.4.1) Requirement already satisfied: six>1.9 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.15.0) Requirement already satisfied: ecdsa<0.15 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.14.1) Requirement already satisfied: docker>=2.5.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (5.0.0) Requirement already satisfied: cfn-lint>=0.4.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.49.0) Requirement already satisfied: grpcio<2,>=1.29.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (1.32.0) Collecting hdfs<3.0.0,>=2.1.0 Using cached hdfs-2.6.0-py3-none-any.whl (33 kB) Collecting pyarrow>=0.17.1 Using cached pyarrow-3.0.0-cp37-cp37m-win_amd64.whl (12.6 MB) Collecting fastavro<2,>=0.21.4 Using cached fastavro-1.4.0-cp37-cp37m-win_amd64.whl (394 kB) Requirement already satisfied: httplib2<0.18.0,>=0.8 in c:\programdata\anaconda3\envs\env\lib\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.17.4) Collecting pymongo<4.0.0,>=3.8.0 Using cached pymongo-3.11.3-cp37-cp37m-win_amd64.whl (382 kB) Collecting crcmod<2.0,>=1.7 Using cached crcmod-1.7-py3-none-any.whl Collecting avro-python3!=1.9.2,<1.10.0,>=1.8.1 Using cached avro_python3-1.9.2.1-py3-none-any.whl Requirement already satisfied: typing-extensions<3.8.0,>=3.7.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (3.7.4.3) Requirement already satisfied: future<1.0.0,>=0.18.2 in c:\programdata\anaconda3\envs\env\lib\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.18.2) Collecting oauth2client<5,>=2.0.1 Using cached oauth2client-4.1.3-py2.py3-none-any.whl (98 kB) Collecting pydot<2,>=1.2.0 Using cached pydot-1.4.2-py2.py3-none-any.whl (21 kB) Requirement already satisfied: protobuf<4,>=3.12.2 in c:\programdata\anaconda3\envs\env\lib\site-packages (from apache-beam>=2.24.0->datasets==1.5.0.dev0) (3.15.8) Requirement already satisfied: wrapt in c:\programdata\anaconda3\envs\env\lib\site-packages (from aws-xray-sdk!=0.96,>=0.93->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.12.1) Collecting matplotlib Using cached matplotlib-3.4.1-cp37-cp37m-win_amd64.whl (7.1 MB) Requirement already satisfied: junit-xml~=1.9 in c:\programdata\anaconda3\envs\env\lib\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.9) Requirement already satisfied: jsonpatch in c:\programdata\anaconda3\envs\env\lib\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.32) Requirement already satisfied: jsonschema~=3.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (3.2.0) Requirement already satisfied: networkx~=2.4 in c:\programdata\anaconda3\envs\env\lib\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.5.1) Requirement already satisfied: aws-sam-translator>=1.35.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.35.0) Requirement already satisfied: cffi>=1.12 in c:\programdata\anaconda3\envs\env\lib\site-packages (from cryptography>=2.3.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (1.14.5) Requirement already satisfied: pycparser in c:\programdata\anaconda3\envs\env\lib\site-packages (from cffi>=1.12->cryptography>=2.3.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.20) Requirement already satisfied: pywin32==227 in c:\programdata\anaconda3\envs\env\lib\site-packages (from docker>=2.5.1->moto[s3]==1.3.16->datasets==1.5.0.dev0) (227) Requirement already satisfied: websocket-client>=0.32.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from docker>=2.5.1->moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.58.0) Requirement already satisfied: docopt in c:\programdata\anaconda3\envs\env\lib\site-packages (from hdfs<3.0.0,>=2.1.0->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.6.2) Requirement already satisfied: filelock in c:\programdata\anaconda3\envs\env\lib\site-packages (from huggingface_hub<0.1.0->datasets==1.5.0.dev0) (3.0.12) Requirement already satisfied: pyrsistent>=0.14.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from jsonschema~=3.0->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (0.17.3) Requirement already satisfied: attrs>=17.4.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from jsonschema~=3.0->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (20.3.0) Requirement already satisfied: decorator<5,>=4.3 in c:\programdata\anaconda3\envs\env\lib\site-packages (from networkx~=2.4->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (4.4.2) Requirement already satisfied: rsa>=3.1.4 in c:\programdata\anaconda3\envs\env\lib\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (4.7.2) Requirement already satisfied: pyasn1-modules>=0.0.5 in c:\programdata\anaconda3\envs\env\lib\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.2.8) Requirement already satisfied: pyasn1>=0.1.7 in c:\programdata\anaconda3\envs\env\lib\site-packages (from oauth2client<5,>=2.0.1->apache-beam>=2.24.0->datasets==1.5.0.dev0) (0.4.8) Requirement already satisfied: pyparsing>=2.1.4 in c:\programdata\anaconda3\envs\env\lib\site-packages (from pydot<2,>=1.2.0->apache-beam>=2.24.0->datasets==1.5.0.dev0) (2.4.7) Requirement already satisfied: certifi>=2017.4.17 in c:\programdata\anaconda3\envs\env\lib\site-packages (from requests>=2.19.0->datasets==1.5.0.dev0) (2020.12.5) Requirement already satisfied: chardet<5,>=3.0.2 in c:\programdata\anaconda3\envs\env\lib\site-packages (from requests>=2.19.0->datasets==1.5.0.dev0) (4.0.0) Collecting keras-preprocessing~=1.1.2 Using cached Keras_Preprocessing-1.1.2-py2.py3-none-any.whl (42 kB) Requirement already satisfied: termcolor~=1.1.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (1.1.0) Requirement already satisfied: tensorboard~=2.4 in c:\programdata\anaconda3\envs\env\lib\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (2.5.0) Requirement already satisfied: wheel~=0.35 in c:\programdata\anaconda3\envs\env\lib\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (0.36.2) Collecting opt-einsum~=3.3.0 Using cached opt_einsum-3.3.0-py3-none-any.whl (65 kB) Collecting gast==0.3.3 Using cached gast-0.3.3-py2.py3-none-any.whl (9.7 kB) Collecting google-pasta~=0.2 Using cached google_pasta-0.2.0-py3-none-any.whl (57 kB) Requirement already satisfied: tensorflow-estimator<2.5.0,>=2.4.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from tensorflow>=2.3->datasets==1.5.0.dev0) (2.4.0) Collecting astunparse~=1.6.3 Using cached astunparse-1.6.3-py2.py3-none-any.whl (12 kB) Collecting flatbuffers~=1.12.0 Using cached flatbuffers-1.12-py2.py3-none-any.whl (15 kB) Collecting h5py~=2.10.0 Using cached h5py-2.10.0-cp37-cp37m-win_amd64.whl (2.5 MB) Requirement already satisfied: markdown>=2.6.8 in c:\programdata\anaconda3\envs\env\lib\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (3.3.4) Requirement already satisfied: tensorboard-plugin-wit>=1.6.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.8.0) Requirement already satisfied: google-auth-oauthlib<0.5,>=0.4.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (0.4.4) Requirement already satisfied: tensorboard-data-server<0.7.0,>=0.6.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (0.6.0) Requirement already satisfied: google-auth<2,>=1.6.3 in c:\programdata\anaconda3\envs\env\lib\site-packages (from tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.30.0) Requirement already satisfied: cachetools<5.0,>=2.0.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from google-auth<2,>=1.6.3->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (4.2.2) Requirement already satisfied: requests-oauthlib>=0.7.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from google-auth-oauthlib<0.5,>=0.4.1->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (1.3.0) Requirement already satisfied: oauthlib>=3.0.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from requests-oauthlib>=0.7.0->google-auth-oauthlib<0.5,>=0.4.1->tensorboard~=2.4->tensorflow>=2.3->datasets==1.5.0.dev0) (3.1.0) Requirement already satisfied: regex!=2019.12.17 in c:\programdata\anaconda3\envs\env\lib\site-packages (from transformers->datasets==1.5.0.dev0) (2021.4.4) Requirement already satisfied: tokenizers<0.11,>=0.10.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from transformers->datasets==1.5.0.dev0) (0.10.2) Requirement already satisfied: sacremoses in c:\programdata\anaconda3\envs\env\lib\site-packages (from transformers->datasets==1.5.0.dev0) (0.0.45) Requirement already satisfied: packaging in c:\programdata\anaconda3\envs\env\lib\site-packages (from transformers->datasets==1.5.0.dev0) (20.9) Collecting pathspec<1,>=0.8.1 Using cached pathspec-0.8.1-py2.py3-none-any.whl (28 kB) Requirement already satisfied: click>=7.1.2 in c:\programdata\anaconda3\envs\env\lib\site-packages (from black->datasets==1.5.0.dev0) (7.1.2) Collecting appdirs Using cached appdirs-1.4.4-py2.py3-none-any.whl (9.6 kB) Collecting mypy-extensions>=0.4.3 Using cached mypy_extensions-0.4.3-py2.py3-none-any.whl (4.5 kB) Requirement already satisfied: typed-ast>=1.4.2 in c:\programdata\anaconda3\envs\env\lib\site-packages (from black->datasets==1.5.0.dev0) (1.4.3) Collecting beautifulsoup4 Using cached beautifulsoup4-4.9.3-py3-none-any.whl (115 kB) Requirement already satisfied: soupsieve>1.2 in c:\programdata\anaconda3\envs\env\lib\site-packages (from beautifulsoup4->bs4->datasets==1.5.0.dev0) (2.2.1) Collecting python-Levenshtein Using cached python-Levenshtein-0.12.2.tar.gz (50 kB) Requirement already satisfied: jsonpointer>=1.9 in c:\programdata\anaconda3\envs\env\lib\site-packages (from jsonpatch->cfn-lint>=0.4.0->moto[s3]==1.3.16->datasets==1.5.0.dev0) (2.1) Requirement already satisfied: pillow>=6.2.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (8.2.0) Requirement already satisfied: cycler>=0.10 in c:\programdata\anaconda3\envs\env\lib\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (0.10.0) Requirement already satisfied: kiwisolver>=1.0.1 in c:\programdata\anaconda3\envs\env\lib\site-packages (from matplotlib->bert_score>=0.3.6->datasets==1.5.0.dev0) (1.3.1) Collecting multiprocess Using cached multiprocess-0.70.11-py3-none-any.whl (98 kB) Using cached multiprocess-0.70.10.zip (2.4 MB) Using cached multiprocess-0.70.9-py3-none-any.whl Requirement already satisfied: joblib in c:\programdata\anaconda3\envs\env\lib\site-packages (from nltk->datasets==1.5.0.dev0) (1.0.1) Collecting et-xmlfile Using cached et_xmlfile-1.1.0-py3-none-any.whl (4.7 kB) Requirement already satisfied: pyzstd<0.15.0,>=0.14.4 in c:\programdata\anaconda3\envs\env\lib\site-packages (from py7zr->datasets==1.5.0.dev0) (0.14.4) Collecting pyppmd<0.13.0,>=0.12.1 Using cached pyppmd-0.12.1-cp37-cp37m-win_amd64.whl (32 kB) Collecting pycryptodome>=3.6.6 Using cached pycryptodome-3.10.1-cp35-abi3-win_amd64.whl (1.6 MB) Collecting bcj-cffi<0.6.0,>=0.5.1 Using cached bcj_cffi-0.5.1-cp37-cp37m-win_amd64.whl (21 kB) Collecting multivolumefile<0.3.0,>=0.2.0 Using cached multivolumefile-0.2.3-py3-none-any.whl (17 kB) Requirement already satisfied: iniconfig in c:\programdata\anaconda3\envs\env\lib\site-packages (from pytest->datasets==1.5.0.dev0) (1.1.1) Requirement already satisfied: py>=1.8.2 in c:\programdata\anaconda3\envs\env\lib\site-packages (from pytest->datasets==1.5.0.dev0) (1.10.0) Requirement already satisfied: pluggy<1.0.0a1,>=0.12 in c:\programdata\anaconda3\envs\env\lib\site-packages (from pytest->datasets==1.5.0.dev0) (0.13.1) Requirement already satisfied: atomicwrites>=1.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from pytest->datasets==1.5.0.dev0) (1.4.0) Requirement already satisfied: colorama in c:\programdata\anaconda3\envs\env\lib\site-packages (from pytest->datasets==1.5.0.dev0) (0.4.4) Collecting pytest-forked Using cached pytest_forked-1.3.0-py2.py3-none-any.whl (4.7 kB) Collecting execnet>=1.1 Using cached execnet-1.8.0-py2.py3-none-any.whl (39 kB) Requirement already satisfied: apipkg>=1.4 in c:\programdata\anaconda3\envs\env\lib\site-packages (from execnet>=1.1->pytest-xdist->datasets==1.5.0.dev0) (1.5) Collecting portalocker==2.0.0 Using cached portalocker-2.0.0-py2.py3-none-any.whl (11 kB) Requirement already satisfied: scikit-learn>=0.21.3 in c:\programdata\anaconda3\envs\env\lib\site-packages (from seqeval->datasets==1.5.0.dev0) (0.24.2) Requirement already satisfied: threadpoolctl>=2.0.0 in c:\programdata\anaconda3\envs\env\lib\site-packages (from scikit-learn>=0.21.3->seqeval->datasets==1.5.0.dev0) (2.1.0) Building wheels for collected packages: python-Levenshtein Building wheel for python-Levenshtein (setup.py) ... error ERROR: Command errored out with exit status 1: command: 'C:\ProgramData\Anaconda3\envs\env\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\setup.py'"'"'; __file__='"'"'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(__file__);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' bdist_wheel -d 'C:\Users\VKC~1\AppData\Local\Temp\pip-wheel-8jh7fm18' cwd: C:\Users\VKC~1\AppData\Local\Temp\pip-install-ynt_dbm4\python-levenshtein_c02e7e6f9def4629a475349654670ae9\ Complete output (27 lines): running bdist_wheel running build running build_py creating build creating build\lib.win-amd64-3.7 creating build\lib.win-amd64-3.7\Levenshtein copying Levenshtein\StringMatcher.py -> build\lib.win-amd64-3.7\Levenshtein copying Levenshtein\__init__.py -> build\lib.win-amd64-3.7\Levenshtein running egg_info writing python_Levenshtein.egg-info\PKG-INFO writing dependency_links to python_Levenshtein.egg-info\dependency_links.txt writing entry points to python_Levenshtein.egg-info\entry_points.txt writing namespace_packages to python_Levenshtein.egg-info\namespace_packages.txt writing requirements to python_Levenshtein.egg-info\requires.txt writing top-level names to python_Levenshtein.egg-info\top_level.txt reading manifest file 'python_Levenshtein.egg-info\SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no previously-included files matching '*pyc' found anywhere in distribution warning: no previously-included files matching '*so' found anywhere in distribution warning: no previously-included files matching '.project' found anywhere in distribution warning: no previously-included files matching '.pydevproject' found anywhere in distribution writing manifest file 'python_Levenshtein.egg-info\SOURCES.txt' copying Levenshtein\_levenshtein.c -> build\lib.win-amd64-3.7\Levenshtein copying Levenshtein\_levenshtein.h -> build\lib.win-amd64-3.7\Levenshtein running build_ext building 'Levenshtein._levenshtein' extension error: Microsoft Visual C++ 14.0 or greater is required. Get it with "Microsoft C++ Build Tools": https://visualstudio.microsoft.com/visual-cpp-build-tools/ ---------------------------------------- ERROR: Failed building wheel for python-Levenshtein Running setup.py clean for python-Levenshtein Failed to build python-Levenshtein Installing collected packages: python-Levenshtein, pytest-forked, pyppmd, pymongo, pyflakes, pydot, pycryptodome, pycodestyle, pyarrow, portalocker, pathspec, pandas, opt-einsum, oauth2client, nltk, mypy-extensions, multivolumefile, multiprocess, moto, mccabe, matplotlib, keras-preprocessing, huggingface-hub, hdfs, h5py, google-pasta, gast, flatbuffers, fastavro, execnet, et-xmlfile, entrypoints, crcmod, beautifulsoup4, bcj-cffi, avro-python3, astunparse, appdirs, zstandard, tldextract, tensorflow, sklearn, seqeval, sacrebleu, rouge-score, rarfile, pytest-xdist, py7zr, openpyxl, mwparserfromhell, lxml, langdetect, jiwer, isort, flake8, elasticsearch, datasets, conllu, bs4, black, bert-score, apache-beam Running setup.py install for python-Levenshtein ... error ERROR: Command errored out with exit status 1: command: 'C:\ProgramData\Anaconda3\envs\env\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\setup.py'"'"'; __file__='"'"'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(__file__);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' install --record 'C:\Users\VKC~1\AppData\Local\Temp\pip-record-v7l7zitb\install-record.txt' --single-version-externally-managed --compile --install-headers 'C:\ProgramData\Anaconda3\envs\env\Include\python-Levenshtein' cwd: C:\Users\VKC~1\AppData\Local\Temp\pip-install-ynt_dbm4\python-levenshtein_c02e7e6f9def4629a475349654670ae9\ Complete output (27 lines): running install running build running build_py creating build creating build\lib.win-amd64-3.7 creating build\lib.win-amd64-3.7\Levenshtein copying Levenshtein\StringMatcher.py -> build\lib.win-amd64-3.7\Levenshtein copying Levenshtein\__init__.py -> build\lib.win-amd64-3.7\Levenshtein running egg_info writing python_Levenshtein.egg-info\PKG-INFO writing dependency_links to python_Levenshtein.egg-info\dependency_links.txt writing entry points to python_Levenshtein.egg-info\entry_points.txt writing namespace_packages to python_Levenshtein.egg-info\namespace_packages.txt writing requirements to python_Levenshtein.egg-info\requires.txt writing top-level names to python_Levenshtein.egg-info\top_level.txt reading manifest file 'python_Levenshtein.egg-info\SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no previously-included files matching '*pyc' found anywhere in distribution warning: no previously-included files matching '*so' found anywhere in distribution warning: no previously-included files matching '.project' found anywhere in distribution warning: no previously-included files matching '.pydevproject' found anywhere in distribution writing manifest file 'python_Levenshtein.egg-info\SOURCES.txt' copying Levenshtein\_levenshtein.c -> build\lib.win-amd64-3.7\Levenshtein copying Levenshtein\_levenshtein.h -> build\lib.win-amd64-3.7\Levenshtein running build_ext building 'Levenshtein._levenshtein' extension error: Microsoft Visual C++ 14.0 or greater is required. Get it with "Microsoft C++ Build Tools": https://visualstudio.microsoft.com/visual-cpp-build-tools/ ---------------------------------------- ERROR: Command errored out with exit status 1: 'C:\ProgramData\Anaconda3\envs\env\python.exe' -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\setup.py'"'"'; __file__='"'"'C:\\Users\\VKC~1\\AppData\\Local\\Temp\\pip-install-ynt_dbm4\\python-levenshtein_c02e7e6f9def4629a475349654670ae9\\setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(__file__);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' install --record 'C:\Users\VKC~1\AppData\Local\Temp\pip-record-v7l7zitb\install-record.txt' --single-version-externally-managed --compile --install-headers 'C:\ProgramData\Anaconda3\envs\env\Include\python-Levenshtein' Check the logs for full command output. ``` Here are conda and python versions: ```bat (env) C:\testing\datasets>conda --version conda 4.9.2 (env) C:\testing\datasets>python --version Python 3.7.10 ``` Please help me out. Thanks.
2021-05-03T15:18:01Z
https://github.com/huggingface/datasets/issues/2301
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2301/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2300/comments
https://api.github.com/repos/huggingface/datasets/issues/2300/timeline
2023-02-28T17:43:51Z
null
completed
MDU6SXNzdWU4NzM5MjgxNjk=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" } ]
null
2,300
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
Add VoxPopuli
https://api.github.com/repos/huggingface/datasets/issues/2300/events
null
https://api.github.com/repos/huggingface/datasets/issues/2300/labels{/name}
2021-05-02T12:17:40Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
null
873,928,169
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" }, { "color": "d93f0b", "default": false, "description": "", "id": 2725241052, "name": "speech", "node_id": "MDU6TGFiZWwyNzI1MjQxMDUy", "url": "https://api.github.com/repos/huggingface/datasets/labels/speech" } ]
https://api.github.com/repos/huggingface/datasets/issues/2300
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Adding a Dataset - **Name:** Voxpopuli - **Description:** VoxPopuli is raw data is collected from 2009-2020 European Parliament event recordings - **Paper:** https://arxiv.org/abs/2101.00390 - **Data:** https://github.com/facebookresearch/voxpopuli - **Motivation:** biggest unlabeled speech dataset **Note**: Since the dataset is so huge, we should only add the config `10k` in the beginning. Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
2023-02-28T17:43:52Z
https://github.com/huggingface/datasets/issues/2300
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2300/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2299/comments
https://api.github.com/repos/huggingface/datasets/issues/2299/timeline
2021-05-03T08:17:38Z
null
completed
MDU6SXNzdWU4NzM5MTQ3MTc=
closed
[]
null
2,299
{ "avatar_url": "https://avatars.githubusercontent.com/u/82856229?v=4", "events_url": "https://api.github.com/users/Jasonbuchanan1983/events{/privacy}", "followers_url": "https://api.github.com/users/Jasonbuchanan1983/followers", "following_url": "https://api.github.com/users/Jasonbuchanan1983/following{/other_user}", "gists_url": "https://api.github.com/users/Jasonbuchanan1983/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Jasonbuchanan1983", "id": 82856229, "login": "Jasonbuchanan1983", "node_id": "MDQ6VXNlcjgyODU2MjI5", "organizations_url": "https://api.github.com/users/Jasonbuchanan1983/orgs", "received_events_url": "https://api.github.com/users/Jasonbuchanan1983/received_events", "repos_url": "https://api.github.com/users/Jasonbuchanan1983/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Jasonbuchanan1983/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Jasonbuchanan1983/subscriptions", "type": "User", "url": "https://api.github.com/users/Jasonbuchanan1983" }
My iPhone
https://api.github.com/repos/huggingface/datasets/issues/2299/events
null
https://api.github.com/repos/huggingface/datasets/issues/2299/labels{/name}
2021-05-02T11:11:11Z
null
false
null
null
873,914,717
[]
https://api.github.com/repos/huggingface/datasets/issues/2299
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Adding a Dataset - **Name:** *name of the dataset* - **Description:** *short description of the dataset (or link to social media or blog post)* - **Paper:** *link to the dataset paper if available* - **Data:** *link to the Github repository or current dataset location* - **Motivation:** *what are some good reasons to have this dataset* Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
2021-07-23T09:24:16Z
https://github.com/huggingface/datasets/issues/2299
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2299/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2298/comments
https://api.github.com/repos/huggingface/datasets/issues/2298/timeline
2021-05-03T13:54:53Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI4NDk2NjM2
closed
[]
false
2,298
{ "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "events_url": "https://api.github.com/users/TevenLeScao/events{/privacy}", "followers_url": "https://api.github.com/users/TevenLeScao/followers", "following_url": "https://api.github.com/users/TevenLeScao/following{/other_user}", "gists_url": "https://api.github.com/users/TevenLeScao/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/TevenLeScao", "id": 26709476, "login": "TevenLeScao", "node_id": "MDQ6VXNlcjI2NzA5NDc2", "organizations_url": "https://api.github.com/users/TevenLeScao/orgs", "received_events_url": "https://api.github.com/users/TevenLeScao/received_events", "repos_url": "https://api.github.com/users/TevenLeScao/repos", "site_admin": false, "starred_url": "https://api.github.com/users/TevenLeScao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TevenLeScao/subscriptions", "type": "User", "url": "https://api.github.com/users/TevenLeScao" }
Mapping in the distributed setting
https://api.github.com/repos/huggingface/datasets/issues/2298/events
null
https://api.github.com/repos/huggingface/datasets/issues/2298/labels{/name}
2021-05-01T21:23:05Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2298.diff", "html_url": "https://github.com/huggingface/datasets/pull/2298", "merged_at": "2021-05-03T13:54:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/2298.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2298" }
873,771,942
[]
https://api.github.com/repos/huggingface/datasets/issues/2298
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
The barrier trick for distributed mapping as discussed on Thursday with @lhoestq
2021-05-03T13:54:53Z
https://github.com/huggingface/datasets/pull/2298
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2298/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2296/comments
https://api.github.com/repos/huggingface/datasets/issues/2296/timeline
2021-05-03T08:17:31Z
null
completed
MDU6SXNzdWU4NzI5NzQ5MDc=
closed
[]
null
2,296
{ "avatar_url": "https://avatars.githubusercontent.com/u/82880142?v=4", "events_url": "https://api.github.com/users/zinnyi/events{/privacy}", "followers_url": "https://api.github.com/users/zinnyi/followers", "following_url": "https://api.github.com/users/zinnyi/following{/other_user}", "gists_url": "https://api.github.com/users/zinnyi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zinnyi", "id": 82880142, "login": "zinnyi", "node_id": "MDQ6VXNlcjgyODgwMTQy", "organizations_url": "https://api.github.com/users/zinnyi/orgs", "received_events_url": "https://api.github.com/users/zinnyi/received_events", "repos_url": "https://api.github.com/users/zinnyi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zinnyi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zinnyi/subscriptions", "type": "User", "url": "https://api.github.com/users/zinnyi" }
1
https://api.github.com/repos/huggingface/datasets/issues/2296/events
null
https://api.github.com/repos/huggingface/datasets/issues/2296/labels{/name}
2021-04-30T17:53:49Z
null
false
null
null
872,974,907
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
https://api.github.com/repos/huggingface/datasets/issues/2296
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Adding a Dataset - **Name:** *name of the dataset* - **Description:** *short description of the dataset (or link to social media or blog post)* - **Paper:** *link to the dataset paper if available* - **Data:** *link to the Github repository or current dataset location* - **Motivation:** *what are some good reasons to have this dataset* Instructions to add a new dataset can be found [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md).
2021-05-03T08:17:31Z
https://github.com/huggingface/datasets/issues/2296
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2296/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2295/comments
https://api.github.com/repos/huggingface/datasets/issues/2295/timeline
2021-07-08T08:11:49Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI3NzY0NDk3
closed
[]
false
2,295
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Create ExtractManager
https://api.github.com/repos/huggingface/datasets/issues/2295/events
{ "closed_at": "2021-07-21T15:36:49Z", "closed_issues": 29, "created_at": "2021-06-08T18:48:33Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-08-05T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/6", "id": 6836458, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/6/labels", "node_id": "MDk6TWlsZXN0b25lNjgzNjQ1OA==", "number": 6, "open_issues": 0, "state": "closed", "title": "1.10", "updated_at": "2021-07-21T15:36:49Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/6" }
https://api.github.com/repos/huggingface/datasets/issues/2295/labels{/name}
2021-04-30T17:13:34Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2295.diff", "html_url": "https://github.com/huggingface/datasets/pull/2295", "merged_at": "2021-07-08T08:11:49Z", "patch_url": "https://github.com/huggingface/datasets/pull/2295.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2295" }
872,902,867
[ { "color": "B67A40", "default": false, "description": "Restructuring existing code without changing its external behavior", "id": 2851292821, "name": "refactoring", "node_id": "MDU6TGFiZWwyODUxMjkyODIx", "url": "https://api.github.com/repos/huggingface/datasets/labels/refactoring" } ]
https://api.github.com/repos/huggingface/datasets/issues/2295
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Perform refactoring to decouple extract functionality.
2021-07-12T14:12:03Z
https://github.com/huggingface/datasets/pull/2295
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2295/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2294/comments
https://api.github.com/repos/huggingface/datasets/issues/2294/timeline
null
null
null
MDU6SXNzdWU4NzIxMzYwNzU=
open
[]
null
2,294
{ "avatar_url": "https://avatars.githubusercontent.com/u/31714566?v=4", "events_url": "https://api.github.com/users/VerdureChen/events{/privacy}", "followers_url": "https://api.github.com/users/VerdureChen/followers", "following_url": "https://api.github.com/users/VerdureChen/following{/other_user}", "gists_url": "https://api.github.com/users/VerdureChen/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/VerdureChen", "id": 31714566, "login": "VerdureChen", "node_id": "MDQ6VXNlcjMxNzE0NTY2", "organizations_url": "https://api.github.com/users/VerdureChen/orgs", "received_events_url": "https://api.github.com/users/VerdureChen/received_events", "repos_url": "https://api.github.com/users/VerdureChen/repos", "site_admin": false, "starred_url": "https://api.github.com/users/VerdureChen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/VerdureChen/subscriptions", "type": "User", "url": "https://api.github.com/users/VerdureChen" }
Slow #0 when using map to tokenize.
https://api.github.com/repos/huggingface/datasets/issues/2294/events
null
https://api.github.com/repos/huggingface/datasets/issues/2294/labels{/name}
2021-04-30T08:00:33Z
null
false
null
null
872,136,075
[]
https://api.github.com/repos/huggingface/datasets/issues/2294
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Hi, _datasets_ is really amazing! I am following [run_mlm_no_trainer.py](url) to pre-train BERT, and it uses `tokenized_datasets = raw_datasets.map( tokenize_function, batched=True, num_proc=args.preprocessing_num_workers, remove_columns=column_names, load_from_cache_file=not args.overwrite_cache, )` to tokenize by multiprocessing. However, I have found that when `num_proc`>1,the process _#0_ is much slower than others. It looks like this: ![image](https://user-images.githubusercontent.com/31714566/116665555-81246280-a9cc-11eb-8a37-6e608ab310d0.png) It takes more than 12 hours for #0, while others just about half an hour. Could anyone tell me it is normal or not, and is there any methods to speed up it?
2021-05-04T11:00:11Z
https://github.com/huggingface/datasets/issues/2294
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2294/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2293/comments
https://api.github.com/repos/huggingface/datasets/issues/2293/timeline
2021-04-30T06:54:25Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI3MDQzNzQ3
closed
[]
false
2,293
{ "avatar_url": "https://avatars.githubusercontent.com/u/52530809?v=4", "events_url": "https://api.github.com/users/BobbyManion/events{/privacy}", "followers_url": "https://api.github.com/users/BobbyManion/followers", "following_url": "https://api.github.com/users/BobbyManion/following{/other_user}", "gists_url": "https://api.github.com/users/BobbyManion/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/BobbyManion", "id": 52530809, "login": "BobbyManion", "node_id": "MDQ6VXNlcjUyNTMwODA5", "organizations_url": "https://api.github.com/users/BobbyManion/orgs", "received_events_url": "https://api.github.com/users/BobbyManion/received_events", "repos_url": "https://api.github.com/users/BobbyManion/repos", "site_admin": false, "starred_url": "https://api.github.com/users/BobbyManion/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BobbyManion/subscriptions", "type": "User", "url": "https://api.github.com/users/BobbyManion" }
imdb dataset from Don't Stop Pretraining Paper
https://api.github.com/repos/huggingface/datasets/issues/2293/events
null
https://api.github.com/repos/huggingface/datasets/issues/2293/labels{/name}
2021-04-30T06:40:48Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2293.diff", "html_url": "https://github.com/huggingface/datasets/pull/2293", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2293.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2293" }
872,079,385
[]
https://api.github.com/repos/huggingface/datasets/issues/2293
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
2021-04-30T06:54:25Z
https://github.com/huggingface/datasets/pull/2293
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2293/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2292/comments
https://api.github.com/repos/huggingface/datasets/issues/2292/timeline
2021-04-30T13:03:12Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI2MjgzNTYy
closed
[]
false
2,292
{ "avatar_url": "https://avatars.githubusercontent.com/u/32505743?v=4", "events_url": "https://api.github.com/users/laksh9950/events{/privacy}", "followers_url": "https://api.github.com/users/laksh9950/followers", "following_url": "https://api.github.com/users/laksh9950/following{/other_user}", "gists_url": "https://api.github.com/users/laksh9950/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/laksh9950", "id": 32505743, "login": "laksh9950", "node_id": "MDQ6VXNlcjMyNTA1NzQz", "organizations_url": "https://api.github.com/users/laksh9950/orgs", "received_events_url": "https://api.github.com/users/laksh9950/received_events", "repos_url": "https://api.github.com/users/laksh9950/repos", "site_admin": false, "starred_url": "https://api.github.com/users/laksh9950/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/laksh9950/subscriptions", "type": "User", "url": "https://api.github.com/users/laksh9950" }
Fixed typo seperate->separate
https://api.github.com/repos/huggingface/datasets/issues/2292/events
null
https://api.github.com/repos/huggingface/datasets/issues/2292/labels{/name}
2021-04-29T16:40:53Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2292.diff", "html_url": "https://github.com/huggingface/datasets/pull/2292", "merged_at": "2021-04-30T13:03:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/2292.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2292" }
871,230,183
[]
https://api.github.com/repos/huggingface/datasets/issues/2292
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
2021-04-30T13:29:18Z
https://github.com/huggingface/datasets/pull/2292
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2292/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2291/comments
https://api.github.com/repos/huggingface/datasets/issues/2291/timeline
2021-04-29T16:34:34Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI2MjcyNzE5
closed
[]
false
2,291
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Don't copy recordbatches in memory during a table deepcopy
https://api.github.com/repos/huggingface/datasets/issues/2291/events
null
https://api.github.com/repos/huggingface/datasets/issues/2291/labels{/name}
2021-04-29T16:26:05Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2291.diff", "html_url": "https://github.com/huggingface/datasets/pull/2291", "merged_at": "2021-04-29T16:34:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/2291.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2291" }
871,216,757
[]
https://api.github.com/repos/huggingface/datasets/issues/2291
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix issue #2276 and hopefully #2134 The recordbatches of the `IndexedTableMixin` used to speed up queries to the table were copied in memory during a table deepcopy. This resulted in `concatenate_datasets`, `load_from_disk` and other methods to always bring the data in memory. I fixed the copy similarly to #2287 and updated the test to make sure it doesn't happen again (added a test for deepcopy + make sure that the immutable arrow objects are passed to the copied table without being copied). The issue was not caught by our tests because the total allocated bytes value in PyArrow isn't updated when deepcopying recordbatches: the copy in memory wasn't detected. This behavior looks like a bug in PyArrow, I'll open a ticket on JIRA. Thanks @samsontmr , @TaskManager91 and @mariosasko for the help
2021-04-29T16:34:35Z
https://github.com/huggingface/datasets/pull/2291
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 1, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2291/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2290/comments
https://api.github.com/repos/huggingface/datasets/issues/2290/timeline
2021-05-06T17:25:25Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI2MjEyNTIz
closed
[]
false
2,290
{ "avatar_url": "https://avatars.githubusercontent.com/u/54144149?v=4", "events_url": "https://api.github.com/users/phiwi/events{/privacy}", "followers_url": "https://api.github.com/users/phiwi/followers", "following_url": "https://api.github.com/users/phiwi/following{/other_user}", "gists_url": "https://api.github.com/users/phiwi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/phiwi", "id": 54144149, "login": "phiwi", "node_id": "MDQ6VXNlcjU0MTQ0MTQ5", "organizations_url": "https://api.github.com/users/phiwi/orgs", "received_events_url": "https://api.github.com/users/phiwi/received_events", "repos_url": "https://api.github.com/users/phiwi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/phiwi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/phiwi/subscriptions", "type": "User", "url": "https://api.github.com/users/phiwi" }
Bbaw egyptian
https://api.github.com/repos/huggingface/datasets/issues/2290/events
null
https://api.github.com/repos/huggingface/datasets/issues/2290/labels{/name}
2021-04-29T15:27:58Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2290.diff", "html_url": "https://github.com/huggingface/datasets/pull/2290", "merged_at": "2021-05-06T17:25:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/2290.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2290" }
871,145,817
[]
https://api.github.com/repos/huggingface/datasets/issues/2290
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This is the "hieroglyph corpus" that I could unfortunately not contribute during the marathon. I re-extracted it again now, so that it is in the state as used in my paper (seee documentation). I hope it satiesfies your requirements and wish every scientist out their loads of fun deciphering a 5.000 years old language :-)
2021-05-06T17:25:25Z
https://github.com/huggingface/datasets/pull/2290
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 3, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/2290/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2289/comments
https://api.github.com/repos/huggingface/datasets/issues/2289/timeline
2021-04-30T18:28:16Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI2MTg5MDU3
closed
[]
false
2,289
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Allow collaborators to self-assign issues
https://api.github.com/repos/huggingface/datasets/issues/2289/events
null
https://api.github.com/repos/huggingface/datasets/issues/2289/labels{/name}
2021-04-29T15:07:06Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2289.diff", "html_url": "https://github.com/huggingface/datasets/pull/2289", "merged_at": "2021-04-30T18:28:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/2289.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2289" }
871,118,573
[]
https://api.github.com/repos/huggingface/datasets/issues/2289
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Allow collaborators (without write access to the repository) to self-assign issues. In order to self-assign an issue, they have to comment it with the word: `#take` or `#self-assign`.
2021-04-30T18:28:16Z
https://github.com/huggingface/datasets/pull/2289
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2289/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2288/comments
https://api.github.com/repos/huggingface/datasets/issues/2288/timeline
2021-06-15T13:49:26Z
null
completed
MDU6SXNzdWU4NzExMTEyMzU=
closed
[]
null
2,288
{ "avatar_url": "https://avatars.githubusercontent.com/u/17052700?v=4", "events_url": "https://api.github.com/users/sstojanoska/events{/privacy}", "followers_url": "https://api.github.com/users/sstojanoska/followers", "following_url": "https://api.github.com/users/sstojanoska/following{/other_user}", "gists_url": "https://api.github.com/users/sstojanoska/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sstojanoska", "id": 17052700, "login": "sstojanoska", "node_id": "MDQ6VXNlcjE3MDUyNzAw", "organizations_url": "https://api.github.com/users/sstojanoska/orgs", "received_events_url": "https://api.github.com/users/sstojanoska/received_events", "repos_url": "https://api.github.com/users/sstojanoska/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sstojanoska/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sstojanoska/subscriptions", "type": "User", "url": "https://api.github.com/users/sstojanoska" }
Load_dataset for local CSV files
https://api.github.com/repos/huggingface/datasets/issues/2288/events
null
https://api.github.com/repos/huggingface/datasets/issues/2288/labels{/name}
2021-04-29T15:01:10Z
null
false
null
null
871,111,235
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2288
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
The method load_dataset fails to correctly load a dataset from csv. Moreover, I am working on a token-classification task ( POS tagging) , where each row in my CSV contains two columns each of them having a list of strings. row example: ```tokens | labels ['I' , 'am', 'John'] | ['PRON', 'AUX', 'PROPN' ] ``` The method, loads each list as a string: (i.g "['I' , 'am', 'John']"). To solve this issue, I copied the Datasets.Features, created Sequence types ( instead of Value) and tried to cast the features type ``` new_features['tokens'] = Sequence(feature=Value(dtype='string', id=None)) new_features['labels'] = Sequence(feature=ClassLabel(num_classes=len(tag2idx), names=list(unique_tags))) dataset = dataset.cast(new_features) ``` but I got the following error ``` ArrowNotImplementedError: Unsupported cast from string to list using function cast_list ``` Moreover, I tried to set feature parameter in load_dataset method, to my new_features, but this fails as well. How can this be solved ?
2021-06-15T13:49:26Z
https://github.com/huggingface/datasets/issues/2288
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2288/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2287/comments
https://api.github.com/repos/huggingface/datasets/issues/2287/timeline
2021-04-29T16:34:22Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI2MTQ0MTQ3
closed
[]
false
2,287
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Avoid copying table's record batches
https://api.github.com/repos/huggingface/datasets/issues/2287/events
null
https://api.github.com/repos/huggingface/datasets/issues/2287/labels{/name}
2021-04-29T14:15:01Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2287.diff", "html_url": "https://github.com/huggingface/datasets/pull/2287", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2287.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2287" }
871,063,374
[]
https://api.github.com/repos/huggingface/datasets/issues/2287
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Fixes #2276
2021-04-29T16:34:23Z
https://github.com/huggingface/datasets/pull/2287
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2287/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2286/comments
https://api.github.com/repos/huggingface/datasets/issues/2286/timeline
2021-04-29T14:07:28Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI2MTE5MTE2
closed
[]
false
2,286
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix metadata validation with config names
https://api.github.com/repos/huggingface/datasets/issues/2286/events
null
https://api.github.com/repos/huggingface/datasets/issues/2286/labels{/name}
2021-04-29T13:44:32Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2286.diff", "html_url": "https://github.com/huggingface/datasets/pull/2286", "merged_at": "2021-04-29T14:07:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/2286.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2286" }
871,032,393
[]
https://api.github.com/repos/huggingface/datasets/issues/2286
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
I noticed in https://github.com/huggingface/datasets/pull/2280 that the metadata validator doesn't parse the tags in the readme properly when then contain the tags per config.
2021-04-29T14:07:29Z
https://github.com/huggingface/datasets/pull/2286
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2286/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2285/comments
https://api.github.com/repos/huggingface/datasets/issues/2285/timeline
2021-05-19T07:22:39Z
null
completed
MDU6SXNzdWU4NzEwMDUyMzY=
closed
[]
null
2,285
{ "avatar_url": "https://avatars.githubusercontent.com/u/46021411?v=4", "events_url": "https://api.github.com/users/danieldiezmallo/events{/privacy}", "followers_url": "https://api.github.com/users/danieldiezmallo/followers", "following_url": "https://api.github.com/users/danieldiezmallo/following{/other_user}", "gists_url": "https://api.github.com/users/danieldiezmallo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/danieldiezmallo", "id": 46021411, "login": "danieldiezmallo", "node_id": "MDQ6VXNlcjQ2MDIxNDEx", "organizations_url": "https://api.github.com/users/danieldiezmallo/orgs", "received_events_url": "https://api.github.com/users/danieldiezmallo/received_events", "repos_url": "https://api.github.com/users/danieldiezmallo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/danieldiezmallo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/danieldiezmallo/subscriptions", "type": "User", "url": "https://api.github.com/users/danieldiezmallo" }
Help understanding how to build a dataset for language modeling as with the old TextDataset
https://api.github.com/repos/huggingface/datasets/issues/2285/events
null
https://api.github.com/repos/huggingface/datasets/issues/2285/labels{/name}
2021-04-29T13:16:45Z
null
false
null
null
871,005,236
[]
https://api.github.com/repos/huggingface/datasets/issues/2285
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Hello, I am trying to load a custom dataset that I will then use for language modeling. The dataset consists of a text file that has a whole document in each line, meaning that each line overpasses the normal 512 tokens limit of most tokenizers. I would like to understand what is the process to build a text dataset that tokenizes each line, having previously split the documents in the dataset into lines of a "tokenizable" size, as the old TextDataset class would do, where you only had to do the following, and a tokenized dataset without text loss would be available to pass to a DataCollator: ``` model_checkpoint = 'distilbert-base-uncased' from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained(model_checkpoint) from transformers import TextDataset dataset = TextDataset( tokenizer=tokenizer, file_path="path/to/text_file.txt", block_size=512, ) ``` For now, what I have is the following, which, of course, throws an error because each line is longer than the maximum block size in the tokenizer: ``` import datasets dataset = datasets.load_dataset('path/to/text_file.txt') model_checkpoint = 'distilbert-base-uncased' tokenizer = AutoTokenizer.from_pretrained(model_checkpoint) def tokenize_function(examples): return tokenizer(examples["text"]) tokenized_datasets = dataset.map(tokenize_function, batched=True, num_proc=4, remove_columns=["text"]) tokenized_datasets ``` So what would be the "standard" way of creating a dataset in the way it was done before? Thank you very much for the help :))
2021-05-19T07:22:45Z
https://github.com/huggingface/datasets/issues/2285
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2285/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2284/comments
https://api.github.com/repos/huggingface/datasets/issues/2284/timeline
2021-04-29T12:54:34Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI2MDM5MDc5
closed
[]
false
2,284
{ "avatar_url": "https://avatars.githubusercontent.com/u/52530809?v=4", "events_url": "https://api.github.com/users/BobbyManion/events{/privacy}", "followers_url": "https://api.github.com/users/BobbyManion/followers", "following_url": "https://api.github.com/users/BobbyManion/following{/other_user}", "gists_url": "https://api.github.com/users/BobbyManion/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/BobbyManion", "id": 52530809, "login": "BobbyManion", "node_id": "MDQ6VXNlcjUyNTMwODA5", "organizations_url": "https://api.github.com/users/BobbyManion/orgs", "received_events_url": "https://api.github.com/users/BobbyManion/received_events", "repos_url": "https://api.github.com/users/BobbyManion/repos", "site_admin": false, "starred_url": "https://api.github.com/users/BobbyManion/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BobbyManion/subscriptions", "type": "User", "url": "https://api.github.com/users/BobbyManion" }
Initialize Imdb dataset as used in Don't Stop Pretraining Paper
https://api.github.com/repos/huggingface/datasets/issues/2284/events
null
https://api.github.com/repos/huggingface/datasets/issues/2284/labels{/name}
2021-04-29T11:52:38Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2284.diff", "html_url": "https://github.com/huggingface/datasets/pull/2284", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2284.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2284" }
870,932,710
[]
https://api.github.com/repos/huggingface/datasets/issues/2284
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
2021-04-29T12:54:34Z
https://github.com/huggingface/datasets/pull/2284
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2284/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2283/comments
https://api.github.com/repos/huggingface/datasets/issues/2283/timeline
2021-04-29T11:50:24Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI2MDM0MDk5
closed
[]
false
2,283
{ "avatar_url": "https://avatars.githubusercontent.com/u/52530809?v=4", "events_url": "https://api.github.com/users/BobbyManion/events{/privacy}", "followers_url": "https://api.github.com/users/BobbyManion/followers", "following_url": "https://api.github.com/users/BobbyManion/following{/other_user}", "gists_url": "https://api.github.com/users/BobbyManion/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/BobbyManion", "id": 52530809, "login": "BobbyManion", "node_id": "MDQ6VXNlcjUyNTMwODA5", "organizations_url": "https://api.github.com/users/BobbyManion/orgs", "received_events_url": "https://api.github.com/users/BobbyManion/received_events", "repos_url": "https://api.github.com/users/BobbyManion/repos", "site_admin": false, "starred_url": "https://api.github.com/users/BobbyManion/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BobbyManion/subscriptions", "type": "User", "url": "https://api.github.com/users/BobbyManion" }
Initialize imdb dataset from don't stop pretraining paper
https://api.github.com/repos/huggingface/datasets/issues/2283/events
null
https://api.github.com/repos/huggingface/datasets/issues/2283/labels{/name}
2021-04-29T11:44:54Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2283.diff", "html_url": "https://github.com/huggingface/datasets/pull/2283", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2283.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2283" }
870,926,475
[]
https://api.github.com/repos/huggingface/datasets/issues/2283
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
2021-04-29T11:50:24Z
https://github.com/huggingface/datasets/pull/2283
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2283/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2282/comments
https://api.github.com/repos/huggingface/datasets/issues/2282/timeline
2021-04-29T11:43:51Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI2MDEyMzM3
closed
[]
false
2,282
{ "avatar_url": "https://avatars.githubusercontent.com/u/52530809?v=4", "events_url": "https://api.github.com/users/BobbyManion/events{/privacy}", "followers_url": "https://api.github.com/users/BobbyManion/followers", "following_url": "https://api.github.com/users/BobbyManion/following{/other_user}", "gists_url": "https://api.github.com/users/BobbyManion/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/BobbyManion", "id": 52530809, "login": "BobbyManion", "node_id": "MDQ6VXNlcjUyNTMwODA5", "organizations_url": "https://api.github.com/users/BobbyManion/orgs", "received_events_url": "https://api.github.com/users/BobbyManion/received_events", "repos_url": "https://api.github.com/users/BobbyManion/repos", "site_admin": false, "starred_url": "https://api.github.com/users/BobbyManion/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BobbyManion/subscriptions", "type": "User", "url": "https://api.github.com/users/BobbyManion" }
Initialize imdb dataset from don't stop pretraining paper
https://api.github.com/repos/huggingface/datasets/issues/2282/events
null
https://api.github.com/repos/huggingface/datasets/issues/2282/labels{/name}
2021-04-29T11:17:56Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2282.diff", "html_url": "https://github.com/huggingface/datasets/pull/2282", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2282.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2282" }
870,900,332
[]
https://api.github.com/repos/huggingface/datasets/issues/2282
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
2021-04-29T11:43:51Z
https://github.com/huggingface/datasets/pull/2282
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2282/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2281/comments
https://api.github.com/repos/huggingface/datasets/issues/2281/timeline
2021-04-29T13:41:34Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI1OTI2MjAw
closed
[]
false
2,281
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Update multi_woz_v22 checksum
https://api.github.com/repos/huggingface/datasets/issues/2281/events
null
https://api.github.com/repos/huggingface/datasets/issues/2281/labels{/name}
2021-04-29T09:09:11Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2281.diff", "html_url": "https://github.com/huggingface/datasets/pull/2281", "merged_at": "2021-04-29T13:41:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/2281.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2281" }
870,792,784
[]
https://api.github.com/repos/huggingface/datasets/issues/2281
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix issue https://github.com/huggingface/datasets/issues/1876 The files were changed in https://github.com/budzianowski/multiwoz/pull/72
2021-04-29T13:41:35Z
https://github.com/huggingface/datasets/pull/2281
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2281/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2280/comments
https://api.github.com/repos/huggingface/datasets/issues/2280/timeline
2021-04-29T16:41:16Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI1OTE2Mzcy
closed
[]
false
2,280
{ "avatar_url": "https://avatars.githubusercontent.com/u/32505743?v=4", "events_url": "https://api.github.com/users/laksh9950/events{/privacy}", "followers_url": "https://api.github.com/users/laksh9950/followers", "following_url": "https://api.github.com/users/laksh9950/following{/other_user}", "gists_url": "https://api.github.com/users/laksh9950/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/laksh9950", "id": 32505743, "login": "laksh9950", "node_id": "MDQ6VXNlcjMyNTA1NzQz", "organizations_url": "https://api.github.com/users/laksh9950/orgs", "received_events_url": "https://api.github.com/users/laksh9950/received_events", "repos_url": "https://api.github.com/users/laksh9950/repos", "site_admin": false, "starred_url": "https://api.github.com/users/laksh9950/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/laksh9950/subscriptions", "type": "User", "url": "https://api.github.com/users/laksh9950" }
Fixed typo seperate->separate
https://api.github.com/repos/huggingface/datasets/issues/2280/events
null
https://api.github.com/repos/huggingface/datasets/issues/2280/labels{/name}
2021-04-29T08:55:46Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2280.diff", "html_url": "https://github.com/huggingface/datasets/pull/2280", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2280.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2280" }
870,780,431
[]
https://api.github.com/repos/huggingface/datasets/issues/2280
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
2021-04-29T16:41:22Z
https://github.com/huggingface/datasets/pull/2280
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2280/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2279/comments
https://api.github.com/repos/huggingface/datasets/issues/2279/timeline
2021-04-29T07:42:42Z
null
completed
MDU6SXNzdWU4NzA0MzE2NjI=
closed
[]
null
2,279
{ "avatar_url": "https://avatars.githubusercontent.com/u/11379648?v=4", "events_url": "https://api.github.com/users/tginart/events{/privacy}", "followers_url": "https://api.github.com/users/tginart/followers", "following_url": "https://api.github.com/users/tginart/following{/other_user}", "gists_url": "https://api.github.com/users/tginart/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/tginart", "id": 11379648, "login": "tginart", "node_id": "MDQ6VXNlcjExMzc5NjQ4", "organizations_url": "https://api.github.com/users/tginart/orgs", "received_events_url": "https://api.github.com/users/tginart/received_events", "repos_url": "https://api.github.com/users/tginart/repos", "site_admin": false, "starred_url": "https://api.github.com/users/tginart/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tginart/subscriptions", "type": "User", "url": "https://api.github.com/users/tginart" }
Compatibility with Ubuntu 18 and GLIBC 2.27?
https://api.github.com/repos/huggingface/datasets/issues/2279/events
null
https://api.github.com/repos/huggingface/datasets/issues/2279/labels{/name}
2021-04-28T22:08:07Z
null
false
null
null
870,431,662
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2279
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug For use on Ubuntu systems, it seems that datasets requires GLIBC 2.29. However, Ubuntu 18 runs with GLIBC 2.27 and it seems [non-trivial to upgrade GLIBC to 2.29 for Ubuntu 18 users](https://www.digitalocean.com/community/questions/how-install-glibc-2-29-or-higher-in-ubuntu-18-04). I'm not sure if there is anything that can be done about this, but I'd like to confirm that using huggingface/datasets requires either an upgrade to Ubuntu 19/20 or a hand-rolled install of a higher version of GLIBC. ## Steps to reproduce the bug 1. clone the transformers repo 2. move to examples/pytorch/language-modeling 3. run example command: ```python run_clm.py --model_name_or_path gpt2 --dataset_name wikitext --dataset_config_name wikitext-2-raw-v1 --do_train --do_eval --output_dir /tmp/test-clm``` ## Expected results As described in the transformers repo. ## Actual results ```Traceback (most recent call last): File "run_clm.py", line 34, in <module> from transformers import ( File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/transformers/__init__.py", line 2487, in __getattr__ return super().__getattr__(name) File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/transformers/file_utils.py", line 1699, in __getattr__ module = self._get_module(self._class_to_module[name]) File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/transformers/__init__.py", line 2481, in _get_module return importlib.import_module("." + module_name, self.__name__) File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/importlib/__init__.py", line 127, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/transformers/models/__init__.py", line 19, in <module> from . import ( File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/transformers/models/layoutlm/__init__.py", line 23, in <module> from .tokenization_layoutlm import LayoutLMTokenizer File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/transformers/models/layoutlm/tokenization_layoutlm.py", line 19, in <module> from ..bert.tokenization_bert import BertTokenizer File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/transformers/models/bert/tokenization_bert.py", line 23, in <module> from ...tokenization_utils import PreTrainedTokenizer, _is_control, _is_punctuation, _is_whitespace File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/transformers/tokenization_utils.py", line 26, in <module> from .tokenization_utils_base import ( File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/transformers/tokenization_utils_base.py", line 68, in <module> from tokenizers import AddedToken File "/home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/tokenizers/__init__.py", line 79, in <module> from .tokenizers import ( ImportError: /lib/x86_64-linux-gnu/libm.so.6: version `GLIBC_2.29' not found (required by /home/tginart/anaconda3/envs/huggingface/lib/python3.7/site-packages/tokenizers/tokenizers.cpython-37m-x86_64-linux-gnu.so) ``` ## Versions Paste the output of the following code: ``` - Datasets: 1.6.1 - Python: 3.7.10 (default, Feb 26 2021, 18:47:35) [GCC 7.3.0] - Platform: Linux-4.15.0-128-generic-x86_64-with-debian-buster-sid ```
2021-04-29T07:42:42Z
https://github.com/huggingface/datasets/issues/2279
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2279/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2278/comments
https://api.github.com/repos/huggingface/datasets/issues/2278/timeline
2021-05-06T16:14:23Z
null
completed
MDU6SXNzdWU4NzAwODgwNTk=
closed
[]
null
2,278
{ "avatar_url": "https://avatars.githubusercontent.com/u/51174606?v=4", "events_url": "https://api.github.com/users/Yossillamm/events{/privacy}", "followers_url": "https://api.github.com/users/Yossillamm/followers", "following_url": "https://api.github.com/users/Yossillamm/following{/other_user}", "gists_url": "https://api.github.com/users/Yossillamm/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Yossillamm", "id": 51174606, "login": "Yossillamm", "node_id": "MDQ6VXNlcjUxMTc0NjA2", "organizations_url": "https://api.github.com/users/Yossillamm/orgs", "received_events_url": "https://api.github.com/users/Yossillamm/received_events", "repos_url": "https://api.github.com/users/Yossillamm/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Yossillamm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Yossillamm/subscriptions", "type": "User", "url": "https://api.github.com/users/Yossillamm" }
Loss result inGptNeoForCasual
https://api.github.com/repos/huggingface/datasets/issues/2278/events
null
https://api.github.com/repos/huggingface/datasets/issues/2278/labels{/name}
2021-04-28T15:39:52Z
null
false
null
null
870,088,059
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2278
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Is there any way you give the " loss" and "logits" results in the gpt neo api?
2021-05-06T16:14:23Z
https://github.com/huggingface/datasets/issues/2278
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2278/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2277/comments
https://api.github.com/repos/huggingface/datasets/issues/2277/timeline
null
null
null
MDExOlB1bGxSZXF1ZXN0NjI1MzI5NjIz
open
[]
false
2,277
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Create CacheManager
https://api.github.com/repos/huggingface/datasets/issues/2277/events
{ "closed_at": null, "closed_issues": 2, "created_at": "2021-07-21T15:34:56Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-08-30T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/8", "id": 6968069, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/8/labels", "node_id": "MI_kwDODunzps4AalMF", "number": 8, "open_issues": 4, "state": "open", "title": "1.12", "updated_at": "2021-10-13T10:26:33Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/8" }
https://api.github.com/repos/huggingface/datasets/issues/2277/labels{/name}
2021-04-28T15:23:42Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2277.diff", "html_url": "https://github.com/huggingface/datasets/pull/2277", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2277.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2277" }
870,071,994
[ { "color": "B67A40", "default": false, "description": "Restructuring existing code without changing its external behavior", "id": 2851292821, "name": "refactoring", "node_id": "MDU6TGFiZWwyODUxMjkyODIx", "url": "https://api.github.com/repos/huggingface/datasets/labels/refactoring" } ]
https://api.github.com/repos/huggingface/datasets/issues/2277
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Perform refactoring to decouple cache functionality (method `as_dataset`).
2022-07-06T15:19:48Z
https://github.com/huggingface/datasets/pull/2277
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2277/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2276/comments
https://api.github.com/repos/huggingface/datasets/issues/2276/timeline
2021-05-03T08:41:55Z
null
completed
MDU6SXNzdWU4NzAwMTA1MTE=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
2,276
{ "avatar_url": "https://avatars.githubusercontent.com/u/7063207?v=4", "events_url": "https://api.github.com/users/chbensch/events{/privacy}", "followers_url": "https://api.github.com/users/chbensch/followers", "following_url": "https://api.github.com/users/chbensch/following{/other_user}", "gists_url": "https://api.github.com/users/chbensch/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/chbensch", "id": 7063207, "login": "chbensch", "node_id": "MDQ6VXNlcjcwNjMyMDc=", "organizations_url": "https://api.github.com/users/chbensch/orgs", "received_events_url": "https://api.github.com/users/chbensch/received_events", "repos_url": "https://api.github.com/users/chbensch/repos", "site_admin": false, "starred_url": "https://api.github.com/users/chbensch/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chbensch/subscriptions", "type": "User", "url": "https://api.github.com/users/chbensch" }
concatenate_datasets loads all the data into memory
https://api.github.com/repos/huggingface/datasets/issues/2276/events
null
https://api.github.com/repos/huggingface/datasets/issues/2276/labels{/name}
2021-04-28T14:27:21Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
870,010,511
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2276
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug When I try to concatenate 2 datasets (10GB each) , the entire data is loaded into memory instead of being written directly to disk. Interestingly, this happens when trying to save the new dataset to disk or concatenating it again. ![image](https://user-images.githubusercontent.com/7063207/116420321-2b21b480-a83e-11eb-9006-8f6ca729fb6f.png) ## Steps to reproduce the bug ```python from datasets import concatenate_datasets, load_from_disk test_sampled_pro = load_from_disk("test_sampled_pro") val_sampled_pro = load_from_disk("val_sampled_pro") big_set = concatenate_datasets([test_sampled_pro, val_sampled_pro]) # Loaded to memory big_set.save_to_disk("big_set") # Loaded to memory big_set = concatenate_datasets([big_set, val_sampled_pro]) ``` ## Expected results The data should be loaded into memory in batches and then saved directly to disk. ## Actual results The entire data set is loaded into the memory and then saved to the hard disk. ## Versions Paste the output of the following code: ```python - Datasets: 1.6.1 - Python: 3.8.8 (default, Apr 13 2021, 19:58:26) [GCC 7.3.0] - Platform: Linux-5.4.72-microsoft-standard-WSL2-x86_64-with-glibc2.10 ```
2021-05-03T08:41:55Z
https://github.com/huggingface/datasets/issues/2276
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2276/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2275/comments
https://api.github.com/repos/huggingface/datasets/issues/2275/timeline
2021-05-17T13:34:18Z
null
completed
MDU6SXNzdWU4NjkzNzgzMTE=
closed
[]
null
2,275
{ "avatar_url": "https://avatars.githubusercontent.com/u/17426779?v=4", "events_url": "https://api.github.com/users/puzzler10/events{/privacy}", "followers_url": "https://api.github.com/users/puzzler10/followers", "following_url": "https://api.github.com/users/puzzler10/following{/other_user}", "gists_url": "https://api.github.com/users/puzzler10/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/puzzler10", "id": 17426779, "login": "puzzler10", "node_id": "MDQ6VXNlcjE3NDI2Nzc5", "organizations_url": "https://api.github.com/users/puzzler10/orgs", "received_events_url": "https://api.github.com/users/puzzler10/received_events", "repos_url": "https://api.github.com/users/puzzler10/repos", "site_admin": false, "starred_url": "https://api.github.com/users/puzzler10/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/puzzler10/subscriptions", "type": "User", "url": "https://api.github.com/users/puzzler10" }
SNLI dataset has labels of -1
https://api.github.com/repos/huggingface/datasets/issues/2275/events
null
https://api.github.com/repos/huggingface/datasets/issues/2275/labels{/name}
2021-04-28T00:32:25Z
null
false
null
null
869,378,311
[]
https://api.github.com/repos/huggingface/datasets/issues/2275
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
There are a number of rows with a label of -1 in the SNLI dataset. The dataset descriptions [here](https://nlp.stanford.edu/projects/snli/) and [here](https://github.com/huggingface/datasets/tree/master/datasets/snli) don't list -1 as a label possibility, and neither does the dataset viewer. As examples, see index 107 or 124 of the test set. It isn't clear what these labels mean. I found a [line of code](https://github.com/huggingface/datasets/blob/80e59ef178d3bb2090d091bc32315c655eb0633d/datasets/snli/snli.py#L94) that seems to put them in but it seems still unclear why they are there. The current workaround is to just drop the rows from any model being trained. Perhaps the documentation should be updated.
2021-05-17T13:34:18Z
https://github.com/huggingface/datasets/issues/2275
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2275/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2274/comments
https://api.github.com/repos/huggingface/datasets/issues/2274/timeline
2021-04-29T09:57:50Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI0NTkyMjQx
closed
[]
false
2,274
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Always update metadata in arrow schema
https://api.github.com/repos/huggingface/datasets/issues/2274/events
null
https://api.github.com/repos/huggingface/datasets/issues/2274/labels{/name}
2021-04-27T19:21:57Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2274.diff", "html_url": "https://github.com/huggingface/datasets/pull/2274", "merged_at": "2021-04-29T09:57:50Z", "patch_url": "https://github.com/huggingface/datasets/pull/2274.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2274" }
869,186,276
[]
https://api.github.com/repos/huggingface/datasets/issues/2274
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
We store a redundant copy of the features in the metadata of the schema of the arrow table. This is used to recover the features when doing `Dataset.from_file`. These metadata are updated after each transfor, that changes the feature types. For each function that transforms the feature types of the dataset, I added a step in the tests to make sure the metadata in the arrow schema are up to date. I also added a line to update the metadata directly in the Dataset.__init__ method. This way even a dataset instantiated with __init__ will have a table with the right metadata. Fix #2271. cc @mariosasko
2022-06-03T08:31:19Z
https://github.com/huggingface/datasets/pull/2274
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2274/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2273/comments
https://api.github.com/repos/huggingface/datasets/issues/2273/timeline
2021-04-29T13:59:47Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI0NDcxODc1
closed
[]
false
2,273
{ "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}", "followers_url": "https://api.github.com/users/bhavitvyamalik/followers", "following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}", "gists_url": "https://api.github.com/users/bhavitvyamalik/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bhavitvyamalik", "id": 19718818, "login": "bhavitvyamalik", "node_id": "MDQ6VXNlcjE5NzE4ODE4", "organizations_url": "https://api.github.com/users/bhavitvyamalik/orgs", "received_events_url": "https://api.github.com/users/bhavitvyamalik/received_events", "repos_url": "https://api.github.com/users/bhavitvyamalik/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bhavitvyamalik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhavitvyamalik/subscriptions", "type": "User", "url": "https://api.github.com/users/bhavitvyamalik" }
Added CUAD metrics
https://api.github.com/repos/huggingface/datasets/issues/2273/events
null
https://api.github.com/repos/huggingface/datasets/issues/2273/labels{/name}
2021-04-27T16:49:12Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2273.diff", "html_url": "https://github.com/huggingface/datasets/pull/2273", "merged_at": "2021-04-29T13:59:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/2273.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2273" }
869,046,290
[]
https://api.github.com/repos/huggingface/datasets/issues/2273
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
`EM`, `F1`, `AUPR`, `Precision@80%Recall`, and `Precision@90%Recall` metrics supported for CUAD
2021-04-29T13:59:47Z
https://github.com/huggingface/datasets/pull/2273
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2273/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2272/comments
https://api.github.com/repos/huggingface/datasets/issues/2272/timeline
2021-04-30T12:54:27Z
null
completed
MDU6SXNzdWU4NjkwMTc5Nzc=
closed
[]
null
2,272
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Bug in Dataset.class_encode_column
https://api.github.com/repos/huggingface/datasets/issues/2272/events
null
https://api.github.com/repos/huggingface/datasets/issues/2272/labels{/name}
2021-04-27T16:13:18Z
null
false
null
null
869,017,977
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2272
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
## Describe the bug All the rest of the columns except the one passed to `Dataset.class_encode_column` are discarded. ## Expected results All the original columns should be kept. This needs regression tests.
2021-04-30T12:54:27Z
https://github.com/huggingface/datasets/issues/2272
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2272/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2271/comments
https://api.github.com/repos/huggingface/datasets/issues/2271/timeline
2022-06-01T17:13:21Z
null
completed
MDU6SXNzdWU4NjkwMDIxNDE=
closed
[]
null
2,271
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Synchronize table metadata with features
https://api.github.com/repos/huggingface/datasets/issues/2271/events
null
https://api.github.com/repos/huggingface/datasets/issues/2271/labels{/name}
2021-04-27T15:55:13Z
null
false
null
null
869,002,141
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2271
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
**Is your feature request related to a problem? Please describe.** As pointed out in this [comment](https://github.com/huggingface/datasets/pull/2145#discussion_r621326767): > Metadata stored in the schema is just a redundant information regarding the feature types. It is used when calling Dataset.from_file to know which feature types to use. These metadata are stored in the schema of the pyarrow table by using `update_metadata_with_features`. However this something that's almost never tested properly. **Describe the solution you'd like** We should find a way to always make sure that the metadata (in `self.data.schema.metadata`) are synced with the actual feature types (in `self.info.features`).
2022-06-01T17:13:21Z
https://github.com/huggingface/datasets/issues/2271
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2271/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2270/comments
https://api.github.com/repos/huggingface/datasets/issues/2270/timeline
2021-04-28T17:39:27Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI0MzU5Njky
closed
[]
false
2,270
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix iterable interface expected by numpy
https://api.github.com/repos/huggingface/datasets/issues/2270/events
null
https://api.github.com/repos/huggingface/datasets/issues/2270/labels{/name}
2021-04-27T14:35:56Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2270.diff", "html_url": "https://github.com/huggingface/datasets/pull/2270", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2270.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2270" }
868,913,660
[]
https://api.github.com/repos/huggingface/datasets/issues/2270
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Numpy expects the old iterable interface with `__getitem__` instead of `__iter__`.
2021-04-28T17:39:27Z
https://github.com/huggingface/datasets/pull/2270
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2270/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2269/comments
https://api.github.com/repos/huggingface/datasets/issues/2269/timeline
2021-04-27T14:21:56Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI0MzMwNDA3
closed
[]
false
2,269
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix query table with iterable
https://api.github.com/repos/huggingface/datasets/issues/2269/events
null
https://api.github.com/repos/huggingface/datasets/issues/2269/labels{/name}
2021-04-27T13:59:38Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2269.diff", "html_url": "https://github.com/huggingface/datasets/pull/2269", "merged_at": "2021-04-27T14:21:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/2269.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2269" }
868,878,468
[]
https://api.github.com/repos/huggingface/datasets/issues/2269
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
The benchmark runs are failing on master because it tries to use an iterable to query the dataset. However there's currently an issue caused by the use of `np.array` instead of `np.fromiter` on the iterable. This PR fixes it
2021-04-27T14:21:57Z
https://github.com/huggingface/datasets/pull/2269
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2269/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2268/comments
https://api.github.com/repos/huggingface/datasets/issues/2268/timeline
2021-04-27T13:43:20Z
null
null
MDExOlB1bGxSZXF1ZXN0NjI0MjQyODg1
closed
[]
false
2,268
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Don't use pyarrow 4.0.0 since it segfaults when casting a sliced ListArray of integers
https://api.github.com/repos/huggingface/datasets/issues/2268/events
null
https://api.github.com/repos/huggingface/datasets/issues/2268/labels{/name}
2021-04-27T11:58:28Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2268.diff", "html_url": "https://github.com/huggingface/datasets/pull/2268", "merged_at": "2021-04-27T13:43:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/2268.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2268" }
868,773,380
[]
https://api.github.com/repos/huggingface/datasets/issues/2268
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This test `tests/test_table.py::test_concatenation_table_cast` segfaults with the latest update of pyarrow 4.0.0. Setting `pyarrow<4.0.0` for now. I'll open an issue on JIRA once I know more about the origin of the issue
2021-06-12T12:44:49Z
https://github.com/huggingface/datasets/pull/2268
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2268/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2267/comments
https://api.github.com/repos/huggingface/datasets/issues/2267/timeline
null
null
null
MDU6SXNzdWU4NjgyOTExMjk=
open
[]
null
2,267
{ "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "events_url": "https://api.github.com/users/timothyjlaurent/events{/privacy}", "followers_url": "https://api.github.com/users/timothyjlaurent/followers", "following_url": "https://api.github.com/users/timothyjlaurent/following{/other_user}", "gists_url": "https://api.github.com/users/timothyjlaurent/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/timothyjlaurent", "id": 2000204, "login": "timothyjlaurent", "node_id": "MDQ6VXNlcjIwMDAyMDQ=", "organizations_url": "https://api.github.com/users/timothyjlaurent/orgs", "received_events_url": "https://api.github.com/users/timothyjlaurent/received_events", "repos_url": "https://api.github.com/users/timothyjlaurent/repos", "site_admin": false, "starred_url": "https://api.github.com/users/timothyjlaurent/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/timothyjlaurent/subscriptions", "type": "User", "url": "https://api.github.com/users/timothyjlaurent" }
DatasetDict save load Failing test in 1.6 not in 1.5
https://api.github.com/repos/huggingface/datasets/issues/2267/events
null
https://api.github.com/repos/huggingface/datasets/issues/2267/labels{/name}
2021-04-27T00:03:25Z
null
false
null
null
868,291,129
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2267
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug We have a test that saves a DatasetDict to disk and then loads it from disk. In 1.6 there is an incompatibility in the schema. Downgrading to `>1.6` -- fixes the problem. ## Steps to reproduce the bug ```python ### Load a dataset dict from jsonl path = '/test/foo' ds_dict.save_to_disk(path) ds_from_disk = DatasetDict.load_from_disk(path). ## <-- this is where I see the error on 1.6 ``` ## Expected results Upgrading to 1.6 shouldn't break that test. We should be able to serialize to and from disk. ## Actual results ``` # Infer features if None inferred_features = Features.from_arrow_schema(arrow_table.schema) if self.info.features is None: self.info.features = inferred_features # Infer fingerprint if None if self._fingerprint is None: self._fingerprint = generate_fingerprint(self) # Sanity checks assert self.features is not None, "Features can't be None in a Dataset object" assert self._fingerprint is not None, "Fingerprint can't be None in a Dataset object" if self.info.features.type != inferred_features.type: > raise ValueError( "External features info don't match the dataset:\nGot\n{}\nwith type\n{}\n\nbut expected something like\n{}\nwith type\n{}".format( self.info.features, self.info.features.type, inferred_features, inferred_features.type ) ) E ValueError: External features info don't match the dataset: E Got E {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'child': Value(dtype='int64', id=None), 'child_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'color': Value(dtype='string', id=None), 'head': Value(dtype='int64', id=None), 'head_span': {'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None)}, 'label': Value(dtype='string', id=None)}], 'spans': [{'end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'token_end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'disabled': Value(dtype='bool', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'start': Value(dtype='int64', id=None), 'text': Value(dtype='string', id=None), 'ws': Value(dtype='bool', id=None)}]} E with type E struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list<item: int64>, encoding__offsets: list<item: list<item: int64>>, encoding__overflowing: list<item: null>, encoding__tokens: list<item: string>, encoding__words: list<item: int64>, ner_ids: list<item: int64>, ner_labels: list<item: string>, relations: list<item: struct<child: int64, child_span: struct<end: int64, label: string, start: int64, token_end: int64, token_start: int64>, color: string, head: int64, head_span: struct<end: int64, label: string, start: int64, token_end: int64, token_start: int64>, label: string>>, spans: list<item: struct<end: int64, label: string, start: int64, text: string, token_end: int64, token_start: int64, type: string>>, text: string, tokens: list<item: struct<disabled: bool, end: int64, id: int64, start: int64, text: string, ws: bool>>> E E but expected something like E {'_input_hash': Value(dtype='int64', id=None), '_task_hash': Value(dtype='int64', id=None), '_view_id': Value(dtype='string', id=None), 'answer': Value(dtype='string', id=None), 'encoding__ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'encoding__offsets': Sequence(feature=Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), length=-1, id=None), 'encoding__overflowing': Sequence(feature=Value(dtype='null', id=None), length=-1, id=None), 'encoding__tokens': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'encoding__words': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_ids': Sequence(feature=Value(dtype='int64', id=None), length=-1, id=None), 'ner_labels': Sequence(feature=Value(dtype='string', id=None), length=-1, id=None), 'relations': [{'head': Value(dtype='int64', id=None), 'child': Value(dtype='int64', id=None), 'head_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'child_span': {'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'label': Value(dtype='string', id=None)}, 'color': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'spans': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'token_start': Value(dtype='int64', id=None), 'token_end': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'type': Value(dtype='string', id=None), 'label': Value(dtype='string', id=None)}], 'text': Value(dtype='string', id=None), 'tokens': [{'text': Value(dtype='string', id=None), 'start': Value(dtype='int64', id=None), 'end': Value(dtype='int64', id=None), 'id': Value(dtype='int64', id=None), 'ws': Value(dtype='bool', id=None), 'disabled': Value(dtype='bool', id=None)}]} E with type E struct<_input_hash: int64, _task_hash: int64, _view_id: string, answer: string, encoding__ids: list<item: int64>, encoding__offsets: list<item: list<item: int64>>, encoding__overflowing: list<item: null>, encoding__tokens: list<item: string>, encoding__words: list<item: int64>, ner_ids: list<item: int64>, ner_labels: list<item: string>, relations: list<item: struct<head: int64, child: int64, head_span: struct<start: int64, end: int64, token_start: int64, token_end: int64, label: string>, child_span: struct<start: int64, end: int64, token_start: int64, token_end: int64, label: string>, color: string, label: string>>, spans: list<item: struct<text: string, start: int64, token_start: int64, token_end: int64, end: int64, type: string, label: string>>, text: string, tokens: list<item: struct<text: string, start: int64, end: int64, id: int64, ws: bool, disabled: bool>>> ../../../../../.virtualenvs/tf_ner_rel_lib/lib/python3.8/site-packages/datasets/arrow_dataset.py:274: ValueError ``` ## Versions - Datasets: 1.6.1 - Python: 3.8.5 (default, Jan 26 2021, 10:01:04) [Clang 12.0.0 (clang-1200.0.32.2)] - Platform: macOS-10.15.7-x86_64-i386-64bit ```
2021-05-28T15:27:34Z
https://github.com/huggingface/datasets/issues/2267
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2267/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2266/comments
https://api.github.com/repos/huggingface/datasets/issues/2266/timeline
2021-04-29T10:00:04Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIzNDY1OTI5
closed
[]
false
2,266
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Make tests run faster
https://api.github.com/repos/huggingface/datasets/issues/2266/events
null
https://api.github.com/repos/huggingface/datasets/issues/2266/labels{/name}
2021-04-26T15:55:40Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2266.diff", "html_url": "https://github.com/huggingface/datasets/pull/2266", "merged_at": "2021-04-29T10:00:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/2266.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2266" }
867,864,353
[]
https://api.github.com/repos/huggingface/datasets/issues/2266
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
From 7min to 2min to run pytest. Ideally we should keep the whole CI run time below 10min. In this PR I removed the remote tests that were never used. I also replaced nested parametrized tests with unit tests. This makes me think that we could still add more high level tests to check for a few combinations of parameters (but not all of them since there are too many of them). Let me know what you think Finally in another PR we can also separate in two circleci jobs: - the tests of the code code of the lib - the tests of the all the dataset/metric scripts.
2021-04-29T10:00:13Z
https://github.com/huggingface/datasets/pull/2266
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2266/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2265/comments
https://api.github.com/repos/huggingface/datasets/issues/2265/timeline
2021-04-26T09:47:47Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIzMTUyOTg5
closed
[]
false
2,265
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Update black
https://api.github.com/repos/huggingface/datasets/issues/2265/events
null
https://api.github.com/repos/huggingface/datasets/issues/2265/labels{/name}
2021-04-26T09:35:09Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2265.diff", "html_url": "https://github.com/huggingface/datasets/pull/2265", "merged_at": "2021-04-26T09:47:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/2265.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2265" }
867,490,646
[]
https://api.github.com/repos/huggingface/datasets/issues/2265
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Latest black version 21.4b0 requires to reformat most dataset scripts and also the core code of the lib. This makes the CI currently fail on master
2021-04-26T09:47:48Z
https://github.com/huggingface/datasets/pull/2265
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2265/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2264/comments
https://api.github.com/repos/huggingface/datasets/issues/2264/timeline
2021-04-26T10:08:14Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIzMTQwODA1
closed
[]
false
2,264
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix memory issue in multiprocessing: Don't pickle table index
https://api.github.com/repos/huggingface/datasets/issues/2264/events
null
https://api.github.com/repos/huggingface/datasets/issues/2264/labels{/name}
2021-04-26T09:21:35Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2264.diff", "html_url": "https://github.com/huggingface/datasets/pull/2264", "merged_at": "2021-04-26T10:08:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/2264.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2264" }
867,476,228
[]
https://api.github.com/repos/huggingface/datasets/issues/2264
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
The table index is currently being pickled when doing multiprocessing, which brings all the record batches of the dataset in memory. I fixed that by not pickling the index attributes. Therefore each process has to rebuild the index when unpickling the table. Fix issue #2256 We'll do a patch release asap !
2021-04-26T10:30:28Z
https://github.com/huggingface/datasets/pull/2264
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2264/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2263/comments
https://api.github.com/repos/huggingface/datasets/issues/2263/timeline
2021-04-29T09:30:20Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIzMDk0NTcy
closed
[]
false
2,263
{ "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}", "followers_url": "https://api.github.com/users/bhavitvyamalik/followers", "following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}", "gists_url": "https://api.github.com/users/bhavitvyamalik/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bhavitvyamalik", "id": 19718818, "login": "bhavitvyamalik", "node_id": "MDQ6VXNlcjE5NzE4ODE4", "organizations_url": "https://api.github.com/users/bhavitvyamalik/orgs", "received_events_url": "https://api.github.com/users/bhavitvyamalik/received_events", "repos_url": "https://api.github.com/users/bhavitvyamalik/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bhavitvyamalik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhavitvyamalik/subscriptions", "type": "User", "url": "https://api.github.com/users/bhavitvyamalik" }
test data added, dataset_infos updated
https://api.github.com/repos/huggingface/datasets/issues/2263/events
null
https://api.github.com/repos/huggingface/datasets/issues/2263/labels{/name}
2021-04-26T08:27:18Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2263.diff", "html_url": "https://github.com/huggingface/datasets/pull/2263", "merged_at": "2021-04-29T09:30:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/2263.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2263" }
867,420,912
[]
https://api.github.com/repos/huggingface/datasets/issues/2263
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Fixes #2262. Thanks for pointing out issue with dataset @jinmang2!
2021-04-29T09:30:21Z
https://github.com/huggingface/datasets/pull/2263
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2263/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2262/comments
https://api.github.com/repos/huggingface/datasets/issues/2262/timeline
2021-04-29T09:30:20Z
null
completed
MDU6SXNzdWU4NjczMjUzNTE=
closed
[]
null
2,262
{ "avatar_url": "https://avatars.githubusercontent.com/u/37775784?v=4", "events_url": "https://api.github.com/users/jinmang2/events{/privacy}", "followers_url": "https://api.github.com/users/jinmang2/followers", "following_url": "https://api.github.com/users/jinmang2/following{/other_user}", "gists_url": "https://api.github.com/users/jinmang2/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jinmang2", "id": 37775784, "login": "jinmang2", "node_id": "MDQ6VXNlcjM3Nzc1Nzg0", "organizations_url": "https://api.github.com/users/jinmang2/orgs", "received_events_url": "https://api.github.com/users/jinmang2/received_events", "repos_url": "https://api.github.com/users/jinmang2/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jinmang2/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jinmang2/subscriptions", "type": "User", "url": "https://api.github.com/users/jinmang2" }
NewsPH NLI dataset script fails to access test data.
https://api.github.com/repos/huggingface/datasets/issues/2262/events
null
https://api.github.com/repos/huggingface/datasets/issues/2262/labels{/name}
2021-04-26T06:44:41Z
null
false
null
null
867,325,351
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2262
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
In Newsph-NLI Dataset (#1192), it fails to access test data. According to the script below, the download manager will download the train data when trying to download the test data. https://github.com/huggingface/datasets/blob/2a2dd6316af2cc7fdf24e4779312e8ee0c7ed98b/datasets/newsph_nli/newsph_nli.py#L71 If you download it according to the script above, you can see that train and test receive the same data as shown below. ```python >>> from datasets import load_dataset >>> newsph_nli = load_dataset(path="./datasets/newsph_nli.py") >>> newsph_nli DatasetDict({ train: Dataset({ features: ['premise', 'hypothesis', 'label'], num_rows: 420000 }) test: Dataset({ features: ['premise', 'hypothesis', 'label'], num_rows: 420000 }) validation: Dataset({ features: ['premise', 'hypothesis', 'label'], num_rows: 90000 }) }) >>> newsph_nli["train"][0] {'hypothesis': 'Ito ang dineklara ni Atty. Romulo Macalintal, abogado ni Robredo, kaugnay ng pagsisimula ng preliminary conference ngayong hapon sa Presidential Electoral Tribunal (PET).', 'label': 1, 'premise': '"Hindi ko ugali ang mamulitika; mas gusto kong tahimik na magtrabaho. Pero sasabihin ko ito ngayon: ang tapang, lakas, at diskarte, hindi nadadaan sa mapanirang salita. Ang kailangan ng taumbayan ay tapang sa gawa," ayon kay Robredo sa inilabas nitong statement.'} >>> newsph_nli["test"][0] {'hypothesis': 'Ito ang dineklara ni Atty. Romulo Macalintal, abogado ni Robredo, kaugnay ng pagsisimula ng preliminary conference ngayong hapon sa Presidential Electoral Tribunal (PET).', 'label': 1, 'premise': '"Hindi ko ugali ang mamulitika; mas gusto kong tahimik na magtrabaho. Pero sasabihin ko ito ngayon: ang tapang, lakas, at diskarte, hindi nadadaan sa mapanirang salita. Ang kailangan ng taumbayan ay tapang sa gawa," ayon kay Robredo sa inilabas nitong statement.'} ``` In local, I modified the code of the source as below and got the correct result. ```python 71 test_path = os.path.join(download_path, "test.csv") ``` ```python >>> from datasets import load_dataset >>> newsph_nli = load_dataset(path="./datasets/newsph_nli.py") >>> newsph_nli DatasetDict({ train: Dataset({ features: ['premise', 'hypothesis', 'label'], num_rows: 420000 }) test: Dataset({ features: ['premise', 'hypothesis', 'label'], num_rows: 9000 }) validation: Dataset({ features: ['premise', 'hypothesis', 'label'], num_rows: 90000 }) }) >>> newsph_nli["train"][0] {'hypothesis': 'Ito ang dineklara ni Atty. Romulo Macalintal, abogado ni Robredo, kaugnay ng pagsisimula ng preliminary conference ngayong hapon sa Presidential Electoral Tribunal (PET).', 'label': 1, 'premise': '"Hindi ko ugali ang mamulitika; mas gusto kong tahimik na magtrabaho. Pero sasabihin ko ito ngayon: ang tapang, lakas, at diskarte, hindi nadadaan sa mapanirang salita. Ang kailangan ng taumbayan ay tapang sa gawa," ayon kay Robredo sa inilabas nitong statement.'} >>> newsph_nli["test"][0] {'hypothesis': '-- JAI (@JaiPaller) September 13, 2019', 'label': 1, 'premise': 'Pinag-iingat ng Konsulado ng Pilipinas sa Dubai ang publiko, partikular ang mga donor, laban sa mga scam na gumagamit ng mga charitable organization.'} ``` I don't have experience with open source pull requests, so I suggest that you reflect them in the source. Thank you for reading :)
2021-04-29T09:32:03Z
https://github.com/huggingface/datasets/issues/2262
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2262/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2261/comments
https://api.github.com/repos/huggingface/datasets/issues/2261/timeline
2021-05-17T16:48:57Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIyODIxNzQw
closed
[]
false
2,261
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Improve ReadInstruction logic and update docs
https://api.github.com/repos/huggingface/datasets/issues/2261/events
null
https://api.github.com/repos/huggingface/datasets/issues/2261/labels{/name}
2021-04-25T19:07:26Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2261.diff", "html_url": "https://github.com/huggingface/datasets/pull/2261", "merged_at": "2021-05-17T16:48:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/2261.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2261" }
867,088,818
[]
https://api.github.com/repos/huggingface/datasets/issues/2261
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Improve ReadInstruction logic and docs.
2021-05-17T18:24:44Z
https://github.com/huggingface/datasets/pull/2261
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2261/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2260/comments
https://api.github.com/repos/huggingface/datasets/issues/2260/timeline
2021-05-07T08:36:17Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIyNzMwODYx
closed
[]
false
2,260
{ "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}", "followers_url": "https://api.github.com/users/bhavitvyamalik/followers", "following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}", "gists_url": "https://api.github.com/users/bhavitvyamalik/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bhavitvyamalik", "id": 19718818, "login": "bhavitvyamalik", "node_id": "MDQ6VXNlcjE5NzE4ODE4", "organizations_url": "https://api.github.com/users/bhavitvyamalik/orgs", "received_events_url": "https://api.github.com/users/bhavitvyamalik/received_events", "repos_url": "https://api.github.com/users/bhavitvyamalik/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bhavitvyamalik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhavitvyamalik/subscriptions", "type": "User", "url": "https://api.github.com/users/bhavitvyamalik" }
GooAQ dataset added
https://api.github.com/repos/huggingface/datasets/issues/2260/events
null
https://api.github.com/repos/huggingface/datasets/issues/2260/labels{/name}
2021-04-25T09:26:48Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2260.diff", "html_url": "https://github.com/huggingface/datasets/pull/2260", "merged_at": "2021-05-07T08:36:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/2260.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2260" }
866,961,697
[]
https://api.github.com/repos/huggingface/datasets/issues/2260
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
@lhoestq here the dataset is stored with Git LFS. Should I add option for manual downloading of dataset using `git lfs pull` post repo cloning or can we accommodate this in the current `download_and_extract`?
2021-05-07T08:36:17Z
https://github.com/huggingface/datasets/pull/2260
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2260/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2259/comments
https://api.github.com/repos/huggingface/datasets/issues/2259/timeline
2021-06-28T08:21:27Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIyNjc2ODA0
closed
[]
false
2,259
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Add support for Split.ALL
https://api.github.com/repos/huggingface/datasets/issues/2259/events
null
https://api.github.com/repos/huggingface/datasets/issues/2259/labels{/name}
2021-04-25T01:45:42Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2259.diff", "html_url": "https://github.com/huggingface/datasets/pull/2259", "merged_at": "2021-06-28T08:21:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/2259.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2259" }
866,880,092
[]
https://api.github.com/repos/huggingface/datasets/issues/2259
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
The title says it all.
2021-06-28T08:21:27Z
https://github.com/huggingface/datasets/pull/2259
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2259/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2258/comments
https://api.github.com/repos/huggingface/datasets/issues/2258/timeline
2021-04-26T16:54:04Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIyNjcxNTQy
closed
[]
false
2,258
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Fix incorrect update_metadata_with_features calls in ArrowDataset
https://api.github.com/repos/huggingface/datasets/issues/2258/events
null
https://api.github.com/repos/huggingface/datasets/issues/2258/labels{/name}
2021-04-25T00:48:38Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2258.diff", "html_url": "https://github.com/huggingface/datasets/pull/2258", "merged_at": "2021-04-26T16:54:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/2258.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2258" }
866,870,588
[]
https://api.github.com/repos/huggingface/datasets/issues/2258
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Fixes bugs in the `unpdate_metadata_with_features` calls (caused by changes in #2151)
2021-04-26T17:16:30Z
https://github.com/huggingface/datasets/pull/2258
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2258/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2257/comments
https://api.github.com/repos/huggingface/datasets/issues/2257/timeline
2021-04-27T16:16:32Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIyNTkwMDQw
closed
[]
false
2,257
{ "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}", "followers_url": "https://api.github.com/users/bhavitvyamalik/followers", "following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}", "gists_url": "https://api.github.com/users/bhavitvyamalik/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bhavitvyamalik", "id": 19718818, "login": "bhavitvyamalik", "node_id": "MDQ6VXNlcjE5NzE4ODE4", "organizations_url": "https://api.github.com/users/bhavitvyamalik/orgs", "received_events_url": "https://api.github.com/users/bhavitvyamalik/received_events", "repos_url": "https://api.github.com/users/bhavitvyamalik/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bhavitvyamalik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhavitvyamalik/subscriptions", "type": "User", "url": "https://api.github.com/users/bhavitvyamalik" }
added metrics for CUAD
https://api.github.com/repos/huggingface/datasets/issues/2257/events
null
https://api.github.com/repos/huggingface/datasets/issues/2257/labels{/name}
2021-04-24T14:09:54Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2257.diff", "html_url": "https://github.com/huggingface/datasets/pull/2257", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2257.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2257" }
866,755,203
[]
https://api.github.com/repos/huggingface/datasets/issues/2257
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
For now I've added F1, AUPR, Precision at 80% recall, and Precision at 90%. Last 3 metrics were reported in the [paper](https://arxiv.org/pdf/2103.06268.pdf). Please let me know if we require `exact_match` metric too here
2021-04-29T09:53:38Z
https://github.com/huggingface/datasets/pull/2257
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2257/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2256/comments
https://api.github.com/repos/huggingface/datasets/issues/2256/timeline
2021-04-26T17:12:15Z
null
completed
MDU6SXNzdWU4NjY3MDg2MDk=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
2,256
{ "avatar_url": "https://avatars.githubusercontent.com/u/8143425?v=4", "events_url": "https://api.github.com/users/roskoN/events{/privacy}", "followers_url": "https://api.github.com/users/roskoN/followers", "following_url": "https://api.github.com/users/roskoN/following{/other_user}", "gists_url": "https://api.github.com/users/roskoN/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/roskoN", "id": 8143425, "login": "roskoN", "node_id": "MDQ6VXNlcjgxNDM0MjU=", "organizations_url": "https://api.github.com/users/roskoN/orgs", "received_events_url": "https://api.github.com/users/roskoN/received_events", "repos_url": "https://api.github.com/users/roskoN/repos", "site_admin": false, "starred_url": "https://api.github.com/users/roskoN/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/roskoN/subscriptions", "type": "User", "url": "https://api.github.com/users/roskoN" }
Running `datase.map` with `num_proc > 1` uses a lot of memory
https://api.github.com/repos/huggingface/datasets/issues/2256/events
null
https://api.github.com/repos/huggingface/datasets/issues/2256/labels{/name}
2021-04-24T09:56:20Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
866,708,609
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2256
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug Running `datase.map` with `num_proc > 1` leads to a tremendous memory usage that requires swapping on disk and it becomes very slow. ## Steps to reproduce the bug ```python from datasets import load_dataset dstc8_datset = load_dataset("roskoN/dstc8-reddit-corpus", keep_in_memory=False) def _prepare_sample(batch): return {"input_ids": list(), "attention_mask": list()} for split_name, dataset_split in list(dstc8_datset.items()): print(f"Processing {split_name}") encoded_dataset_split = dataset_split.map( function=_prepare_sample, batched=True, num_proc=4, remove_columns=dataset_split.column_names, batch_size=10, writer_batch_size=10, keep_in_memory=False, ) print(encoded_dataset_split) path = f"./data/encoded_{split_name}" encoded_dataset_split.save_to_disk(path) ``` ## Expected results Memory usage should stay within reasonable boundaries. ## Actual results This is htop-output from running the provided script. ![image](https://user-images.githubusercontent.com/8143425/115954836-66954980-a4f3-11eb-8340-0153bdc3a475.png) ## Versions ``` - Datasets: 1.6.0 - Python: 3.8.8 (default, Apr 13 2021, 19:58:26) [GCC 7.3.0] - Platform: Linux-4.19.128-microsoft-standard-x86_64-with-glibc2.10 ``` Running on WSL2
2021-04-26T17:12:15Z
https://github.com/huggingface/datasets/issues/2256
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2256/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2255/comments
https://api.github.com/repos/huggingface/datasets/issues/2255/timeline
2021-05-18T13:31:35Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIyMTc0Njg4
closed
[]
false
2,255
{ "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "events_url": "https://api.github.com/users/SBrandeis/events{/privacy}", "followers_url": "https://api.github.com/users/SBrandeis/followers", "following_url": "https://api.github.com/users/SBrandeis/following{/other_user}", "gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SBrandeis", "id": 33657802, "login": "SBrandeis", "node_id": "MDQ6VXNlcjMzNjU3ODAy", "organizations_url": "https://api.github.com/users/SBrandeis/orgs", "received_events_url": "https://api.github.com/users/SBrandeis/received_events", "repos_url": "https://api.github.com/users/SBrandeis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions", "type": "User", "url": "https://api.github.com/users/SBrandeis" }
Task casting for text classification & question answering
https://api.github.com/repos/huggingface/datasets/issues/2255/events
null
https://api.github.com/repos/huggingface/datasets/issues/2255/labels{/name}
2021-04-23T16:00:41Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2255.diff", "html_url": "https://github.com/huggingface/datasets/pull/2255", "merged_at": "2021-05-18T13:31:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/2255.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2255" }
866,242,892
[]
https://api.github.com/repos/huggingface/datasets/issues/2255
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This PR implements task preparation for a given task, in the continuation of #2143 Task taxonomy follows πŸ€— Transformers's pipelines taxonomy: https://github.com/huggingface/transformers/tree/master/src/transformers/pipelines Edit by @lewtun: This PR implements support for the following tasks: * `text-classification` * `question-answering` The intended usage is as follows: ```python # Load a dataset with default column names / features ds = load_dataset("dataset_name") # Cast column names / features to schema. Casting is defined in the dataset's `DatasetInfo` ds = ds.prepare_for_task(task="text-classification") # Casting can also be realised during load ds = load_dataset("dataset_name", task="text-classification") # We can also combine shared tasks across dataset concatenation ds1 = load_dataset("dataset_name_1", task="text-classification") ds2 = load_dataset("dataset_name_2", task="text-classification") # If the tasks have the same schema, so will `ds_concat` ds_concat = concatenate_datasets([ds1, ds2]) ``` Note that the current implementation assumes that `DatasetInfo.task_templates` has been pre-defined by the user / contributor when overriding the `MyDataset(GeneratorBasedBuilder)._info` function. As pointed out by @SBrandeis, for evaluation we'll need a way to detect which datasets are already have a compatible schema so we don't have to edit hundreds of dataset scripts. One possibility is to check if the schema features are a subset of the dataset ones, e.g. ```python squad = load_dataset("./datasets/squad", split="train") qa = QuestionAnswering() schema = Features({**qa.input_schema, **qa.label_schema}) assert all(item in squad.features.items() for item in schema.items()) ```
2021-05-18T13:31:36Z
https://github.com/huggingface/datasets/pull/2255
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2255/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2254/comments
https://api.github.com/repos/huggingface/datasets/issues/2254/timeline
2021-04-27T16:30:48Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIyMTE1NDI0
closed
[]
false
2,254
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Update format, fingerprint and indices after add_item
https://api.github.com/repos/huggingface/datasets/issues/2254/events
null
https://api.github.com/repos/huggingface/datasets/issues/2254/labels{/name}
2021-04-23T14:31:49Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2254.diff", "html_url": "https://github.com/huggingface/datasets/pull/2254", "merged_at": "2021-04-27T16:30:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/2254.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2254" }
866,169,312
[]
https://api.github.com/repos/huggingface/datasets/issues/2254
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Added fingerprint and format update wrappers + update the indices by adding the index of the newly added item in the table.
2021-04-27T16:30:49Z
https://github.com/huggingface/datasets/pull/2254
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2254/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2253/comments
https://api.github.com/repos/huggingface/datasets/issues/2253/timeline
2021-04-27T15:02:59Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIyMDA2Njg3
closed
[]
false
2,253
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Perform minor refactoring: use config
https://api.github.com/repos/huggingface/datasets/issues/2253/events
null
https://api.github.com/repos/huggingface/datasets/issues/2253/labels{/name}
2021-04-23T11:45:47Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2253.diff", "html_url": "https://github.com/huggingface/datasets/pull/2253", "merged_at": "2021-04-27T15:02:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/2253.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2253" }
866,034,321
[ { "color": "B67A40", "default": false, "description": "Restructuring existing code without changing its external behavior", "id": 2851292821, "name": "refactoring", "node_id": "MDU6TGFiZWwyODUxMjkyODIx", "url": "https://api.github.com/repos/huggingface/datasets/labels/refactoring" } ]
https://api.github.com/repos/huggingface/datasets/issues/2253
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Perform minor refactoring related to `config`.
2021-05-27T09:12:45Z
https://github.com/huggingface/datasets/pull/2253
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2253/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2252/comments
https://api.github.com/repos/huggingface/datasets/issues/2252/timeline
2024-01-26T15:10:28Z
null
completed
MDU6SXNzdWU4NjU4NzA3MTA=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
2,252
{ "avatar_url": "https://avatars.githubusercontent.com/u/29157715?v=4", "events_url": "https://api.github.com/users/hwijeen/events{/privacy}", "followers_url": "https://api.github.com/users/hwijeen/followers", "following_url": "https://api.github.com/users/hwijeen/following{/other_user}", "gists_url": "https://api.github.com/users/hwijeen/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hwijeen", "id": 29157715, "login": "hwijeen", "node_id": "MDQ6VXNlcjI5MTU3NzE1", "organizations_url": "https://api.github.com/users/hwijeen/orgs", "received_events_url": "https://api.github.com/users/hwijeen/received_events", "repos_url": "https://api.github.com/users/hwijeen/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hwijeen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hwijeen/subscriptions", "type": "User", "url": "https://api.github.com/users/hwijeen" }
Slow dataloading with big datasets issue persists
https://api.github.com/repos/huggingface/datasets/issues/2252/events
null
https://api.github.com/repos/huggingface/datasets/issues/2252/labels{/name}
2021-04-23T08:18:20Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
865,870,710
[]
https://api.github.com/repos/huggingface/datasets/issues/2252
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Hi, I reported too slow data fetching when data is large(#2210) a couple of weeks ago, and @lhoestq referred me to the fix (#2122). However, the problem seems to persist. Here is the profiled results: 1) Running with 60GB ``` Action | Mean duration (s) |Num calls | Total time (s) | Percentage % | ------------------------------------------------------------------------------------------------------------------------------------ Total | - |_ | 517.96 | 100 % | ------------------------------------------------------------------------------------------------------------------------------------ model_backward | 0.26144 |100 | 26.144 | 5.0475 | model_forward | 0.11123 |100 | 11.123 | 2.1474 | get_train_batch | 0.097121 |100 | 9.7121 | 1.8751 | ``` 3) Running with 600GB, datasets==1.6.0 ``` Action | Mean duration (s) |Num calls | Total time (s) | Percentage % | ------------------------------------------------------------------------------------------------------------------------------------ Total | - |_ | 4563.2 | 100 % | ------------------------------------------------------------------------------------------------------------------------------------ get_train_batch | 5.1279 |100 | 512.79 | 11.237 | model_backward | 4.8394 |100 | 483.94 | 10.605 | model_forward | 0.12162 |100 | 12.162 | 0.26652 | ``` I see that `get_train_batch` lags when data is large. Could this be related to different issues? I would be happy to provide necessary information to investigate.
2024-01-26T15:10:28Z
https://github.com/huggingface/datasets/issues/2252
{ "+1": 5, "-1": 0, "confused": 0, "eyes": 4, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 9, "url": "https://api.github.com/repos/huggingface/datasets/issues/2252/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2251/comments
https://api.github.com/repos/huggingface/datasets/issues/2251/timeline
null
null
null
MDU6SXNzdWU4NjU4NDg3MDU=
open
[]
null
2,251
{ "avatar_url": "https://avatars.githubusercontent.com/u/44570724?v=4", "events_url": "https://api.github.com/users/nlee0212/events{/privacy}", "followers_url": "https://api.github.com/users/nlee0212/followers", "following_url": "https://api.github.com/users/nlee0212/following{/other_user}", "gists_url": "https://api.github.com/users/nlee0212/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nlee0212", "id": 44570724, "login": "nlee0212", "node_id": "MDQ6VXNlcjQ0NTcwNzI0", "organizations_url": "https://api.github.com/users/nlee0212/orgs", "received_events_url": "https://api.github.com/users/nlee0212/received_events", "repos_url": "https://api.github.com/users/nlee0212/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nlee0212/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nlee0212/subscriptions", "type": "User", "url": "https://api.github.com/users/nlee0212" }
while running run_qa.py, ran into a value error
https://api.github.com/repos/huggingface/datasets/issues/2251/events
null
https://api.github.com/repos/huggingface/datasets/issues/2251/labels{/name}
2021-04-23T07:51:03Z
null
false
null
null
865,848,705
[]
https://api.github.com/repos/huggingface/datasets/issues/2251
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
command: python3 run_qa.py --model_name_or_path hyunwoongko/kobart --dataset_name squad_kor_v2 --do_train --do_eval --per_device_train_batch_size 8 --learning_rate 3e-5 --num_train_epochs 3 --max_seq_length 512 --doc_stride 128 --output_dir /tmp/debug_squad/ error: ValueError: External features info don't match the dataset: Got {'id': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'context': Value(dtype='string', id=None), 'question': Value(dtype='string', id=None), 'answer': {'text': Value(dtype='string', id=None), 'answer_start': Value(dtype='int32', id=None), 'html_answer_start': Value(dtype='int32', id=None)}, 'url': Value(dtype='string', id=None), 'raw_html': Value(dtype='string', id=None)} with type struct<answer: struct<text: string, answer_start: int32, html_answer_start: int32>, context: string, id: string, question: string, raw_html: string, title: string, url: string> but expected something like {'answer': {'answer_start': Value(dtype='int32', id=None), 'html_answer_start': Value(dtype='int32', id=None), 'text': Value(dtype='string', id=None)}, 'context': Value(dtype='string', id=None), 'id': Value(dtype='string', id=None), 'question': Value(dtype='string', id=None), 'raw_html': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'url': Value(dtype='string', id=None)} with type struct<answer: struct<answer_start: int32, html_answer_start: int32, text: string>, context: string, id: string, question: string, raw_html: string, title: string, url: string> I didn't encounter this error 4 hours ago. any solutions for this kind of issue? looks like gained dataset format refers to 'Data Fields', while expected refers to 'Data Instances'.
2021-04-23T07:51:03Z
https://github.com/huggingface/datasets/issues/2251
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2251/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2250/comments
https://api.github.com/repos/huggingface/datasets/issues/2250/timeline
2022-03-30T08:29:47Z
null
completed
MDU6SXNzdWU4NjU0MDI0NDk=
closed
[]
null
2,250
{ "avatar_url": "https://avatars.githubusercontent.com/u/14968123?v=4", "events_url": "https://api.github.com/users/alighofrani95/events{/privacy}", "followers_url": "https://api.github.com/users/alighofrani95/followers", "following_url": "https://api.github.com/users/alighofrani95/following{/other_user}", "gists_url": "https://api.github.com/users/alighofrani95/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alighofrani95", "id": 14968123, "login": "alighofrani95", "node_id": "MDQ6VXNlcjE0OTY4MTIz", "organizations_url": "https://api.github.com/users/alighofrani95/orgs", "received_events_url": "https://api.github.com/users/alighofrani95/received_events", "repos_url": "https://api.github.com/users/alighofrani95/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alighofrani95/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alighofrani95/subscriptions", "type": "User", "url": "https://api.github.com/users/alighofrani95" }
some issue in loading local txt file as Dataset for run_mlm.py
https://api.github.com/repos/huggingface/datasets/issues/2250/events
null
https://api.github.com/repos/huggingface/datasets/issues/2250/labels{/name}
2021-04-22T19:39:13Z
null
false
null
null
865,402,449
[]
https://api.github.com/repos/huggingface/datasets/issues/2250
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
![image](https://user-images.githubusercontent.com/14968123/115773877-18cef300-a3c6-11eb-8e58-a9cbfd1001ec.png) first of all, I tried to load 3 .txt files as a dataset (sure that the directory and permission is OK.), I face with the below error. > FileNotFoundError: [Errno 2] No such file or directory: 'c' by removing one of the training .txt files It's fixed and although if I put all file as training it's ok ![image](https://user-images.githubusercontent.com/14968123/115774207-867b1f00-a3c6-11eb-953b-905cfb112d25.png) ![image](https://user-images.githubusercontent.com/14968123/115774264-9b57b280-a3c6-11eb-9f36-7b109f0e5a31.png) after this, my question is how could I use this defined Dataset for run_mlm.py for from scratch pretraining. by using --train_file path_to_train_file just can use one .txt , .csv or, .json file. I tried to set my defined Dataset as --dataset_name but the below issue occurs. > Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/datasets/load.py", line 336, in prepare_module local_path = cached_path(file_path, download_config=download_config) File "/usr/local/lib/python3.7/dist-packages/datasets/utils/file_utils.py", line 291, in cached_path use_auth_token=download_config.use_auth_token, File "/usr/local/lib/python3.7/dist-packages/datasets/utils/file_utils.py", line 621, in get_from_cache raise FileNotFoundError("Couldn't find file at {}".format(url)) FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/master/datasets/dataset/dataset.py > During handling of the above exception, another exception occurred: > Traceback (most recent call last): File "run_mlm.py", line 486, in <module> main() File "run_mlm.py", line 242, in main datasets = load_dataset(data_args.dataset_name, data_args.dataset_config_name, cache_dir=model_args.cache_dir) File "/usr/local/lib/python3.7/dist-packages/datasets/load.py", line 719, in load_dataset use_auth_token=use_auth_token, File "/usr/local/lib/python3.7/dist-packages/datasets/load.py", line 347, in prepare_module combined_path, github_file_path FileNotFoundError: Couldn't find file locally at dataset/dataset.py, or remotely at https://raw.githubusercontent.com/huggingface/datasets/1.6.0/datasets/dataset/dataset.py. The file is also not present on the master branch on github.
2022-03-30T08:29:47Z
https://github.com/huggingface/datasets/issues/2250
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2250/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2249/comments
https://api.github.com/repos/huggingface/datasets/issues/2249/timeline
null
null
null
MDExOlB1bGxSZXF1ZXN0NjIxMzU1MzE3
open
[]
false
2,249
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Allow downloading/processing/caching only specific splits
https://api.github.com/repos/huggingface/datasets/issues/2249/events
{ "closed_at": null, "closed_issues": 2, "created_at": "2021-07-21T15:34:56Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-08-30T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/8", "id": 6968069, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/8/labels", "node_id": "MI_kwDODunzps4AalMF", "number": 8, "open_issues": 4, "state": "open", "title": "1.12", "updated_at": "2021-10-13T10:26:33Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/8" }
https://api.github.com/repos/huggingface/datasets/issues/2249/labels{/name}
2021-04-22T17:51:44Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2249.diff", "html_url": "https://github.com/huggingface/datasets/pull/2249", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2249.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2249" }
865,257,826
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2249
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Allow downloading/processing/caching only specific splits without downloading/processing/caching the other splits. This PR implements two steps to handle only specific splits: - it allows processing/caching only specific splits into Arrow files - for some simple cases, it allows downloading only specific splits (which is more intricate as it depends on the user-defined method `_split_generators`) This PR makes several assumptions: - `DownloadConfig` contains the configuration settings for downloading - the parameter `split` passed to `load_dataset` is just a parameter for loading (from cache), not for downloading
2022-07-06T15:19:48Z
https://github.com/huggingface/datasets/pull/2249
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2249/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2248/comments
https://api.github.com/repos/huggingface/datasets/issues/2248/timeline
2021-04-27T15:29:20Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIxMDEyNzg5
closed
[]
false
2,248
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Implement Dataset to JSON
https://api.github.com/repos/huggingface/datasets/issues/2248/events
{ "closed_at": "2021-05-31T16:20:53Z", "closed_issues": 3, "created_at": "2021-04-09T13:16:31Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-05-14T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/3", "id": 6644287, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/3/labels", "node_id": "MDk6TWlsZXN0b25lNjY0NDI4Nw==", "number": 3, "open_issues": 0, "state": "closed", "title": "1.7", "updated_at": "2021-05-31T16:20:53Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/3" }
https://api.github.com/repos/huggingface/datasets/issues/2248/labels{/name}
2021-04-22T11:46:51Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2248.diff", "html_url": "https://github.com/huggingface/datasets/pull/2248", "merged_at": "2021-04-27T15:29:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/2248.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2248" }
864,853,447
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2248
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Implement `Dataset.to_json`.
2021-04-27T15:29:21Z
https://github.com/huggingface/datasets/pull/2248
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2248/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2247/comments
https://api.github.com/repos/huggingface/datasets/issues/2247/timeline
2021-07-26T13:28:51Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIwOTgzNzY3
closed
[]
false
2,247
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Implement Dataset from Parquet
https://api.github.com/repos/huggingface/datasets/issues/2247/events
{ "closed_at": "2021-09-02T05:34:03Z", "closed_issues": 2, "created_at": "2021-07-09T05:49:00Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-07-30T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/7", "id": 6931350, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/7/labels", "node_id": "MDk6TWlsZXN0b25lNjkzMTM1MA==", "number": 7, "open_issues": 0, "state": "closed", "title": "1.11", "updated_at": "2021-09-02T05:34:03Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/7" }
https://api.github.com/repos/huggingface/datasets/issues/2247/labels{/name}
2021-04-22T11:01:38Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2247.diff", "html_url": "https://github.com/huggingface/datasets/pull/2247", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2247.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2247" }
864,817,520
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2247
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Implement instantiation of Dataset from Parquet file.
2021-07-26T13:28:52Z
https://github.com/huggingface/datasets/pull/2247
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2247/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2246/comments
https://api.github.com/repos/huggingface/datasets/issues/2246/timeline
2021-04-26T16:13:59Z
null
null
MDExOlB1bGxSZXF1ZXN0NjIwNDg3OTUw
closed
[]
false
2,246
{ "avatar_url": "https://avatars.githubusercontent.com/u/39116809?v=4", "events_url": "https://api.github.com/users/norabelrose/events{/privacy}", "followers_url": "https://api.github.com/users/norabelrose/followers", "following_url": "https://api.github.com/users/norabelrose/following{/other_user}", "gists_url": "https://api.github.com/users/norabelrose/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/norabelrose", "id": 39116809, "login": "norabelrose", "node_id": "MDQ6VXNlcjM5MTE2ODA5", "organizations_url": "https://api.github.com/users/norabelrose/orgs", "received_events_url": "https://api.github.com/users/norabelrose/received_events", "repos_url": "https://api.github.com/users/norabelrose/repos", "site_admin": false, "starred_url": "https://api.github.com/users/norabelrose/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/norabelrose/subscriptions", "type": "User", "url": "https://api.github.com/users/norabelrose" }
Faster map w/ input_columns & faster slicing w/ Iterable keys
https://api.github.com/repos/huggingface/datasets/issues/2246/events
null
https://api.github.com/repos/huggingface/datasets/issues/2246/labels{/name}
2021-04-21T19:49:07Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2246.diff", "html_url": "https://github.com/huggingface/datasets/pull/2246", "merged_at": "2021-04-26T16:13:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/2246.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2246" }
864,220,031
[]
https://api.github.com/repos/huggingface/datasets/issues/2246
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
@lhoestq Fixes #2193 - `map` now uses `with_format` to only load needed columns in memory when `input_columns` is set - Slicing datasets with Iterables of indices now uses a new `Table.fast_gather` method, implemented with `np.searchsorted`, to find the appropriate batch indices all at once. `pa.concat_tables` is no longer used for this; we just call `pa.Table.from_batches` with a list of all the batch slices. Together these changes have sped up batched `map()` calls over subsets of columns quite considerably in my initial testing.
2021-04-26T16:13:59Z
https://github.com/huggingface/datasets/pull/2246
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2246/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2245/comments
https://api.github.com/repos/huggingface/datasets/issues/2245/timeline
2021-05-10T17:31:22Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE5NjQzMjQ3
closed
[]
false
2,245
{ "avatar_url": "https://avatars.githubusercontent.com/u/42388668?v=4", "events_url": "https://api.github.com/users/NikhilBartwal/events{/privacy}", "followers_url": "https://api.github.com/users/NikhilBartwal/followers", "following_url": "https://api.github.com/users/NikhilBartwal/following{/other_user}", "gists_url": "https://api.github.com/users/NikhilBartwal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NikhilBartwal", "id": 42388668, "login": "NikhilBartwal", "node_id": "MDQ6VXNlcjQyMzg4NjY4", "organizations_url": "https://api.github.com/users/NikhilBartwal/orgs", "received_events_url": "https://api.github.com/users/NikhilBartwal/received_events", "repos_url": "https://api.github.com/users/NikhilBartwal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NikhilBartwal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NikhilBartwal/subscriptions", "type": "User", "url": "https://api.github.com/users/NikhilBartwal" }
Add `key` type and duplicates verification with hashing
https://api.github.com/repos/huggingface/datasets/issues/2245/events
null
https://api.github.com/repos/huggingface/datasets/issues/2245/labels{/name}
2021-04-20T20:03:19Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2245.diff", "html_url": "https://github.com/huggingface/datasets/pull/2245", "merged_at": "2021-05-10T17:31:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/2245.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2245" }
863,191,655
[]
https://api.github.com/repos/huggingface/datasets/issues/2245
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Closes #2230 There is currently no verification for the data type and the uniqueness of the keys yielded by the `dataset_builder`. This PR is currently a work in progress with the following goals: - [x] Adding `hash_salt` to `ArrowWriter` so that the keys belonging to different splits have different hash - [x] Add `key` arrtibute to `ArrowWriter.write()` for hashing - [x] Add a hashing class which takes an input key of certain type (`str`/`int`/anything convertible to string) and produces a 128-bit hash using `hashlib.md5` - [x] Creating a function giving a custom error message when non-unique keys are found **[This will take care of type-checking for keys]** - [x] Checking for duplicate keys in `writer.write()` for each batch [**NOTE**: This PR is currently concerned with `GeneratorBasedBuilder` only, for simplification. A subsequent PR will be made in future for `ArrowBasedBuilder`] @lhoestq Thank you for the feedback. It would be great to have your guidance on this!
2021-05-10T18:04:37Z
https://github.com/huggingface/datasets/pull/2245
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2245/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2244/comments
https://api.github.com/repos/huggingface/datasets/issues/2244/timeline
null
null
null
MDExOlB1bGxSZXF1ZXN0NjE5NTAyODc0
open
[]
false
2,244
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Set specific cache directories per test function call
https://api.github.com/repos/huggingface/datasets/issues/2244/events
{ "closed_at": null, "closed_issues": 2, "created_at": "2021-07-21T15:34:56Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }, "description": "Next minor release", "due_on": "2021-08-30T07:00:00Z", "html_url": "https://github.com/huggingface/datasets/milestone/8", "id": 6968069, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/8/labels", "node_id": "MI_kwDODunzps4AalMF", "number": 8, "open_issues": 4, "state": "open", "title": "1.12", "updated_at": "2021-10-13T10:26:33Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/8" }
https://api.github.com/repos/huggingface/datasets/issues/2244/labels{/name}
2021-04-20T17:06:22Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2244.diff", "html_url": "https://github.com/huggingface/datasets/pull/2244", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2244.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2244" }
863,029,946
[]
https://api.github.com/repos/huggingface/datasets/issues/2244
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Implement specific cache directories (datasets, metrics and modules) per test function call. Currently, the cache directories are set within the temporary test directory, but they are shared across all test function calls. This PR implements specific cache directories for each test function call, so that tests are atomic and there are no side effects.
2022-07-06T15:19:48Z
https://github.com/huggingface/datasets/pull/2244
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2244/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2243/comments
https://api.github.com/repos/huggingface/datasets/issues/2243/timeline
2021-05-03T17:54:32Z
null
completed
MDU6SXNzdWU4NjI5MDkzODk=
closed
[]
null
2,243
{ "avatar_url": "https://avatars.githubusercontent.com/u/2743060?v=4", "events_url": "https://api.github.com/users/villmow/events{/privacy}", "followers_url": "https://api.github.com/users/villmow/followers", "following_url": "https://api.github.com/users/villmow/following{/other_user}", "gists_url": "https://api.github.com/users/villmow/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/villmow", "id": 2743060, "login": "villmow", "node_id": "MDQ6VXNlcjI3NDMwNjA=", "organizations_url": "https://api.github.com/users/villmow/orgs", "received_events_url": "https://api.github.com/users/villmow/received_events", "repos_url": "https://api.github.com/users/villmow/repos", "site_admin": false, "starred_url": "https://api.github.com/users/villmow/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/villmow/subscriptions", "type": "User", "url": "https://api.github.com/users/villmow" }
Map is slow and processes batches one after another
https://api.github.com/repos/huggingface/datasets/issues/2243/events
null
https://api.github.com/repos/huggingface/datasets/issues/2243/labels{/name}
2021-04-20T14:58:20Z
null
false
null
null
862,909,389
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2243
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Describe the bug I have a somewhat unclear bug to me, where I can't figure out what the problem is. The code works as expected on a small subset of my dataset (2000 samples) on my local machine, but when I execute the same code with a larger dataset (1.4 million samples) this problem occurs. Thats why I can't give exact steps to reproduce, I'm sorry. I process a large dataset in a two step process. I first call map on a dataset I load from disk and create a new dataset from it. This works like expected and `map` uses all workers I started it with. Then I process the dataset created by the first step, again with `map`, which is really slow and starting only one or two process at a time. Number of processes is the same for both steps. pseudo code: ```python ds = datasets.load_from_disk("path") new_dataset = ds.map(work, batched=True, ...) # fast uses all processes final_dataset = new_dataset.map(work2, batched=True, ...) # slow starts one process after another ``` ## Expected results Second stage should be as fast as the first stage. ## Versions Paste the output of the following code: - Datasets: 1.5.0 - Python: 3.8.8 (default, Feb 24 2021, 21:46:12) - Platform: Linux-5.4.0-60-generic-x86_64-with-glibc2.10 Do you guys have any idea? Thanks a lot!
2021-05-03T17:54:33Z
https://github.com/huggingface/datasets/issues/2243
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2243/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2242/comments
https://api.github.com/repos/huggingface/datasets/issues/2242/timeline
2021-04-20T15:02:45Z
null
completed
MDU6SXNzdWU4NjI4NzAyMDU=
closed
[]
null
2,242
{ "avatar_url": "https://avatars.githubusercontent.com/u/6735707?v=4", "events_url": "https://api.github.com/users/martavillegas/events{/privacy}", "followers_url": "https://api.github.com/users/martavillegas/followers", "following_url": "https://api.github.com/users/martavillegas/following{/other_user}", "gists_url": "https://api.github.com/users/martavillegas/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/martavillegas", "id": 6735707, "login": "martavillegas", "node_id": "MDQ6VXNlcjY3MzU3MDc=", "organizations_url": "https://api.github.com/users/martavillegas/orgs", "received_events_url": "https://api.github.com/users/martavillegas/received_events", "repos_url": "https://api.github.com/users/martavillegas/repos", "site_admin": false, "starred_url": "https://api.github.com/users/martavillegas/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/martavillegas/subscriptions", "type": "User", "url": "https://api.github.com/users/martavillegas" }
Link to datasets viwer on Quick Tour page returns "502 Bad Gateway"
https://api.github.com/repos/huggingface/datasets/issues/2242/events
null
https://api.github.com/repos/huggingface/datasets/issues/2242/labels{/name}
2021-04-20T14:19:51Z
null
false
null
null
862,870,205
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2242
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Link to datasets viwer (https://huggingface.co/datasets/viewer/) on Quick Tour page (https://huggingface.co/docs/datasets/quicktour.html) returns "502 Bad Gateway" The same error with https://huggingface.co/datasets/viewer/?dataset=glue&config=mrpc
2021-04-20T15:02:45Z
https://github.com/huggingface/datasets/issues/2242
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2242/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2241/comments
https://api.github.com/repos/huggingface/datasets/issues/2241/timeline
2021-04-23T15:36:15Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE5MjI0MzIw
closed
[]
false
2,241
{ "avatar_url": "https://avatars.githubusercontent.com/u/7669893?v=4", "events_url": "https://api.github.com/users/cahya-wirawan/events{/privacy}", "followers_url": "https://api.github.com/users/cahya-wirawan/followers", "following_url": "https://api.github.com/users/cahya-wirawan/following{/other_user}", "gists_url": "https://api.github.com/users/cahya-wirawan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cahya-wirawan", "id": 7669893, "login": "cahya-wirawan", "node_id": "MDQ6VXNlcjc2Njk4OTM=", "organizations_url": "https://api.github.com/users/cahya-wirawan/orgs", "received_events_url": "https://api.github.com/users/cahya-wirawan/received_events", "repos_url": "https://api.github.com/users/cahya-wirawan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cahya-wirawan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cahya-wirawan/subscriptions", "type": "User", "url": "https://api.github.com/users/cahya-wirawan" }
Add SLR32 to OpenSLR
https://api.github.com/repos/huggingface/datasets/issues/2241/events
null
https://api.github.com/repos/huggingface/datasets/issues/2241/labels{/name}
2021-04-20T11:02:45Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2241.diff", "html_url": "https://github.com/huggingface/datasets/pull/2241", "merged_at": "2021-04-23T15:36:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/2241.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2241" }
862,696,460
[]
https://api.github.com/repos/huggingface/datasets/issues/2241
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
I would like to add SLR32 to OpenSLR. It contains four South African languages: Afrikaans, Sesotho, Setswana and isiXhosa
2021-04-23T16:21:24Z
https://github.com/huggingface/datasets/pull/2241
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2241/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2240/comments
https://api.github.com/repos/huggingface/datasets/issues/2240/timeline
2021-04-21T09:54:57Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE5MDkyODc5
closed
[]
false
2,240
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Clarify how to load wikihow
https://api.github.com/repos/huggingface/datasets/issues/2240/events
null
https://api.github.com/repos/huggingface/datasets/issues/2240/labels{/name}
2021-04-20T08:02:58Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2240.diff", "html_url": "https://github.com/huggingface/datasets/pull/2240", "merged_at": "2021-04-21T09:54:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/2240.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2240" }
862,537,856
[]
https://api.github.com/repos/huggingface/datasets/issues/2240
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Explain clearer how to load the dataset in the manual download instructions. En relation with #2239.
2021-04-21T09:54:57Z
https://github.com/huggingface/datasets/pull/2240
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2240/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2239/comments
https://api.github.com/repos/huggingface/datasets/issues/2239/timeline
2021-04-20T16:33:11Z
null
completed
MDU6SXNzdWU4NjE5MDQzMDY=
closed
[]
null
2,239
{ "avatar_url": "https://avatars.githubusercontent.com/u/4686956?v=4", "events_url": "https://api.github.com/users/odellus/events{/privacy}", "followers_url": "https://api.github.com/users/odellus/followers", "following_url": "https://api.github.com/users/odellus/following{/other_user}", "gists_url": "https://api.github.com/users/odellus/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/odellus", "id": 4686956, "login": "odellus", "node_id": "MDQ6VXNlcjQ2ODY5NTY=", "organizations_url": "https://api.github.com/users/odellus/orgs", "received_events_url": "https://api.github.com/users/odellus/received_events", "repos_url": "https://api.github.com/users/odellus/repos", "site_admin": false, "starred_url": "https://api.github.com/users/odellus/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/odellus/subscriptions", "type": "User", "url": "https://api.github.com/users/odellus" }
Error loading wikihow dataset
https://api.github.com/repos/huggingface/datasets/issues/2239/events
null
https://api.github.com/repos/huggingface/datasets/issues/2239/labels{/name}
2021-04-19T21:02:31Z
null
false
null
null
861,904,306
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2239
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
## Describe the bug When attempting to load wikihow into a dataset with ```python from datasets import load_dataset dataset = load_dataset('wikihow', data_dir='./wikihow') ``` I get the message: ``` AttributeError: 'BuilderConfig' object has no attribute 'filename' ``` at the end of a [full stack trace](https://gist.github.com/odellus/602c3b2de52f541d353b1022f320ffc2). ## Steps to reproduce the bug I have followed the instructions for creating a wikihow dataset. The [wikihow dataset site](https://huggingface.co/datasets/wikihow) says to use ```python from datasets import load_dataset dataset = load_dataset('wikihow') ``` to load the dataset. I do so and I get the message ``` AssertionError: The dataset wikihow with config all requires manual data. Please follow the manual download instructions: You need to manually download two wikihow files. An overview of which files to download can be seen at https://github.com/mahnazkoupaee/WikiHow-Dataset. You need to download the following two files manually: 1) https://ucsb.app.box.com/s/ap23l8gafpezf4tq3wapr6u8241zz358 and save the file under <path/to/folder>/wikihowAll.csv 2) https://ucsb.app.box.com/s/7yq601ijl1lzvlfu4rjdbbxforzd2oag and save the file under <path/to/folder>/wikihowSep.csv The <path/to/folder> can e.g. be "~/manual_wikihow_data". Wikihow can then be loaded using the following command `datasets.load_dataset("wikihow", data_dir="<path/to/folder>")`. . Manual data can be loaded with `datasets.load_dataset(wikihow, data_dir='<path/to/manual/data>') ``` So I create a directory `./wikihow` and download `wikihowAll.csv` and `wikihowSep.csv` into the new directory. Then I run ```python from datasets import load_dataset dataset = load_dataset('wikihow', data_dir='./wikihow') ``` that's when I get the [stack trace](https://gist.github.com/odellus/602c3b2de52f541d353b1022f320ffc2) ## Expected results I expected it to load the downloaded files into a dataset. ## Actual results ```python Using custom data configuration default-data_dir=.%2Fwikihow Downloading and preparing dataset wikihow/default (download: Unknown size, generated: Unknown size, post-processed: Unknown size, total: Unknown size) to /home/azureuser/.cache/huggingface/datasets/wikihow/default-data_dir=.%2Fwikihow/0.0.0/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2... --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) <ipython-input-9-5e4d40142f30> in <module> ----> 1 dataset = load_dataset('wikihow',data_dir='./wikihow') ~/.local/lib/python3.6/site-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, script_version, use_auth_token, **config_kwargs) 745 try_from_hf_gcs=try_from_hf_gcs, 746 base_path=base_path,--> 747 use_auth_token=use_auth_token, 748 ) 749 ~/.local/lib/python3.6/site-packages/datasets/builder.py in download_and_prepare(self, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, **download_and_prepare_kwargs) 577 if not downloaded_from_gcs: 578 self._download_and_prepare( --> 579 dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs 580 ) 581 # Sync info ~/.local/lib/python3.6/site-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_split_kwargs) 632 split_dict = SplitDict(dataset_name=self.name) 633 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) --> 634 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) 635 636 # Checksums verification ~/.cache/huggingface/modules/datasets_modules/datasets/wikihow/58f42f8f0e4d459811a0f69aaab35870093830ccd58006769e7e1eb3e0e686c2/wikihow.py in _split_generators(self, dl_manager) 132 133 path_to_manual_file = os.path.join( --> 134 os.path.abspath(os.path.expanduser(dl_manager.manual_dir)), self.config.filename 135 ) 136 AttributeError: 'BuilderConfig' object has no attribute 'filename' ``` ## Versions Paste the output of the following code: ```python import datasets import sys import platform print(f""" - Datasets: {datasets.__version__} - Python: {sys.version} - Platform: {platform.platform()} """) ``` ``` - Datasets: 1.5.0 - Python: 3.6.9 (default, Jan 26 2021, 15:33:00) [GCC 8.4.0] - Platform: Linux-5.4.0-1046-azure-x86_64-with-Ubuntu-18.04-bionic ```
2021-04-20T16:33:11Z
https://github.com/huggingface/datasets/issues/2239
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2239/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2238/comments
https://api.github.com/repos/huggingface/datasets/issues/2238/timeline
2021-04-23T15:32:05Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE4MTY5NzM5
closed
[]
false
2,238
{ "avatar_url": "https://avatars.githubusercontent.com/u/32985207?v=4", "events_url": "https://api.github.com/users/dkajtoch/events{/privacy}", "followers_url": "https://api.github.com/users/dkajtoch/followers", "following_url": "https://api.github.com/users/dkajtoch/following{/other_user}", "gists_url": "https://api.github.com/users/dkajtoch/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dkajtoch", "id": 32985207, "login": "dkajtoch", "node_id": "MDQ6VXNlcjMyOTg1MjA3", "organizations_url": "https://api.github.com/users/dkajtoch/orgs", "received_events_url": "https://api.github.com/users/dkajtoch/received_events", "repos_url": "https://api.github.com/users/dkajtoch/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dkajtoch/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dkajtoch/subscriptions", "type": "User", "url": "https://api.github.com/users/dkajtoch" }
NLU evaluation data
https://api.github.com/repos/huggingface/datasets/issues/2238/events
null
https://api.github.com/repos/huggingface/datasets/issues/2238/labels{/name}
2021-04-19T16:47:20Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2238.diff", "html_url": "https://github.com/huggingface/datasets/pull/2238", "merged_at": "2021-04-23T15:32:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/2238.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2238" }
861,518,291
[]
https://api.github.com/repos/huggingface/datasets/issues/2238
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
New intent classification dataset from https://github.com/xliuhw/NLU-Evaluation-Data
2021-04-23T15:32:05Z
https://github.com/huggingface/datasets/pull/2238
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2238/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2237/comments
https://api.github.com/repos/huggingface/datasets/issues/2237/timeline
null
null
null
MDU6SXNzdWU4NjE0Mjc0Mzk=
open
[]
null
2,237
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Update Dataset.dataset_size after transformed with map
https://api.github.com/repos/huggingface/datasets/issues/2237/events
null
https://api.github.com/repos/huggingface/datasets/issues/2237/labels{/name}
2021-04-19T15:19:38Z
null
false
null
null
861,427,439
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2237
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
After loading a dataset, if we transform it by using `.map` its `dataset_size` attirbute is not updated.
2021-04-20T14:22:05Z
https://github.com/huggingface/datasets/issues/2237
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2237/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2236/comments
https://api.github.com/repos/huggingface/datasets/issues/2236/timeline
null
null
null
MDU6SXNzdWU4NjEzODgxNDU=
open
[]
null
2,236
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "https://api.github.com/users/sarahwie/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sarahwie", "id": 8027676, "login": "sarahwie", "node_id": "MDQ6VXNlcjgwMjc2NzY=", "organizations_url": "https://api.github.com/users/sarahwie/orgs", "received_events_url": "https://api.github.com/users/sarahwie/received_events", "repos_url": "https://api.github.com/users/sarahwie/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sarahwie/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sarahwie/subscriptions", "type": "User", "url": "https://api.github.com/users/sarahwie" }
Request to add StrategyQA dataset
https://api.github.com/repos/huggingface/datasets/issues/2236/events
null
https://api.github.com/repos/huggingface/datasets/issues/2236/labels{/name}
2021-04-19T14:46:26Z
null
false
null
null
861,388,145
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
https://api.github.com/repos/huggingface/datasets/issues/2236
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
## Request to add StrategyQA dataset - **Name:** StrategyQA - **Description:** open-domain QA [(project page)](https://allenai.org/data/strategyqa) - **Paper:** [url](https://arxiv.org/pdf/2101.02235.pdf) - **Data:** [here](https://allenai.org/data/strategyqa) - **Motivation:** uniquely-formulated dataset that also includes a question-decomposition breakdown and associated Wikipedia annotations for each step. Good for multi-hop reasoning modeling.
2021-04-19T14:46:26Z
https://github.com/huggingface/datasets/issues/2236
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/2236/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2235/comments
https://api.github.com/repos/huggingface/datasets/issues/2235/timeline
2021-04-19T12:49:19Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE3Nzc0NDUw
closed
[]
false
2,235
{ "avatar_url": "https://avatars.githubusercontent.com/u/22492839?v=4", "events_url": "https://api.github.com/users/PierreColombo/events{/privacy}", "followers_url": "https://api.github.com/users/PierreColombo/followers", "following_url": "https://api.github.com/users/PierreColombo/following{/other_user}", "gists_url": "https://api.github.com/users/PierreColombo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/PierreColombo", "id": 22492839, "login": "PierreColombo", "node_id": "MDQ6VXNlcjIyNDkyODM5", "organizations_url": "https://api.github.com/users/PierreColombo/orgs", "received_events_url": "https://api.github.com/users/PierreColombo/received_events", "repos_url": "https://api.github.com/users/PierreColombo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/PierreColombo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/PierreColombo/subscriptions", "type": "User", "url": "https://api.github.com/users/PierreColombo" }
Update README.md
https://api.github.com/repos/huggingface/datasets/issues/2235/events
null
https://api.github.com/repos/huggingface/datasets/issues/2235/labels{/name}
2021-04-19T08:21:02Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2235.diff", "html_url": "https://github.com/huggingface/datasets/pull/2235", "merged_at": "2021-04-19T12:49:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/2235.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2235" }
861,040,716
[]
https://api.github.com/repos/huggingface/datasets/issues/2235
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Adding relevant citations (paper accepted at AAAI 2020 & EMNLP 2020) to the benchmark
2021-04-19T12:49:19Z
https://github.com/huggingface/datasets/pull/2235
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2235/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2234/comments
https://api.github.com/repos/huggingface/datasets/issues/2234/timeline
2021-04-19T07:51:36Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE3MzI4NDU3
closed
[]
false
2,234
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
Fix bash snippet formatting in ADD_NEW_DATASET.md
https://api.github.com/repos/huggingface/datasets/issues/2234/events
null
https://api.github.com/repos/huggingface/datasets/issues/2234/labels{/name}
2021-04-17T16:01:08Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2234.diff", "html_url": "https://github.com/huggingface/datasets/pull/2234", "merged_at": "2021-04-19T07:51:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/2234.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2234" }
860,442,246
[]
https://api.github.com/repos/huggingface/datasets/issues/2234
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
This PR indents the paragraphs around the bash snippets in ADD_NEW_DATASET.md to fix formatting.
2021-04-19T10:57:31Z
https://github.com/huggingface/datasets/pull/2234
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2234/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2233/comments
https://api.github.com/repos/huggingface/datasets/issues/2233/timeline
2021-04-19T08:56:42Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE3MDYwMTkw
closed
[]
false
2,233
{ "avatar_url": "https://avatars.githubusercontent.com/u/42388668?v=4", "events_url": "https://api.github.com/users/NikhilBartwal/events{/privacy}", "followers_url": "https://api.github.com/users/NikhilBartwal/followers", "following_url": "https://api.github.com/users/NikhilBartwal/following{/other_user}", "gists_url": "https://api.github.com/users/NikhilBartwal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NikhilBartwal", "id": 42388668, "login": "NikhilBartwal", "node_id": "MDQ6VXNlcjQyMzg4NjY4", "organizations_url": "https://api.github.com/users/NikhilBartwal/orgs", "received_events_url": "https://api.github.com/users/NikhilBartwal/received_events", "repos_url": "https://api.github.com/users/NikhilBartwal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NikhilBartwal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NikhilBartwal/subscriptions", "type": "User", "url": "https://api.github.com/users/NikhilBartwal" }
Fix `xnli` dataset tuple key
https://api.github.com/repos/huggingface/datasets/issues/2233/events
null
https://api.github.com/repos/huggingface/datasets/issues/2233/labels{/name}
2021-04-16T19:12:42Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2233.diff", "html_url": "https://github.com/huggingface/datasets/pull/2233", "merged_at": "2021-04-19T08:56:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/2233.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2233" }
860,097,084
[]
https://api.github.com/repos/huggingface/datasets/issues/2233
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Closes #2229 The `xnli` dataset yields a tuple key in case of `ar` which is inconsistant with the acceptable key types (str/int). The key was thus ported to `str` keeping the original information intact.
2021-04-19T08:56:42Z
https://github.com/huggingface/datasets/pull/2233
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2233/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2232/comments
https://api.github.com/repos/huggingface/datasets/issues/2232/timeline
2021-04-21T09:33:08Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE3MDQyNTI4
closed
[]
false
2,232
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Start filling GLUE dataset card
https://api.github.com/repos/huggingface/datasets/issues/2232/events
null
https://api.github.com/repos/huggingface/datasets/issues/2232/labels{/name}
2021-04-16T18:37:37Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2232.diff", "html_url": "https://github.com/huggingface/datasets/pull/2232", "merged_at": "2021-04-21T09:33:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/2232.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2232" }
860,075,931
[]
https://api.github.com/repos/huggingface/datasets/issues/2232
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
The dataset card was pretty much empty. I added the descriptions (mainly from TFDS since the script is the same), and I also added the tasks tags as well as examples for a subset of the tasks. cc @sgugger
2021-04-21T09:33:09Z
https://github.com/huggingface/datasets/pull/2232
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2232/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2231/comments
https://api.github.com/repos/huggingface/datasets/issues/2231/timeline
2021-04-16T15:10:04Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE2ODYyNTEx
closed
[]
false
2,231
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
Fix map when removing columns on a formatted dataset
https://api.github.com/repos/huggingface/datasets/issues/2231/events
null
https://api.github.com/repos/huggingface/datasets/issues/2231/labels{/name}
2021-04-16T14:08:55Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2231.diff", "html_url": "https://github.com/huggingface/datasets/pull/2231", "merged_at": "2021-04-16T15:10:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/2231.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2231" }
859,850,488
[]
https://api.github.com/repos/huggingface/datasets/issues/2231
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
This should fix issue #2226 The `remove_columns` argument was ignored on formatted datasets
2021-04-16T15:10:05Z
https://github.com/huggingface/datasets/pull/2231
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2231/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2230/comments
https://api.github.com/repos/huggingface/datasets/issues/2230/timeline
2021-05-10T17:31:21Z
null
completed
MDU6SXNzdWU4NTk4MTcxNTk=
closed
[]
null
2,230
{ "avatar_url": "https://avatars.githubusercontent.com/u/42388668?v=4", "events_url": "https://api.github.com/users/NikhilBartwal/events{/privacy}", "followers_url": "https://api.github.com/users/NikhilBartwal/followers", "following_url": "https://api.github.com/users/NikhilBartwal/following{/other_user}", "gists_url": "https://api.github.com/users/NikhilBartwal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NikhilBartwal", "id": 42388668, "login": "NikhilBartwal", "node_id": "MDQ6VXNlcjQyMzg4NjY4", "organizations_url": "https://api.github.com/users/NikhilBartwal/orgs", "received_events_url": "https://api.github.com/users/NikhilBartwal/received_events", "repos_url": "https://api.github.com/users/NikhilBartwal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NikhilBartwal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NikhilBartwal/subscriptions", "type": "User", "url": "https://api.github.com/users/NikhilBartwal" }
Keys yielded while generating dataset are not being checked
https://api.github.com/repos/huggingface/datasets/issues/2230/events
null
https://api.github.com/repos/huggingface/datasets/issues/2230/labels{/name}
2021-04-16T13:29:47Z
null
false
null
null
859,817,159
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
https://api.github.com/repos/huggingface/datasets/issues/2230
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
The keys used in the dataset generation script to ensure the same order is generated on every user's end should be checked for their types (i.e either `str` or `int`) as well as whether they are unique or not. Currently, the keys are not being checked for any of these, as evident from `xnli' dataset generation: https://github.com/huggingface/datasets/blob/56346791aed417306d054d89bd693d6b7eab17f7/datasets/xnli/xnli.py#L196 Even after having a tuple as key, the dataset is generated without any warning. Also, as tested in the case of `anli` dataset (I tweeked the dataset script to use `1` as a key for every example): ``` >>> import datasets >>> nik = datasets.load_dataset('anli') Downloading and preparing dataset anli/plain_text (download: 17.76 MiB, generated: 73.55 MiB, post-processed: Unknown size, total: 91.31 MiB) to C:\Users\nikhil\.cache\huggingface\datasets\anli\plain_text\0.1.0\43fa2c99c10bf8478f1fa0860f7b122c6b277c4c41306255b7641257cf4e3299... 0 examples [00:00, ? examples/s]1 {'uid': '0fd0abfb-659e-4453-b196-c3a64d2d8267', 'premise': 'The Parma trolleybus system (Italian: "Rete filoviaria di Parma" ) forms part of the public transport network of the city and "comune" of Parma, in the region of Emilia-Romagna, northern Italy. In operation since 1953, the system presently comprises four urban routes.', 'hypothesis': 'The trolleybus system has over 2 urban routes', 'label': 'entailment', 'reason': ''} 2021-04-16 12:38:14.483968: I tensorflow/stream_executor/platform/default/dso_loader.cc:49] Successfully opened dynamic library cudart64_110.dll 1 examples [00:01, 1.87s/ examples]1 {'uid': '7ed72ff4-40b7-4f8a-b1b9-6c612aa62c84', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 – 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage/science fiction adventure series "The Champions". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': "Sharron Macready was a popular character through the 1980's.", 'label': 'neutral', 'reason': ''} 1 {'uid': '5d2930a3-62ac-485d-94d7-4e36cbbcd7b5', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 – 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage/science fiction adventure series "The Champions". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': "Bastedo didn't keep any pets because of her views on animal rights.", 'label': 'neutral', 'reason': ''} 1 {'uid': '324db753-ddc9-4a85-a825-f09e2e5aebdd', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 – 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage/science fiction adventure series "The Champions". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Alexandra Bastedo was named by her mother.', 'label': 'neutral', 'reason': ''} 1 {'uid': '4874f429-da0e-406a-90c7-22240ff3ddf8', 'premise': 'Alexandra Lendon Bastedo (9 March 1946 – 12 January 2014) was a British actress, best known for her role as secret agent Sharron Macready in the 1968 British espionage/science fiction adventure series "The Champions". She has been cited as a sex symbol of the 1960s and 1970s. Bastedo was a vegetarian and animal welfare advocate.', 'hypothesis': 'Bastedo cared for all the animals that inhabit the earth.', 'label': 'neutral', 'reason': ''} ``` Here also, the dataset was generated successfuly even hough it had same keys without any warning. The reason appears to stem from here: https://github.com/huggingface/datasets/blob/56346791aed417306d054d89bd693d6b7eab17f7/src/datasets/builder.py#L988 Here, although it has access to every key, but it is not being checked and the example is written directly: https://github.com/huggingface/datasets/blob/56346791aed417306d054d89bd693d6b7eab17f7/src/datasets/builder.py#L992 I would like to take this issue if you allow me. Thank You!
2021-05-10T17:31:21Z
https://github.com/huggingface/datasets/issues/2230
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2230/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2229/comments
https://api.github.com/repos/huggingface/datasets/issues/2229/timeline
2021-04-19T08:56:42Z
null
completed
MDU6SXNzdWU4NTk4MTA2MDI=
closed
[]
null
2,229
{ "avatar_url": "https://avatars.githubusercontent.com/u/42388668?v=4", "events_url": "https://api.github.com/users/NikhilBartwal/events{/privacy}", "followers_url": "https://api.github.com/users/NikhilBartwal/followers", "following_url": "https://api.github.com/users/NikhilBartwal/following{/other_user}", "gists_url": "https://api.github.com/users/NikhilBartwal/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NikhilBartwal", "id": 42388668, "login": "NikhilBartwal", "node_id": "MDQ6VXNlcjQyMzg4NjY4", "organizations_url": "https://api.github.com/users/NikhilBartwal/orgs", "received_events_url": "https://api.github.com/users/NikhilBartwal/received_events", "repos_url": "https://api.github.com/users/NikhilBartwal/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NikhilBartwal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NikhilBartwal/subscriptions", "type": "User", "url": "https://api.github.com/users/NikhilBartwal" }
`xnli` dataset creating a tuple key while yielding instead of `str` or `int`
https://api.github.com/repos/huggingface/datasets/issues/2229/events
null
https://api.github.com/repos/huggingface/datasets/issues/2229/labels{/name}
2021-04-16T13:21:53Z
null
false
null
null
859,810,602
[]
https://api.github.com/repos/huggingface/datasets/issues/2229
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
When using `ds = datasets.load_dataset('xnli', 'ar')`, the dataset generation script uses the following section of code in the egging, which yields a tuple key instead of the specified `str` or `int` key: https://github.com/huggingface/datasets/blob/56346791aed417306d054d89bd693d6b7eab17f7/datasets/xnli/xnli.py#L196 Since, community datasets in Tensorflow Datasets also use HF datasets, this causes a Tuple key error while loading HF's `xnli` dataset. I'm up for sending a fix for this, I think we can simply use `file_idx + "_" + row_idx` as a unique key instead of a tuple.
2021-04-19T08:56:42Z
https://github.com/huggingface/datasets/issues/2229
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2229/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2228/comments
https://api.github.com/repos/huggingface/datasets/issues/2228/timeline
null
null
null
MDExOlB1bGxSZXF1ZXN0NjE2ODE2MTQz
open
[]
false
2,228
{ "avatar_url": "https://avatars.githubusercontent.com/u/22685854?v=4", "events_url": "https://api.github.com/users/jblemoine/events{/privacy}", "followers_url": "https://api.github.com/users/jblemoine/followers", "following_url": "https://api.github.com/users/jblemoine/following{/other_user}", "gists_url": "https://api.github.com/users/jblemoine/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jblemoine", "id": 22685854, "login": "jblemoine", "node_id": "MDQ6VXNlcjIyNjg1ODU0", "organizations_url": "https://api.github.com/users/jblemoine/orgs", "received_events_url": "https://api.github.com/users/jblemoine/received_events", "repos_url": "https://api.github.com/users/jblemoine/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jblemoine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jblemoine/subscriptions", "type": "User", "url": "https://api.github.com/users/jblemoine" }
[WIP] Add ArrayXD support for fixed size list.
https://api.github.com/repos/huggingface/datasets/issues/2228/events
null
https://api.github.com/repos/huggingface/datasets/issues/2228/labels{/name}
2021-04-16T13:04:08Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2228.diff", "html_url": "https://github.com/huggingface/datasets/pull/2228", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2228.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2228" }
859,795,563
[]
https://api.github.com/repos/huggingface/datasets/issues/2228
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Add support for fixed size list for ArrayXD when shape is known . See https://github.com/huggingface/datasets/issues/2146 Since offset are not stored anymore, the file size is now roughly equal to the actual data size.
2022-07-06T15:19:48Z
https://github.com/huggingface/datasets/pull/2228
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2228/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2227/comments
https://api.github.com/repos/huggingface/datasets/issues/2227/timeline
2021-04-16T13:49:39Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE2Nzk1NjMx
closed
[]
false
2,227
{ "avatar_url": "https://avatars.githubusercontent.com/u/33657802?v=4", "events_url": "https://api.github.com/users/SBrandeis/events{/privacy}", "followers_url": "https://api.github.com/users/SBrandeis/followers", "following_url": "https://api.github.com/users/SBrandeis/following{/other_user}", "gists_url": "https://api.github.com/users/SBrandeis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SBrandeis", "id": 33657802, "login": "SBrandeis", "node_id": "MDQ6VXNlcjMzNjU3ODAy", "organizations_url": "https://api.github.com/users/SBrandeis/orgs", "received_events_url": "https://api.github.com/users/SBrandeis/received_events", "repos_url": "https://api.github.com/users/SBrandeis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SBrandeis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SBrandeis/subscriptions", "type": "User", "url": "https://api.github.com/users/SBrandeis" }
Use update_metadata_with_features decorator in class_encode_column method
https://api.github.com/repos/huggingface/datasets/issues/2227/events
null
https://api.github.com/repos/huggingface/datasets/issues/2227/labels{/name}
2021-04-16T12:31:41Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2227.diff", "html_url": "https://github.com/huggingface/datasets/pull/2227", "merged_at": "2021-04-16T13:49:39Z", "patch_url": "https://github.com/huggingface/datasets/pull/2227.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2227" }
859,771,526
[]
https://api.github.com/repos/huggingface/datasets/issues/2227
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Following @mariosasko 's comment
2021-04-16T13:49:40Z
https://github.com/huggingface/datasets/pull/2227
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2227/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2226/comments
https://api.github.com/repos/huggingface/datasets/issues/2226/timeline
2022-10-05T17:32:15Z
null
completed
MDU6SXNzdWU4NTk3MjAzMDI=
closed
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" } ]
null
2,226
{ "avatar_url": "https://avatars.githubusercontent.com/u/2743060?v=4", "events_url": "https://api.github.com/users/villmow/events{/privacy}", "followers_url": "https://api.github.com/users/villmow/followers", "following_url": "https://api.github.com/users/villmow/following{/other_user}", "gists_url": "https://api.github.com/users/villmow/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/villmow", "id": 2743060, "login": "villmow", "node_id": "MDQ6VXNlcjI3NDMwNjA=", "organizations_url": "https://api.github.com/users/villmow/orgs", "received_events_url": "https://api.github.com/users/villmow/received_events", "repos_url": "https://api.github.com/users/villmow/repos", "site_admin": false, "starred_url": "https://api.github.com/users/villmow/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/villmow/subscriptions", "type": "User", "url": "https://api.github.com/users/villmow" }
Batched map fails when removing all columns
https://api.github.com/repos/huggingface/datasets/issues/2226/events
null
https://api.github.com/repos/huggingface/datasets/issues/2226/labels{/name}
2021-04-16T11:17:01Z
null
false
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
null
859,720,302
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
https://api.github.com/repos/huggingface/datasets/issues/2226
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
Hi @lhoestq , I'm hijacking this issue, because I'm currently trying to do the approach you recommend: > Currently the optimal setup for single-column computations is probably to do something like > > ```python > result = dataset.map(f, input_columns="my_col", remove_columns=dataset.column_names) > ``` Here is my code: (see edit, in which I added a simplified version ``` This is the error: ```bash pyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 8964 but got length 1000 ``` I wonder why this error occurs, when I delete every column? Can you give me a hint? ### Edit: I preprocessed my dataset before (using map with the features argument) and saved it to disk. May this be part of the error? I can iterate over the complete dataset and print every sample before calling map. There seems to be no other problem with the dataset. I tried to simplify the code that crashes: ```python # works log.debug(dataset.column_names) log.debug(dataset) for i, sample in enumerate(dataset): log.debug(i, sample) # crashes counted_dataset = dataset.map( lambda x: {"a": list(range(20))}, input_columns=column, remove_columns=dataset.column_names, load_from_cache_file=False, num_proc=num_workers, batched=True, ) ``` ``` pyarrow.lib.ArrowInvalid: Column 1 named tokens expected length 20 but got length 1000 ``` Edit2: May this be a problem with a schema I set when preprocessing the dataset before? I tried to add the `features` argument to the function and then I get a new error: ```python # crashes counted_dataset = dataset.map( lambda x: {"a": list(range(20))}, input_columns=column, remove_columns=dataset.column_names, load_from_cache_file=False, num_proc=num_workers, batched=True, features=datasets.Features( { "a": datasets.Sequence(datasets.Value("int32")) } ) ) ``` ``` File "env/lib/python3.8/site-packages/datasets/arrow_dataset.py", line 1704, in _map_single writer.write_batch(batch) File "env/lib/python3.8/site-packages/datasets/arrow_writer.py", line 312, in write_batch col_type = schema.field(col).type if schema is not None else None File "pyarrow/types.pxi", line 1341, in pyarrow.lib.Schema.field KeyError: 'Column tokens does not exist in schema' ``` _Originally posted by @villmow in https://github.com/huggingface/datasets/issues/2193#issuecomment-820230874_
2022-10-05T17:32:15Z
https://github.com/huggingface/datasets/issues/2226
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2226/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2225/comments
https://api.github.com/repos/huggingface/datasets/issues/2225/timeline
2021-04-15T21:19:09Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE1NzAzMTY4
closed
[]
false
2,225
{ "avatar_url": "https://avatars.githubusercontent.com/u/46733535?v=4", "events_url": "https://api.github.com/users/alexwdong/events{/privacy}", "followers_url": "https://api.github.com/users/alexwdong/followers", "following_url": "https://api.github.com/users/alexwdong/following{/other_user}", "gists_url": "https://api.github.com/users/alexwdong/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alexwdong", "id": 46733535, "login": "alexwdong", "node_id": "MDQ6VXNlcjQ2NzMzNTM1", "organizations_url": "https://api.github.com/users/alexwdong/orgs", "received_events_url": "https://api.github.com/users/alexwdong/received_events", "repos_url": "https://api.github.com/users/alexwdong/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alexwdong/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alexwdong/subscriptions", "type": "User", "url": "https://api.github.com/users/alexwdong" }
fixed one instance of 'train' to 'test'
https://api.github.com/repos/huggingface/datasets/issues/2225/events
null
https://api.github.com/repos/huggingface/datasets/issues/2225/labels{/name}
2021-04-15T04:26:40Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2225.diff", "html_url": "https://github.com/huggingface/datasets/pull/2225", "merged_at": "2021-04-15T21:19:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/2225.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2225" }
858,469,561
[]
https://api.github.com/repos/huggingface/datasets/issues/2225
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
I believe this should be 'test' instead of 'train'
2021-04-15T22:09:50Z
https://github.com/huggingface/datasets/pull/2225
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/2225/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2224/comments
https://api.github.com/repos/huggingface/datasets/issues/2224/timeline
null
null
null
MDU6SXNzdWU4NTc5ODMzNjE=
open
[]
null
2,224
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Raise error if Windows max path length is not disabled
https://api.github.com/repos/huggingface/datasets/issues/2224/events
null
https://api.github.com/repos/huggingface/datasets/issues/2224/labels{/name}
2021-04-14T14:57:20Z
null
false
null
null
857,983,361
[]
https://api.github.com/repos/huggingface/datasets/issues/2224
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
On startup, raise an error if Windows max path length is not disabled; ask the user to disable it. Linked to discussion in #2220.
2021-04-14T14:59:13Z
https://github.com/huggingface/datasets/issues/2224
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2224/reactions" }
false
https://api.github.com/repos/huggingface/datasets/issues/2223/comments
https://api.github.com/repos/huggingface/datasets/issues/2223/timeline
2021-04-15T19:11:25Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE1MjE4MDIz
closed
[]
false
2,223
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Set test cache config
https://api.github.com/repos/huggingface/datasets/issues/2223/events
null
https://api.github.com/repos/huggingface/datasets/issues/2223/labels{/name}
2021-04-14T12:55:24Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2223.diff", "html_url": "https://github.com/huggingface/datasets/pull/2223", "merged_at": "2021-04-15T19:11:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/2223.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2223" }
857,870,800
[]
https://api.github.com/repos/huggingface/datasets/issues/2223
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Currently, running the tests populates the default cache directory `"~/.cache"`. This PR monkey-patches the config to set the cache directory within the temporary test directory, avoiding side effects.
2021-04-15T19:11:25Z
https://github.com/huggingface/datasets/pull/2223
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2223/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2222/comments
https://api.github.com/repos/huggingface/datasets/issues/2222/timeline
2021-04-14T14:46:19Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE1MTk5MTM5
closed
[]
false
2,222
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix too long WindowsFileLock name
https://api.github.com/repos/huggingface/datasets/issues/2222/events
null
https://api.github.com/repos/huggingface/datasets/issues/2222/labels{/name}
2021-04-14T12:26:52Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2222.diff", "html_url": "https://github.com/huggingface/datasets/pull/2222", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2222.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2222" }
857,847,231
[ { "color": "ffffff", "default": true, "description": "This will not be worked on", "id": 1935892913, "name": "wontfix", "node_id": "MDU6TGFiZWwxOTM1ODkyOTEz", "url": "https://api.github.com/repos/huggingface/datasets/labels/wontfix" } ]
https://api.github.com/repos/huggingface/datasets/issues/2222
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Fix WindowsFileLock name longer than allowed MAX_PATH by shortening the basename.
2021-04-14T15:00:25Z
https://github.com/huggingface/datasets/pull/2222
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2222/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2221/comments
https://api.github.com/repos/huggingface/datasets/issues/2221/timeline
2021-04-14T13:50:19Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE1MTg4MTE5
closed
[]
false
2,221
{ "avatar_url": "https://avatars.githubusercontent.com/u/7669893?v=4", "events_url": "https://api.github.com/users/cahya-wirawan/events{/privacy}", "followers_url": "https://api.github.com/users/cahya-wirawan/followers", "following_url": "https://api.github.com/users/cahya-wirawan/following{/other_user}", "gists_url": "https://api.github.com/users/cahya-wirawan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cahya-wirawan", "id": 7669893, "login": "cahya-wirawan", "node_id": "MDQ6VXNlcjc2Njk4OTM=", "organizations_url": "https://api.github.com/users/cahya-wirawan/orgs", "received_events_url": "https://api.github.com/users/cahya-wirawan/received_events", "repos_url": "https://api.github.com/users/cahya-wirawan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cahya-wirawan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cahya-wirawan/subscriptions", "type": "User", "url": "https://api.github.com/users/cahya-wirawan" }
Add SLR70 - SLR80 and SLR86 to OpenSLR dataset
https://api.github.com/repos/huggingface/datasets/issues/2221/events
null
https://api.github.com/repos/huggingface/datasets/issues/2221/labels{/name}
2021-04-14T12:09:18Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2221.diff", "html_url": "https://github.com/huggingface/datasets/pull/2221", "merged_at": "2021-04-14T13:50:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/2221.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2221" }
857,833,770
[]
https://api.github.com/repos/huggingface/datasets/issues/2221
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
I would like to add SLR70, SLR71, SLR72, SLR73, SLR74, SLR75, SLR76, SLR77, SLR78, SLR79, SLR80 and SLR86 to OpenSLR dataset. The languages are: Nigerian English, Chilean Spanish, Columbian Spanish, Peruvian Spanish, Puerto Rico Spanish, Venezuelan Spanish, Basque, Galician, Gujarati and Kannada.
2021-04-14T13:50:19Z
https://github.com/huggingface/datasets/pull/2221
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2221/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2220/comments
https://api.github.com/repos/huggingface/datasets/issues/2220/timeline
2021-04-14T14:59:34Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE1MTM4NDQz
closed
[]
false
2,220
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
Fix infinite loop in WindowsFileLock
https://api.github.com/repos/huggingface/datasets/issues/2220/events
null
https://api.github.com/repos/huggingface/datasets/issues/2220/labels{/name}
2021-04-14T10:49:58Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2220.diff", "html_url": "https://github.com/huggingface/datasets/pull/2220", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/2220.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2220" }
857,774,626
[ { "color": "ffffff", "default": true, "description": "This will not be worked on", "id": 1935892913, "name": "wontfix", "node_id": "MDU6TGFiZWwxOTM1ODkyOTEz", "url": "https://api.github.com/repos/huggingface/datasets/labels/wontfix" } ]
https://api.github.com/repos/huggingface/datasets/issues/2220
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
MEMBER
Raise exception to avoid infinite loop.
2021-04-14T14:59:50Z
https://github.com/huggingface/datasets/pull/2220
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2220/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2219/comments
https://api.github.com/repos/huggingface/datasets/issues/2219/timeline
2021-04-16T08:50:44Z
null
null
MDExOlB1bGxSZXF1ZXN0NjE0NzYxMzA3
closed
[]
false
2,219
{ "avatar_url": "https://avatars.githubusercontent.com/u/19718818?v=4", "events_url": "https://api.github.com/users/bhavitvyamalik/events{/privacy}", "followers_url": "https://api.github.com/users/bhavitvyamalik/followers", "following_url": "https://api.github.com/users/bhavitvyamalik/following{/other_user}", "gists_url": "https://api.github.com/users/bhavitvyamalik/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bhavitvyamalik", "id": 19718818, "login": "bhavitvyamalik", "node_id": "MDQ6VXNlcjE5NzE4ODE4", "organizations_url": "https://api.github.com/users/bhavitvyamalik/orgs", "received_events_url": "https://api.github.com/users/bhavitvyamalik/received_events", "repos_url": "https://api.github.com/users/bhavitvyamalik/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bhavitvyamalik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bhavitvyamalik/subscriptions", "type": "User", "url": "https://api.github.com/users/bhavitvyamalik" }
Added CUAD dataset
https://api.github.com/repos/huggingface/datasets/issues/2219/events
null
https://api.github.com/repos/huggingface/datasets/issues/2219/labels{/name}
2021-04-13T21:05:03Z
null
false
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/2219.diff", "html_url": "https://github.com/huggingface/datasets/pull/2219", "merged_at": "2021-04-16T08:50:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/2219.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/2219" }
857,321,242
[]
https://api.github.com/repos/huggingface/datasets/issues/2219
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
CONTRIBUTOR
Dataset link : https://github.com/TheAtticusProject/cuad/ Working on README.md currently. Closes #2084 and [#1](https://github.com/TheAtticusProject/cuad/issues/1).
2021-04-24T14:25:51Z
https://github.com/huggingface/datasets/pull/2219
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2219/reactions" }
true
https://api.github.com/repos/huggingface/datasets/issues/2218/comments
https://api.github.com/repos/huggingface/datasets/issues/2218/timeline
null
null
null
MDU6SXNzdWU4NTcyMzg0MzU=
open
[]
null
2,218
{ "avatar_url": "https://avatars.githubusercontent.com/u/7276193?v=4", "events_url": "https://api.github.com/users/amarasovic/events{/privacy}", "followers_url": "https://api.github.com/users/amarasovic/followers", "following_url": "https://api.github.com/users/amarasovic/following{/other_user}", "gists_url": "https://api.github.com/users/amarasovic/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/amarasovic", "id": 7276193, "login": "amarasovic", "node_id": "MDQ6VXNlcjcyNzYxOTM=", "organizations_url": "https://api.github.com/users/amarasovic/orgs", "received_events_url": "https://api.github.com/users/amarasovic/received_events", "repos_url": "https://api.github.com/users/amarasovic/repos", "site_admin": false, "starred_url": "https://api.github.com/users/amarasovic/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amarasovic/subscriptions", "type": "User", "url": "https://api.github.com/users/amarasovic" }
Duplicates in the LAMA dataset
https://api.github.com/repos/huggingface/datasets/issues/2218/events
null
https://api.github.com/repos/huggingface/datasets/issues/2218/labels{/name}
2021-04-13T18:59:49Z
null
false
null
null
857,238,435
[]
https://api.github.com/repos/huggingface/datasets/issues/2218
[ "", "" ]
https://api.github.com/repos/huggingface/datasets
NONE
I observed duplicates in the LAMA probing dataset, see a minimal code below. ``` >>> import datasets >>> dataset = datasets.load_dataset('lama') No config specified, defaulting to: lama/trex Reusing dataset lama (/home/anam/.cache/huggingface/datasets/lama/trex/1.1.0/97deffae13eca0a18e77dfb3960bb31741e973586f5c1fe1ec0d6b5eece7bddc) >>> train_dataset = dataset['train'] >>> train_dataset[0] {'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi Κ’yl tʁɔʃy]; 12 March 1815 – 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'} >>> train_dataset[1] {'description': 'language or languages a person has learned from early childhood', 'label': 'native language', 'masked_sentence': 'Louis Jules Trochu ([lwi Κ’yl tʁɔʃy]; 12 March 1815 – 7 October 1896) was a [MASK] military leader and politician.', 'obj_label': 'French', 'obj_surface': 'French', 'obj_uri': 'Q150', 'predicate_id': 'P103', 'sub_label': 'Louis Jules Trochu', 'sub_surface': 'Louis Jules Trochu', 'sub_uri': 'Q441235', 'template': 'The native language of [X] is [Y] .', 'template_negated': '[X] is not owned by [Y] .', 'type': 'N-1', 'uuid': '40b2ed1c-0961-482e-844e-32596b6117c8'} ``` I checked the original data available at https://dl.fbaipublicfiles.com/LAMA/data.zip. This particular duplicated comes from: ``` {"uuid": "40b2ed1c-0961-482e-844e-32596b6117c8", "obj_uri": "Q150", "obj_label": "French", "sub_uri": "Q441235", "sub_label": "Louis Jules Trochu", "predicate_id": "P103", "evidences": [{"sub_surface": "Louis Jules Trochu", "obj_surface": "French", "masked_sentence": "Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician."}, {"sub_surface": "Louis Jules Trochu", "obj_surface": "French", "masked_sentence": "Louis Jules Trochu ([lwi \u0292yl t\u0281\u0254\u0283y]; 12 March 1815 \u2013 7 October 1896) was a [MASK] military leader and politician."}]} ``` What is the best way to deal with these duplicates if I want to use `datasets` to probe with LAMA?
2021-04-14T21:42:27Z
https://github.com/huggingface/datasets/issues/2218
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/2218/reactions" }
false