html_url
stringlengths
46
51
number
int64
1
7.85k
title
stringlengths
1
290
user
dict
labels
listlengths
0
4
state
stringclasses
2 values
locked
bool
1 class
comments
listlengths
0
30
created_at
timestamp[ns, tz=UTC]date
2020-04-14 10:18:02
2025-11-05 18:11:12
updated_at
timestamp[ns, tz=UTC]date
2020-04-27 16:04:17
2025-11-06 09:44:34
closed_at
timestamp[ns, tz=UTC]date
2020-04-14 12:01:40
2025-11-05 16:02:32
author_association
stringclasses
4 values
draft
bool
2 classes
pull_request
dict
body
stringlengths
0
228k
closed_by
dict
reactions
dict
state_reason
stringclasses
4 values
sub_issues_summary
dict
issue_dependencies_summary
dict
is_pull_request
bool
2 classes
https://github.com/huggingface/datasets/pull/171
171
fix squad metric format
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "One thing for SQuAD is that I wanted to be able to use the SQuAD dataset directly in the metrics and I'm not sure it will be possible with this format.\r\n\r\n(maybe it's not really possible in general though)", "This is kinda related to one thing I had in mind which is that we may want to be able to dump our mo...
2020-05-19T18:37:36Z
2020-05-22T13:36:50Z
2020-05-22T13:36:48Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/171.diff", "html_url": "https://github.com/huggingface/datasets/pull/171", "merged_at": "2020-05-22T13:36:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/171.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/171...
The format of the squad metric was wrong. This should fix #143 I tested with ```python3 predictions = [ {'id': '56be4db0acb8001400a502ec', 'prediction_text': 'Denver Broncos'} ] references = [ {'answers': [{'text': 'Denver Broncos'}], 'id': '56be4db0acb8001400a502ec'} ] ```
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/171/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/170
170
Rename anli dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[]
2020-05-19T16:26:57Z
2020-05-20T12:23:09Z
2020-05-20T12:23:08Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/170.diff", "html_url": "https://github.com/huggingface/datasets/pull/170", "merged_at": "2020-05-20T12:23:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/170.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/170...
What we have now as the `anli` dataset is actually the αNLI dataset from the ART challenge dataset. This name is confusing because `anli` is also the name of adversarial NLI (see [https://github.com/facebookresearch/anli](https://github.com/facebookresearch/anli)). I renamed the current `anli` dataset by `art`.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/170/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/169
169
Adding Qanta (Quizbowl) Dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/1382460?v=4", "events_url": "https://api.github.com/users/EntilZha/events{/privacy}", "followers_url": "https://api.github.com/users/EntilZha/followers", "following_url": "https://api.github.com/users/EntilZha/following{/other_user}", "gists_url": "http...
[]
closed
false
[ "Hi @EntilZha - sorry for waiting so long until taking action here. We created a new command and a new recipe of how to add dummy_data. Can you maybe rebase to `master` as explained in 7. of https://github.com/huggingface/nlp/blob/master/CONTRIBUTING.md#how-to-contribute-to-nlp and check that your dummy data is cor...
2020-05-19T16:03:01Z
2020-05-26T12:52:31Z
2020-05-26T12:52:31Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/169.diff", "html_url": "https://github.com/huggingface/datasets/pull/169", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/169.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/169" }
This PR adds the qanta question answering datasets from [Quizbowl: The Case for Incremental Question Answering](https://arxiv.org/abs/1904.04792) and [Trick Me If You Can: Human-in-the-loop Generation of Adversarial Question Answering Examples](https://www.aclweb.org/anthology/Q19-1029/) (adversarial fold) This part...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/169/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/issues/168
168
Loading 'wikitext' dataset fails
{ "avatar_url": "https://avatars.githubusercontent.com/u/25987633?v=4", "events_url": "https://api.github.com/users/itay1itzhak/events{/privacy}", "followers_url": "https://api.github.com/users/itay1itzhak/followers", "following_url": "https://api.github.com/users/itay1itzhak/following{/other_user}", "gists_u...
[]
closed
false
[ "Hi, make sure you have a recent version of pyarrow.\r\n\r\nAre you using it in Google Colab? In this case, this error is probably the same as #128", "Thanks!\r\n\r\nYes I'm using Google Colab, it seems like a duplicate then.", "Closing as it is a duplicate", "Hi,\r\nThe squad bug seems to be fixed, but the l...
2020-05-19T13:04:29Z
2020-05-26T21:46:52Z
2020-05-26T21:46:52Z
NONE
null
null
Loading the 'wikitext' dataset fails with Attribute error: Code to reproduce (From example notebook): import nlp wikitext_dataset = nlp.load_dataset('wikitext') Error: --------------------------------------------------------------------------- AttributeError Traceback (most rece...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/168/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/pull/167
167
[Tests] refactor tests
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[ "Nice !" ]
2020-05-19T11:43:32Z
2020-05-19T16:17:12Z
2020-05-19T16:17:10Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/167.diff", "html_url": "https://github.com/huggingface/datasets/pull/167", "merged_at": "2020-05-19T16:17:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/167.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/167...
This PR separates AWS and Local tests to remove these ugly statements in the script: ```python if "/" not in dataset_name: logging.info("Skip {} because it is a canonical dataset") return ``` To run a `aws` test, one should now run the following command: ```python pytest -s...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/167/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/issues/166
166
Add a method to shuffle a dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[ { "color": "c5def5", "default": false, "description": "Generic discussion on the library", "id": 2067400324, "name": "generic discussion", "node_id": "MDU6TGFiZWwyMDY3NDAwMzI0", "url": "https://api.github.com/repos/huggingface/datasets/labels/generic%20discussion" } ]
closed
false
[ "+1 for the naming convention\r\n\r\nAbout the `shuffle` method, from my understanding it should be done in `Dataloader` (better separation between dataset processing - usage)", "+1 for shuffle in `Dataloader`. \r\nSome `Dataloader` just store idxs of dataset and just shuffle those idxs, which might(?) be faster ...
2020-05-19T10:08:46Z
2020-06-23T15:07:33Z
2020-06-23T15:07:32Z
MEMBER
null
null
Could maybe be a `dataset.shuffle(generator=None, seed=None)` signature method. Also, we could maybe have a clear indication of which method modify in-place and which methods return/cache a modified dataset. I kinda like torch conversion of having an underscore suffix for all the methods which modify a dataset in-pl...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/166/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/165
165
ANLI
{ "avatar_url": "https://avatars.githubusercontent.com/u/6024930?v=4", "events_url": "https://api.github.com/users/douwekiela/events{/privacy}", "followers_url": "https://api.github.com/users/douwekiela/followers", "following_url": "https://api.github.com/users/douwekiela/following{/other_user}", "gists_url":...
[]
closed
false
[]
2020-05-19T07:50:57Z
2020-05-20T12:23:07Z
2020-05-20T12:23:07Z
NONE
null
null
Can I recommend the following: For ANLI, use https://github.com/facebookresearch/anli. As that paper says, "Our dataset is not to be confused with abductive NLI (Bhagavatula et al., 2019), which calls itself αNLI, or ART.". Indeed, the paper cited under what is currently called anli says in the abstract "We int...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/165/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/164
164
Add Spanish POR and NER Datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/3653789?v=4", "events_url": "https://api.github.com/users/mrm8488/events{/privacy}", "followers_url": "https://api.github.com/users/mrm8488/followers", "following_url": "https://api.github.com/users/mrm8488/following{/other_user}", "gists_url": "https:/...
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
[ "Hello @mrm8488, are these datasets official datasets published in an NLP/CL/ML venue?", "What about this one: https://github.com/ccasimiro88/TranslateAlignRetrieve?" ]
2020-05-18T22:18:21Z
2020-05-25T16:28:45Z
2020-05-25T16:28:45Z
CONTRIBUTOR
null
null
Hi guys, In order to cover multilingual support a little step could be adding standard Datasets used for Spanish NER and POS tasks. I can provide it in raw and preprocessed formats.
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/164/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/163
163
[Feature request] Add cos-e v1.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
[ "Sounds good, @mariamabarham do you want to give a look?\r\nI think we should have two configurations so we can allow either version of the dataset to be loaded with the `1.0` version being the default maybe.\r\n\r\nCc some authors of the great cos-e: @nazneenrajani @bmccann", "cos_e v1.0 is related to CQA v1.0 b...
2020-05-18T22:05:26Z
2020-06-16T23:15:25Z
2020-06-16T18:52:06Z
NONE
null
null
I noticed the second release of cos-e (v1.11) is included in this repo. I wanted to request inclusion of v1.0, since this is the version on which results are reported on in [the paper](https://www.aclweb.org/anthology/P19-1487/), and v1.11 has noted [annotation](https://github.com/salesforce/cos-e/issues/2) [issues](ht...
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/163/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/pull/162
162
fix prev files hash in map
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "Awesome! ", "Hi, yes, this seems to fix #160 -- I cloned the branch locally and verified", "Perfect then :)" ]
2020-05-18T21:20:51Z
2020-05-18T21:36:21Z
2020-05-18T21:36:20Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/162.diff", "html_url": "https://github.com/huggingface/datasets/pull/162", "merged_at": "2020-05-18T21:36:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/162.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/162...
Fix the `.map` issue in #160. This makes sure it takes the previous files when computing the hash.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/162/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/issues/161
161
Discussion on version identifier & MockDataLoaderManager for test data
{ "avatar_url": "https://avatars.githubusercontent.com/u/1382460?v=4", "events_url": "https://api.github.com/users/EntilZha/events{/privacy}", "followers_url": "https://api.github.com/users/EntilZha/followers", "following_url": "https://api.github.com/users/EntilZha/following{/other_user}", "gists_url": "http...
[ { "color": "c5def5", "default": false, "description": "Generic discussion on the library", "id": 2067400324, "name": "generic discussion", "node_id": "MDU6TGFiZWwyMDY3NDAwMzI0", "url": "https://api.github.com/repos/huggingface/datasets/labels/generic%20discussion" } ]
open
false
[ "usually you can replace `download` in your dataset script with `download_and_prepare()` - could you share the code for your dataset here? :-) ", "I have an initial version here: https://github.com/EntilZha/nlp/tree/master/datasets/qanta Thats pretty close to what I'll do as a PR, but still want to do some more s...
2020-05-18T20:31:30Z
2020-05-24T18:10:03Z
null
CONTRIBUTOR
null
null
Hi, I'm working on adding a dataset and ran into an error due to `download` not being defined on `MockDataLoaderManager`, but being defined in `nlp/utils/download_manager.py`. The readme step running this: `RUN_SLOW=1 pytest tests/test_dataset_common.py::DatasetTest::test_load_real_dataset_localmydatasetname` triggers ...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/161/reactions" }
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/160
160
caching in map causes same result to be returned for train, validation and test
{ "avatar_url": "https://avatars.githubusercontent.com/u/247881?v=4", "events_url": "https://api.github.com/users/dpressel/events{/privacy}", "followers_url": "https://api.github.com/users/dpressel/followers", "following_url": "https://api.github.com/users/dpressel/following{/other_user}", "gists_url": "https...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
[ "Hi @dpressel, \r\n\r\nthanks for posting your issue! Can you maybe add a complete code snippet that we can copy paste to reproduce the error? For example, I'm not sure where the variable `train_set` comes from in your code and it seems like you are loading multiple datasets at once? ", "Hi, the full example was...
2020-05-18T19:22:03Z
2020-05-18T21:36:20Z
2020-05-18T21:36:20Z
NONE
null
null
hello, I am working on a program that uses the `nlp` library with the `SST2` dataset. The rough outline of the program is: ``` import nlp as nlp_datasets ... parser.add_argument('--dataset', help='HuggingFace Datasets id', default=['glue', 'sst2'], nargs='+') ... dataset = nlp_datasets.load_dataset(*args....
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/160/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/159
159
How can we add more datasets to nlp library?
{ "avatar_url": "https://avatars.githubusercontent.com/u/17886829?v=4", "events_url": "https://api.github.com/users/Tahsin-Mayeesha/events{/privacy}", "followers_url": "https://api.github.com/users/Tahsin-Mayeesha/followers", "following_url": "https://api.github.com/users/Tahsin-Mayeesha/following{/other_user}"...
[]
closed
false
[ "Found it. https://github.com/huggingface/nlp/tree/master/datasets" ]
2020-05-18T18:35:31Z
2020-05-18T18:37:08Z
2020-05-18T18:37:07Z
NONE
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/17886829?v=4", "events_url": "https://api.github.com/users/Tahsin-Mayeesha/events{/privacy}", "followers_url": "https://api.github.com/users/Tahsin-Mayeesha/followers", "following_url": "https://api.github.com/users/Tahsin-Mayeesha/following{/other_user}"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/159/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/pull/158
158
add Toronto Books Corpus
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
[]
2020-05-18T17:54:45Z
2020-06-11T07:49:15Z
2020-05-19T07:34:56Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/158.diff", "html_url": "https://github.com/huggingface/datasets/pull/158", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/158.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/158" }
This PR adds the Toronto Books Corpus. . It on consider TMX and plain text files (Moses) defined in the table **Statistics and TMX/Moses Downloads** [here](http://opus.nlpl.eu/Books.php )
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/158/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/issues/157
157
nlp.load_dataset() gives "TypeError: list_() takes exactly one argument (2 given)"
{ "avatar_url": "https://avatars.githubusercontent.com/u/47444392?v=4", "events_url": "https://api.github.com/users/saahiluppal/events{/privacy}", "followers_url": "https://api.github.com/users/saahiluppal/followers", "following_url": "https://api.github.com/users/saahiluppal/following{/other_user}", "gists_u...
[]
closed
false
[ "You can just run: \r\n`val = nlp.load_dataset('squad')` \r\n\r\nif you want to have just the validation script you can also do:\r\n\r\n`val = nlp.load_dataset('squad', split=\"validation\")`", "If you want to load a local dataset, make sure you include a `./` before the folder name. ", "This happens by just do...
2020-05-18T16:46:38Z
2020-06-05T08:08:58Z
2020-06-05T08:08:58Z
NONE
null
null
I'm trying to load datasets from nlp but there seems to have error saying "TypeError: list_() takes exactly one argument (2 given)" gist can be found here https://gist.github.com/saahiluppal/c4b878f330b10b9ab9762bc0776c0a6a
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/157/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/156
156
SyntaxError with WMT datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/9419158?v=4", "events_url": "https://api.github.com/users/tomhosking/events{/privacy}", "followers_url": "https://api.github.com/users/tomhosking/followers", "following_url": "https://api.github.com/users/tomhosking/following{/other_user}", "gists_url":...
[]
closed
false
[ "Jeez - don't know what happened there :D Should be fixed now! \r\n\r\nThanks a lot for reporting this @tomhosking !", "Hi @patrickvonplaten!\r\n\r\nI'm now getting the below error:\r\n\r\n```\r\n---------------------------------------------------------------------------\r\nTypeError ...
2020-05-18T14:38:18Z
2020-07-23T16:41:55Z
2020-07-23T16:41:55Z
NONE
null
null
The following snippet produces a syntax error: ``` import nlp dataset = nlp.load_dataset('wmt14') print(dataset['train'][0]) ``` ``` Traceback (most recent call last): File "/home/tom/.local/lib/python3.6/site-packages/IPython/core/interactiveshell.py", line 3326, in run_code exec(code_obj, self....
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/156/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/pull/155
155
Include more links in README, fix typos
{ "avatar_url": "https://avatars.githubusercontent.com/u/13381361?v=4", "events_url": "https://api.github.com/users/bharatr21/events{/privacy}", "followers_url": "https://api.github.com/users/bharatr21/followers", "following_url": "https://api.github.com/users/bharatr21/following{/other_user}", "gists_url": "...
[]
closed
false
[ "I fixed a conflict :) thanks !" ]
2020-05-18T09:47:08Z
2020-05-28T08:31:57Z
2020-05-28T08:31:57Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/155.diff", "html_url": "https://github.com/huggingface/datasets/pull/155", "merged_at": "2020-05-28T08:31:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/155.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/155...
Include more links and fix typos in README
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/155/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/154
154
add Ubuntu Dialogs Corpus datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
[]
2020-05-18T09:34:48Z
2020-05-18T10:12:28Z
2020-05-18T10:12:27Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/154.diff", "html_url": "https://github.com/huggingface/datasets/pull/154", "merged_at": "2020-05-18T10:12:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/154.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/154...
This PR adds the Ubuntu Dialog Corpus datasets version 2.0.
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/154/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/issues/153
153
Meta-datasets (GLUE/XTREME/...) – Special care to attributions and citations
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[ { "color": "c5def5", "default": false, "description": "Generic discussion on the library", "id": 2067400324, "name": "generic discussion", "node_id": "MDU6TGFiZWwyMDY3NDAwMzI0", "url": "https://api.github.com/repos/huggingface/datasets/labels/generic%20discussion" } ]
open
false
[ "As @yoavgo suggested, there should be the possibility to call a function like nlp.bib that outputs all bibtex ref from the datasets and models actually used and eventually nlp.bib.forreadme that would output the same info + versions numbers so they can be included in a readme.md file.", "Actually, double checki...
2020-05-18T07:24:22Z
2020-05-18T21:18:16Z
null
MEMBER
null
null
Meta-datasets are interesting in terms of standardized benchmarks but they also have specific behaviors, in particular in terms of attribution and authorship. It's very important that each specific dataset inside a meta dataset is properly referenced and the citation/specific homepage/etc are very visible and accessibl...
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/153/reactions" }
null
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/pull/152
152
Add GLUE config name check
{ "avatar_url": "https://avatars.githubusercontent.com/u/13381361?v=4", "events_url": "https://api.github.com/users/bharatr21/events{/privacy}", "followers_url": "https://api.github.com/users/bharatr21/followers", "following_url": "https://api.github.com/users/bharatr21/following{/other_user}", "gists_url": "...
[]
closed
false
[ "If tests are being added, any guidance on where to add tests would be helpful!\r\n\r\nTagging @thomwolf for review", "Looks good to me. Is this compatible with the way we are doing tests right now @patrickvonplaten ?", "If the tests pass it should be fine :-) \r\n\r\n@Bharat123rox could you check whether the t...
2020-05-18T07:23:43Z
2020-05-27T22:09:12Z
2020-05-27T22:09:12Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/152.diff", "html_url": "https://github.com/huggingface/datasets/pull/152", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/152.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/152" }
Fixes #130 by adding a name check to the Glue class
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/152/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/151
151
Fix JSON tests.
{ "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "events_url": "https://api.github.com/users/jplu/events{/privacy}", "followers_url": "https://api.github.com/users/jplu/followers", "following_url": "https://api.github.com/users/jplu/following{/other_user}", "gists_url": "https://api.githu...
[]
closed
false
[]
2020-05-18T07:17:38Z
2020-05-18T07:21:52Z
2020-05-18T07:21:51Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/151.diff", "html_url": "https://github.com/huggingface/datasets/pull/151", "merged_at": "2020-05-18T07:21:51Z", "patch_url": "https://github.com/huggingface/datasets/pull/151.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/151...
{ "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "events_url": "https://api.github.com/users/jplu/events{/privacy}", "followers_url": "https://api.github.com/users/jplu/followers", "following_url": "https://api.github.com/users/jplu/following{/other_user}", "gists_url": "https://api.githu...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/151/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/150
150
Add WNUT 17 NER dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/20651387?v=4", "events_url": "https://api.github.com/users/stefan-it/events{/privacy}", "followers_url": "https://api.github.com/users/stefan-it/followers", "following_url": "https://api.github.com/users/stefan-it/following{/other_user}", "gists_url": "...
[]
closed
false
[ "The PR looks awesome! \r\nSince you have already added a dataset I imagine the tests as described in 5. of https://github.com/huggingface/nlp/blob/master/CONTRIBUTING.md#how-to-add-a-dataset all pass, right @stefan-it ?\r\n\r\nI think we are then good to merge this :-) @lhoestq ", "Nice !\r\n\r\nOne thing though...
2020-05-17T22:19:04Z
2020-05-26T20:37:59Z
2020-05-26T20:37:59Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/150.diff", "html_url": "https://github.com/huggingface/datasets/pull/150", "merged_at": "2020-05-26T20:37:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/150.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/150...
Hi, this PR adds the WNUT 17 dataset to `nlp`. > Emerging and Rare entity recognition > This shared task focuses on identifying unusual, previously-unseen entities in the context of emerging discussions. Named entities form the basis of many modern approaches to other tasks (like event clustering and summarisati...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/150/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/issues/149
149
[Feature request] Add Ubuntu Dialogue Corpus dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/28959268?v=4", "events_url": "https://api.github.com/users/danth/events{/privacy}", "followers_url": "https://api.github.com/users/danth/followers", "following_url": "https://api.github.com/users/danth/following{/other_user}", "gists_url": "https://api....
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
[ "@AlphaMycelium the Ubuntu Dialogue Corpus [version 2]( https://github.com/rkadlec/ubuntu-ranking-dataset-creator) is added. Note that it requires a manual download by following the download instructions in the [repos]( https://github.com/rkadlec/ubuntu-ranking-dataset-creator).\r\nMaybe we can close this issue for...
2020-05-17T15:42:39Z
2020-05-18T17:01:46Z
2020-05-18T17:01:46Z
NONE
null
null
https://github.com/rkadlec/ubuntu-ranking-dataset-creator or http://dataset.cs.mcgill.ca/ubuntu-corpus-1.0/
{ "avatar_url": "https://avatars.githubusercontent.com/u/28959268?v=4", "events_url": "https://api.github.com/users/danth/events{/privacy}", "followers_url": "https://api.github.com/users/danth/followers", "following_url": "https://api.github.com/users/danth/following{/other_user}", "gists_url": "https://api....
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/149/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/148
148
_download_and_prepare() got an unexpected keyword argument 'verify_infos'
{ "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "events_url": "https://api.github.com/users/richarddwang/events{/privacy}", "followers_url": "https://api.github.com/users/richarddwang/followers", "following_url": "https://api.github.com/users/richarddwang/following{/other_user}", "gist...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
[ "Same error for dataset 'wiki40b'", "Should be fixed on master :)" ]
2020-05-17T01:48:53Z
2020-05-18T07:38:33Z
2020-05-18T07:38:33Z
CONTRIBUTOR
null
null
# Reproduce In Colab, ``` %pip install -q nlp %pip install -q apache_beam mwparserfromhell dataset = nlp.load_dataset('wikipedia') ``` get ``` Downloading and preparing dataset wikipedia/20200501.aa (download: Unknown size, generated: Unknown size, total: Unknown size) to /root/.cache/huggingface/datasets/w...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/148/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/147
147
Error with sklearn train_test_split
{ "avatar_url": "https://avatars.githubusercontent.com/u/6853743?v=4", "events_url": "https://api.github.com/users/ClonedOne/events{/privacy}", "followers_url": "https://api.github.com/users/ClonedOne/followers", "following_url": "https://api.github.com/users/ClonedOne/following{/other_user}", "gists_url": "h...
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
[ "Indeed. Probably we will want to have a similar method directly in the library", "Related: #166 " ]
2020-05-17T00:28:24Z
2020-06-18T16:23:23Z
2020-06-18T16:23:23Z
NONE
null
null
It would be nice if we could use sklearn `train_test_split` to quickly generate subsets from the dataset objects returned by `nlp.load_dataset`. At the moment the code: ```python data = nlp.load_dataset('imdb', cache_dir=data_cache) f_half, s_half = train_test_split(data['train'], test_size=0.5, random_state=seed)...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/147/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/pull/146
146
Add BERTScore to metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/7753366?v=4", "events_url": "https://api.github.com/users/felixgwu/events{/privacy}", "followers_url": "https://api.github.com/users/felixgwu/followers", "following_url": "https://api.github.com/users/felixgwu/following{/other_user}", "gists_url": "http...
[]
closed
false
[]
2020-05-16T22:09:39Z
2020-05-17T22:22:10Z
2020-05-17T22:22:09Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/146.diff", "html_url": "https://github.com/huggingface/datasets/pull/146", "merged_at": "2020-05-17T22:22:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/146.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/146...
This PR adds [BERTScore](https://arxiv.org/abs/1904.09675) to metrics. Here is an example of how to use it. ```sh import nlp bertscore = nlp.load_metric('metrics/bertscore') # or simply nlp.load_metric('bertscore') after this is added to huggingface's s3 bucket predictions = ['example', 'fruit'] references = [[...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 3, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/146/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/145
145
[AWS Tests] Follow-up PR from #144
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-16T13:53:46Z
2020-05-16T13:54:23Z
2020-05-16T13:54:22Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/145.diff", "html_url": "https://github.com/huggingface/datasets/pull/145", "merged_at": "2020-05-16T13:54:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/145.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/145...
I forgot to add this line in PR #145 .
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/145/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/144
144
[AWS tests] AWS test should not run for canonical datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-16T13:39:30Z
2020-05-16T13:44:34Z
2020-05-16T13:44:33Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/144.diff", "html_url": "https://github.com/huggingface/datasets/pull/144", "merged_at": "2020-05-16T13:44:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/144.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/144...
AWS tests should in general not run for canonical datasets. Only local tests will run in this case. This way a PR is able to pass when adding a new dataset. This PR changes to logic to the following: 1) All datasets that are present in `nlp/datasets` are tested only locally. This way when one adds a canonical da...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/144/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/issues/143
143
ArrowTypeError in squad metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/27137566?v=4", "events_url": "https://api.github.com/users/patil-suraj/events{/privacy}", "followers_url": "https://api.github.com/users/patil-suraj/followers", "following_url": "https://api.github.com/users/patil-suraj/following{/other_user}", "gists_u...
[ { "color": "25b21e", "default": false, "description": "A bug in a metric script", "id": 2067393914, "name": "metric bug", "node_id": "MDU6TGFiZWwyMDY3MzkzOTE0", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug" } ]
closed
false
[ "There was an issue in the format, thanks.\r\nNow you can do\r\n```python3\r\nsquad_dset = nlp.load_dataset(\"squad\")\r\nsquad_metric = nlp.load_metric(\"/Users/quentinlhoest/Desktop/hf/nlp-bis/metrics/squad\")\r\npredictions = [\r\n {\"id\": v[\"id\"], \"prediction_text\": v[\"answers\"][\"text\"][0]} # take ...
2020-05-16T12:06:37Z
2020-05-22T13:38:52Z
2020-05-22T13:36:48Z
CONTRIBUTOR
null
null
`squad_metric.compute` is giving following error ``` ArrowTypeError: Could not convert [{'text': 'Denver Broncos'}, {'text': 'Denver Broncos'}, {'text': 'Denver Broncos'}] with type list: was not a dict, tuple, or recognized null value for conversion to struct type ``` This is how my predictions and references lo...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/143/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/pull/142
142
[WMT] Add all wmt
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-16T11:28:46Z
2020-05-17T12:18:21Z
2020-05-17T12:18:20Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/142.diff", "html_url": "https://github.com/huggingface/datasets/pull/142", "merged_at": "2020-05-17T12:18:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/142.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/142...
This PR adds all wmt datasets scripts. At the moment the script is **not** functional for the language pairs "cs-en", "ru-en", "hi-en" because apparently it takes up to a week to get the manual data for these datasets: see http://ufal.mff.cuni.cz/czeng. The datasets are fully functional though for the "big" languag...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/142/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/141
141
[Clean up] remove bogus folder
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[ "Same for the dataset_infos.json at the project root no ?", "Sorry guys, I haven't noticed. Thank you for mentioning it." ]
2020-05-16T11:13:42Z
2020-05-16T13:24:27Z
2020-05-16T13:24:26Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/141.diff", "html_url": "https://github.com/huggingface/datasets/pull/141", "merged_at": "2020-05-16T13:24:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/141.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/141...
@mariamabarham - I think you accidentally placed it there.
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/141/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/140
140
[Tests] run local tests as default
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[ "You are right and I think those are usual best practice :) I'm 100% fine with this^^", "Merging this for now to unblock other PRs." ]
2020-05-16T10:56:06Z
2020-05-16T13:21:44Z
2020-05-16T13:21:43Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/140.diff", "html_url": "https://github.com/huggingface/datasets/pull/140", "merged_at": "2020-05-16T13:21:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/140.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/140...
This PR also enables local tests by default I think it's safer for now to enable both local and aws tests for every commit. The problem currently is that when we do a PR to add a dataset, the dataset is not yet on AWS on therefore not tested on the PR itself. Thus the PR will always be green even if the datasets are...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/140/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/139
139
Add GermEval 2014 NER dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/20651387?v=4", "events_url": "https://api.github.com/users/stefan-it/events{/privacy}", "followers_url": "https://api.github.com/users/stefan-it/followers", "following_url": "https://api.github.com/users/stefan-it/following{/other_user}", "gists_url": "...
[]
closed
false
[ "Had really fun playing around with this new library :heart: ", "That's awesome - thanks @stefan-it :-) \r\n\r\nCould you maybe rebase to master and check if all dummy data tests are fine. I should have included the local tests directly in the test suite so that all PRs are fully checked: #140 - sorry :D ", "@p...
2020-05-15T23:42:09Z
2020-05-16T13:56:37Z
2020-05-16T13:56:22Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/139.diff", "html_url": "https://github.com/huggingface/datasets/pull/139", "merged_at": "2020-05-16T13:56:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/139.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/139...
Hi, this PR adds the GermEval 2014 NER dataset 😃 > The GermEval 2014 NER Shared Task builds on a new dataset with German Named Entity annotation [1] with the following properties: > - The data was sampled from German Wikipedia and News Corpora as a collection of citations. > - The dataset covers over 31,000...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/139/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/issues/138
138
Consider renaming to nld
{ "avatar_url": "https://avatars.githubusercontent.com/u/8059750?v=4", "events_url": "https://api.github.com/users/honnibal/events{/privacy}", "followers_url": "https://api.github.com/users/honnibal/followers", "following_url": "https://api.github.com/users/honnibal/following{/other_user}", "gists_url": "http...
[ { "color": "c5def5", "default": false, "description": "Generic discussion on the library", "id": 2067400324, "name": "generic discussion", "node_id": "MDU6TGFiZWwyMDY3NDAwMzI0", "url": "https://api.github.com/repos/huggingface/datasets/labels/generic%20discussion" } ]
closed
false
[ "I would suggest `nlds`. NLP is a very general, broad and ambiguous term, the library is not about NLP (as in processing) per se, it is about accessing Natural Language related datasets. So the name should reflect its purpose.\r\n", "Chiming in to second everything @honnibal said, and to add that I think the curr...
2020-05-15T20:23:27Z
2022-09-16T05:18:22Z
2020-09-28T00:08:10Z
NONE
null
null
Hey :) Just making a thread here recording what I said on Twitter, as it's impossible to follow discussion there. It's also just really not a good way to talk about this sort of thing. The issue is that modules go into the global namespace, so you shouldn't use variable names that conflict with module names. This...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 33, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 33, "url": "https://api.github.com/repos/huggingface/datasets/issues/138/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/pull/136
136
Update README.md
{ "avatar_url": "https://avatars.githubusercontent.com/u/75369?v=4", "events_url": "https://api.github.com/users/renaud/events{/privacy}", "followers_url": "https://api.github.com/users/renaud/followers", "following_url": "https://api.github.com/users/renaud/following{/other_user}", "gists_url": "https://api....
[]
closed
false
[ "Thanks, this was fixed with #135 :)" ]
2020-05-15T20:01:07Z
2020-05-17T12:17:28Z
2020-05-17T12:17:28Z
NONE
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/136.diff", "html_url": "https://github.com/huggingface/datasets/pull/136", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/136.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/136" }
small typo
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/136/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/135
135
Fix print statement in READ.md
{ "avatar_url": "https://avatars.githubusercontent.com/u/51091425?v=4", "events_url": "https://api.github.com/users/codehunk628/events{/privacy}", "followers_url": "https://api.github.com/users/codehunk628/followers", "following_url": "https://api.github.com/users/codehunk628/following{/other_user}", "gists_u...
[]
closed
false
[ "Indeed, thanks!" ]
2020-05-15T19:52:23Z
2020-05-17T12:14:06Z
2020-05-17T12:14:05Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/135.diff", "html_url": "https://github.com/huggingface/datasets/pull/135", "merged_at": "2020-05-17T12:14:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/135.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/135...
print statement was throwing generator object instead of printing names of available datasets/metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/135/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/134
134
Update README.md
{ "avatar_url": "https://avatars.githubusercontent.com/u/8753078?v=4", "events_url": "https://api.github.com/users/pranv/events{/privacy}", "followers_url": "https://api.github.com/users/pranv/followers", "following_url": "https://api.github.com/users/pranv/following{/other_user}", "gists_url": "https://api.g...
[]
closed
false
[ "the readme got removed, closing this one" ]
2020-05-15T16:56:14Z
2020-05-28T08:21:49Z
2020-05-28T08:21:49Z
NONE
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/134.diff", "html_url": "https://github.com/huggingface/datasets/pull/134", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/134.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/134" }
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/134/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/issues/133
133
[Question] Using/adding a local dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/1668462?v=4", "events_url": "https://api.github.com/users/zphang/events{/privacy}", "followers_url": "https://api.github.com/users/zphang/followers", "following_url": "https://api.github.com/users/zphang/following{/other_user}", "gists_url": "https://ap...
[]
closed
false
[ "Hi @zphang,\r\n\r\nSo you can just give the local path to a dataset script file and it should work.\r\n\r\nHere is an example:\r\n- you can download one of the scripts in the `datasets` folder of the present repo (or clone the repo)\r\n- then you can load it with `load_dataset('PATH/TO/YOUR/LOCAL/SCRIPT.py')`\r\n\...
2020-05-15T16:26:06Z
2020-07-23T16:44:09Z
2020-07-23T16:44:09Z
NONE
null
null
Users may want to either create/modify a local copy of a dataset, or use a custom-built dataset with the same `Dataset` API as externally downloaded datasets. It appears to be possible to point to a local dataset path rather than downloading the external ones, but I'm not exactly sure how to go about doing this. ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 6, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 6, "url": "https://api.github.com/repos/huggingface/datasets/issues/133/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/132
132
[Feature Request] Add the OpenWebText dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/30755778?v=4", "events_url": "https://api.github.com/users/LysandreJik/events{/privacy}", "followers_url": "https://api.github.com/users/LysandreJik/followers", "following_url": "https://api.github.com/users/LysandreJik/following{/other_user}", "gists_u...
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
[ "We're experimenting with hosting the OpenWebText corpus on Zenodo for easier downloading. https://zenodo.org/record/3834942#.Xs1w8i-z2J8", "Closing since it's been added in #660 " ]
2020-05-15T15:57:29Z
2020-10-07T14:22:48Z
2020-10-07T14:22:48Z
MEMBER
null
null
The OpenWebText dataset is an open clone of OpenAI's WebText dataset. It can be used to train ELECTRA as is specified in the [README](https://www.github.com/google-research/electra). More information and the download link are available [here](https://skylion007.github.io/OpenWebTextCorpus/).
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/132/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/131
131
[Feature request] Add Toronto BookCorpus dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "events_url": "https://api.github.com/users/jarednielsen/events{/privacy}", "followers_url": "https://api.github.com/users/jarednielsen/followers", "following_url": "https://api.github.com/users/jarednielsen/following{/other_user}", "gists...
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
[ "As far as I understand, `wikitext` is refer to `WikiText-103` and `WikiText-2` that created by researchers in Salesforce, and mostly used in traditional language modeling.\r\n\r\nYou might want to say `wikipedia`, a dump from wikimedia foundation.\r\n\r\nAlso I would like to have Toronto BookCorpus too ! Though it...
2020-05-15T15:50:44Z
2020-06-28T21:27:31Z
2020-06-28T21:27:31Z
CONTRIBUTOR
null
null
I know the copyright/distribution of this one is complex, but it would be great to have! That, combined with the existing `wikitext`, would provide a complete dataset for pretraining models like BERT.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/131/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/130
130
Loading GLUE dataset loads CoLA by default
{ "avatar_url": "https://avatars.githubusercontent.com/u/1668462?v=4", "events_url": "https://api.github.com/users/zphang/events{/privacy}", "followers_url": "https://api.github.com/users/zphang/followers", "following_url": "https://api.github.com/users/zphang/following{/other_user}", "gists_url": "https://ap...
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
closed
false
[ "As a follow-up to this: It looks like the actual GLUE task name is supplied as the `name` argument. Is there a way to check what `name`s/sub-datasets are available under a grouping like GLUE? That information doesn't seem to be readily available in info from `nlp.list_datasets()`.\r\n\r\nEdit: I found the info und...
2020-05-15T14:55:50Z
2020-05-27T22:08:15Z
2020-05-27T22:08:15Z
NONE
null
null
If I run: ```python dataset = nlp.load_dataset('glue') ``` The resultant dataset seems to be CoLA be default, without throwing any error. This is in contrast to calling: ```python metric = nlp.load_metric("glue") ``` which throws an error telling the user that they need to specify a task in GLUE. Should the...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/130/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/129
129
[Feature request] Add Google Natural Question dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/1175888?v=4", "events_url": "https://api.github.com/users/elyase/events{/privacy}", "followers_url": "https://api.github.com/users/elyase/followers", "following_url": "https://api.github.com/users/elyase/following{/other_user}", "gists_url": "https://ap...
[ { "color": "e99695", "default": false, "description": "Requesting to add a new dataset", "id": 2067376369, "name": "dataset request", "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request" } ]
closed
false
[ "Indeed, I think this one is almost ready cc @lhoestq ", "I'm doing the latest adjustments to make the processing of the dataset run on Dataflow", "Is there an update to this? It will be very beneficial for the QA community!", "Still work in progress :)\r\nThe idea is to have the dataset already processed som...
2020-05-15T14:14:20Z
2020-07-23T13:21:29Z
2020-07-23T13:21:29Z
NONE
null
null
Would be great to have https://github.com/google-research-datasets/natural-questions as an alternative to SQuAD.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/129/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/128
128
Some error inside nlp.load_dataset()
{ "avatar_url": "https://avatars.githubusercontent.com/u/18486287?v=4", "events_url": "https://api.github.com/users/polkaYK/events{/privacy}", "followers_url": "https://api.github.com/users/polkaYK/followers", "following_url": "https://api.github.com/users/polkaYK/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "Google colab has an old version of Apache Arrow built-in.\r\nBe sure you execute the \"pip install\" cell and restart the notebook environment if the colab asks for it.", "Thanks for reply, worked fine!\r\n" ]
2020-05-15T13:01:29Z
2020-05-15T13:10:40Z
2020-05-15T13:10:40Z
NONE
null
null
First of all, nice work! I am going through [this overview notebook](https://colab.research.google.com/github/huggingface/nlp/blob/master/notebooks/Overview.ipynb) In simple step `dataset = nlp.load_dataset('squad', split='validation[:10%]')` I get an error, which is connected with some inner code, I think: `...
{ "avatar_url": "https://avatars.githubusercontent.com/u/18486287?v=4", "events_url": "https://api.github.com/users/polkaYK/events{/privacy}", "followers_url": "https://api.github.com/users/polkaYK/followers", "following_url": "https://api.github.com/users/polkaYK/following{/other_user}", "gists_url": "https:...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/128/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/pull/127
127
Update Overview.ipynb
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-15T11:46:48Z
2020-05-15T11:47:27Z
2020-05-15T11:47:25Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/127.diff", "html_url": "https://github.com/huggingface/datasets/pull/127", "merged_at": "2020-05-15T11:47:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/127.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/127...
update notebook
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/127/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/126
126
remove webis
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-15T11:25:20Z
2020-05-15T11:31:24Z
2020-05-15T11:30:26Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/126.diff", "html_url": "https://github.com/huggingface/datasets/pull/126", "merged_at": "2020-05-15T11:30:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/126.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/126...
Remove webis from dataset folder. Our first dataset script that only lives on AWS :-) https://s3.console.aws.amazon.com/s3/buckets/datasets.huggingface.co/nlp/datasets/webis/tl_dr/?region=us-east-1 @julien-c @jplu
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/126/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/125
125
[Newsroom] add newsroom
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-15T10:34:34Z
2020-05-15T10:37:07Z
2020-05-15T10:37:02Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/125.diff", "html_url": "https://github.com/huggingface/datasets/pull/125", "merged_at": "2020-05-15T10:37:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/125.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/125...
I checked it with the data link of the mail you forwarded @thomwolf => works well!
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/125/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/124
124
Xsum, require manual download of some files
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
[]
2020-05-15T10:26:13Z
2020-05-15T11:04:48Z
2020-05-15T11:04:46Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/124.diff", "html_url": "https://github.com/huggingface/datasets/pull/124", "merged_at": "2020-05-15T11:04:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/124.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/124...
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/124/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/123
123
[Tests] Local => aws
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[ "For each dataset, If there exist a `dataset_info.json`, then the command `nlp-cli test path/to/my/dataset --al_configs` is successful only if the `dataset_infos.json` is correct. The infos are correct if the size and checksums of the downloaded file are correct, and if the number of examples in each split are corr...
2020-05-15T09:12:25Z
2020-05-15T10:06:12Z
2020-05-15T10:03:26Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/123.diff", "html_url": "https://github.com/huggingface/datasets/pull/123", "merged_at": "2020-05-15T10:03:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/123.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/123...
## Change default Test from local => aws As a default we set` aws=True`, `Local=False`, `slow=False` ### 1. RUN_AWS=1 (default) This runs 4 tests per dataset script. a) Does the dataset script have a valid etag / Can it be reached on AWS? b) Can we load its `builder_class`? c) Can we load **all** dataset c...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/123/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/122
122
Final cleanup of readme and metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
[]
2020-05-15T09:00:52Z
2021-09-03T19:40:09Z
2020-05-15T09:02:22Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/122.diff", "html_url": "https://github.com/huggingface/datasets/pull/122", "merged_at": "2020-05-15T09:02:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/122.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/122...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/122/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/121
121
make style
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-15T08:23:36Z
2020-05-15T08:25:39Z
2020-05-15T08:25:38Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/121.diff", "html_url": "https://github.com/huggingface/datasets/pull/121", "merged_at": "2020-05-15T08:25:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/121.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/121...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/121/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/issues/120
120
🐛 `map` not working
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
[ "I didn't assign the output 🤦‍♂️\r\n\r\n```python\r\ndataset.map(test)\r\n```\r\n\r\nshould be :\r\n\r\n```python\r\ndataset = dataset.map(test)\r\n```" ]
2020-05-15T06:43:08Z
2020-05-15T07:02:38Z
2020-05-15T07:02:38Z
NONE
null
null
I'm trying to run a basic example (mapping function to add a prefix). [Here is the colab notebook I'm using.](https://colab.research.google.com/drive/1YH4JCAy0R1MMSc-k_Vlik_s1LEzP_t1h?usp=sharing) ```python import nlp dataset = nlp.load_dataset('squad', split='validation[:10%]') def test(sample): samp...
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/120/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/119
119
🐛 Colab : type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
[ "It's strange, after installing `nlp` on Colab, the `pyarrow` version seems fine from `pip` but not from python :\r\n\r\n```python\r\nimport pyarrow\r\n\r\n!pip show pyarrow\r\nprint(\"version = {}\".format(pyarrow.__version__))\r\n```\r\n\r\n> Name: pyarrow\r\nVersion: 0.17.0\r\nSummary: Python library for Apache ...
2020-05-15T02:27:26Z
2020-05-15T05:11:22Z
2020-05-15T02:45:28Z
NONE
null
null
I'm trying to load CNN/DM dataset on Colab. [Colab notebook](https://colab.research.google.com/drive/11Mf7iNhIyt6GpgA1dBEtg3cyMHmMhtZS?usp=sharing) But I meet this error : > AttributeError: type object 'pyarrow.lib.RecordBatch' has no attribute 'from_struct_array'
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/119/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/118
118
❓ How to apply a map to all subsets ?
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
[ "That's the way!" ]
2020-05-15T01:58:52Z
2020-05-15T07:05:49Z
2020-05-15T07:04:25Z
NONE
null
null
I'm working with CNN/DM dataset, where I have 3 subsets : `train`, `test`, `validation`. Should I apply my map function on the subsets one by one ? ```python import nlp cnn_dm = nlp.load_dataset('cnn_dailymail') for corpus in ['train', 'test', 'validation']: cnn_dm[corpus] = cnn_dm[corpus].map(my_f...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/118/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/117
117
❓ How to remove specific rows of a dataset ?
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
[ "Hi, you can't do that at the moment.", "Can you do it by now? Coz it would be awfully helpful!", "you can convert dataset object to pandas and remove a feature and convert back to dataset .", "That's what I ended up doing too. but it feels like a workaround to a feature that should be added to the datasets c...
2020-05-15T01:25:06Z
2022-07-15T08:36:44Z
2020-05-15T07:04:32Z
NONE
null
null
I saw on the [example notebook](https://colab.research.google.com/github/huggingface/nlp/blob/master/notebooks/Overview.ipynb#scrollTo=efFhDWhlvSVC) how to remove a specific column : ```python dataset.drop('id') ``` But I didn't find how to remove a specific row. **For example, how can I remove all sample w...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/117/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/116
116
🐛 Trying to use ROUGE metric : pyarrow.lib.ArrowInvalid: Column 1 named references expected length 534 but got length 323
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[ { "color": "25b21e", "default": false, "description": "A bug in a metric script", "id": 2067393914, "name": "metric bug", "node_id": "MDU6TGFiZWwyMDY3MzkzOTE0", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20bug" } ]
closed
false
[ "Can you share your data files or a minimally reproducible example?", "Sure, [here is a Colab notebook](https://colab.research.google.com/drive/1uiS89fnHMG7HV_cYxp3r-_LqJQvNNKs9?usp=sharing) reproducing the error.\r\n\r\n> ArrowInvalid: Column 1 named references expected length 36 but got length 56", "This is b...
2020-05-15T01:12:06Z
2020-05-28T23:43:07Z
2020-05-28T23:43:07Z
NONE
null
null
I'm trying to use rouge metric. I have to files : `test.pred.tokenized` and `test.gold.tokenized` with each line containing a sentence. I tried : ```python import nlp rouge = nlp.load_metric('rouge') with open("test.pred.tokenized") as p, open("test.gold.tokenized") as g: for lp, lg in zip(p, g): ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/116/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/115
115
AttributeError: 'dict' object has no attribute 'info'
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
[ "I could access the info by first accessing the different splits :\r\n\r\n```python\r\nimport nlp\r\n\r\ncnn_dm = nlp.load_dataset('cnn_dailymail')\r\nprint(cnn_dm['train'].info)\r\n```\r\n\r\nInformation seems to be duplicated between the subsets :\r\n\r\n```python\r\nprint(cnn_dm[\"train\"].info == cnn_dm[\"test\...
2020-05-15T00:29:47Z
2020-05-17T13:11:00Z
2020-05-17T13:11:00Z
NONE
null
null
I'm trying to access the information of CNN/DM dataset : ```python cnn_dm = nlp.load_dataset('cnn_dailymail') print(cnn_dm.info) ``` returns : > AttributeError: 'dict' object has no attribute 'info'
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/115/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/issues/114
114
Couldn't reach CNN/DM dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
[]
closed
false
[ "Installing from source (instead of Pypi package) solved the problem." ]
2020-05-15T00:16:17Z
2020-05-15T00:19:52Z
2020-05-15T00:19:51Z
NONE
null
null
I can't get CNN / DailyMail dataset. ```python import nlp assert "cnn_dailymail" in [dataset.id for dataset in nlp.list_datasets()] cnn_dm = nlp.load_dataset('cnn_dailymail') ``` [Colab notebook](https://colab.research.google.com/drive/1zQ3bYAVzm1h0mw0yWPqKAg_4EUlSx5Ex?usp=sharing) gives following error ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "events_url": "https://api.github.com/users/astariul/events{/privacy}", "followers_url": "https://api.github.com/users/astariul/followers", "following_url": "https://api.github.com/users/astariul/following{/other_user}", "gists_url": "htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/114/reactions" }
completed
{ "completed": 0, "percent_completed": 0, "total": 0 }
{ "blocked_by": 0, "blocking": 0, "total_blocked_by": 0, "total_blocking": 0 }
false
https://github.com/huggingface/datasets/pull/113
113
Adding docstrings and some doc
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
[]
2020-05-14T23:14:41Z
2020-05-14T23:22:45Z
2020-05-14T23:22:44Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/113.diff", "html_url": "https://github.com/huggingface/datasets/pull/113", "merged_at": "2020-05-14T23:22:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/113.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/113...
Some doc
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/113/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/112
112
Qa4mre - add dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-14T22:17:51Z
2020-05-15T09:16:43Z
2020-05-15T09:16:42Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/112.diff", "html_url": "https://github.com/huggingface/datasets/pull/112", "merged_at": "2020-05-15T09:16:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/112.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/112...
Added dummy data test only for the first config. Will do the rest later. I had to do add some minor hacks to an important function to make it work. There might be a cleaner way to handle it - can you take a look @thomwolf ?
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/112/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/111
111
[Clean-up] remove under construction datastes
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-14T20:52:13Z
2020-05-14T20:52:23Z
2020-05-14T20:52:22Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/111.diff", "html_url": "https://github.com/huggingface/datasets/pull/111", "merged_at": "2020-05-14T20:52:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/111.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/111...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/111/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/110
110
fix reddit tifu dummy data
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-14T20:37:37Z
2020-05-14T20:40:14Z
2020-05-14T20:40:13Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/110.diff", "html_url": "https://github.com/huggingface/datasets/pull/110", "merged_at": "2020-05-14T20:40:13Z", "patch_url": "https://github.com/huggingface/datasets/pull/110.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/110...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/110/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/109
109
[Reclor] fix reclor
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-14T20:16:26Z
2020-05-14T20:19:09Z
2020-05-14T20:19:08Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/109.diff", "html_url": "https://github.com/huggingface/datasets/pull/109", "merged_at": "2020-05-14T20:19:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/109.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/109...
- That's probably one me. Could have made the manual data test more flexible. @mariamabarham
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/109/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/108
108
convert can use manual dir as second argument
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-14T16:52:32Z
2020-05-14T16:52:43Z
2020-05-14T16:52:42Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/108.diff", "html_url": "https://github.com/huggingface/datasets/pull/108", "merged_at": "2020-05-14T16:52:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/108.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/108...
@mariamabarham
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/108/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/107
107
add writer_batch_size to GeneratorBasedBuilder
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "Awesome that's great!" ]
2020-05-14T16:35:39Z
2020-05-14T16:50:30Z
2020-05-14T16:50:29Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/107.diff", "html_url": "https://github.com/huggingface/datasets/pull/107", "merged_at": "2020-05-14T16:50:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/107.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/107...
You can now specify `writer_batch_size` in the builder arguments or directly in `load_dataset`
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/107/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/106
106
Add data dir test command
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "Nice - I think we can merge this. I will update the checksums for `wikihow` then as well" ]
2020-05-14T16:18:39Z
2020-05-14T16:49:11Z
2020-05-14T16:49:10Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/106.diff", "html_url": "https://github.com/huggingface/datasets/pull/106", "merged_at": "2020-05-14T16:49:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/106.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/106...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/106/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/105
105
[New structure on AWS] Adapt paths
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-14T15:55:57Z
2020-05-14T15:56:28Z
2020-05-14T15:56:27Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/105.diff", "html_url": "https://github.com/huggingface/datasets/pull/105", "merged_at": "2020-05-14T15:56:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/105.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/105...
Some small changes so that we have the correct paths. @julien-c
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/105/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/104
104
Add trivia_q
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-14T14:27:19Z
2020-07-12T05:34:20Z
2020-05-14T20:23:32Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/104.diff", "html_url": "https://github.com/huggingface/datasets/pull/104", "merged_at": "2020-05-14T20:23:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/104.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/104...
Currently tested only for one config to pass tests. Needs to add more dummy data later.
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/104/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/103
103
[Manual downloads] add logic proposal for manual downloads and add wikihow
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[ "> Wikihow is an example that needs to manually download two files as stated in: https://github.com/mahnazkoupaee/WikiHow-Dataset.\r\n> \r\n> The user can then store these files under a hard-coded name: `wikihowAll.csv` and `wikihowSep.csv` in this case in a directory of his choice, e.g. `~/wikihow/manual_dir`.\r\n...
2020-05-14T13:30:36Z
2020-05-14T14:27:41Z
2020-05-14T14:27:40Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/103.diff", "html_url": "https://github.com/huggingface/datasets/pull/103", "merged_at": "2020-05-14T14:27:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/103.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/103...
Wikihow is an example that needs to manually download two files as stated in: https://github.com/mahnazkoupaee/WikiHow-Dataset. The user can then store these files under a hard-coded name: `wikihowAll.csv` and `wikihowSep.csv` in this case in a directory of his choice, e.g. `~/wikihow/manual_dir`. The dataset ca...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/103/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/102
102
Run save infos
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "Haha that cornell dialogue dataset - that ran for 3h on my computer as well. The `generate_examples` method in this script is one of the most inefficient code samples I've ever seen :D ", "Indeed it's been 3 hours already\r\n```73111 examples [3:07:48, 2.40 examples/s]```" ]
2020-05-14T13:27:26Z
2020-05-14T15:43:04Z
2020-05-14T15:43:03Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/102.diff", "html_url": "https://github.com/huggingface/datasets/pull/102", "merged_at": "2020-05-14T15:43:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/102.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/102...
I replaced the old checksum file with the new `dataset_infos.json` by running the script on almost all the datasets we have. The only one that is still running on my side is the cornell dialog
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/102/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/101
101
[Reddit] add reddit
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-14T10:25:02Z
2020-05-14T10:27:25Z
2020-05-14T10:27:24Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/101.diff", "html_url": "https://github.com/huggingface/datasets/pull/101", "merged_at": "2020-05-14T10:27:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/101.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/101...
- Everything worked fine @mariamabarham. Made my computer nearly crash, but all seems to be working :-)
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/101/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/100
100
Add per type scores in seqeval metric
{ "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "events_url": "https://api.github.com/users/jplu/events{/privacy}", "followers_url": "https://api.github.com/users/jplu/followers", "following_url": "https://api.github.com/users/jplu/following{/other_user}", "gists_url": "https://api.githu...
[]
closed
false
[ "LGTM :-) Some small suggestions to shorten the code a bit :-) ", "Can you put the kwargs as normal kwargs instead of a dict? (And add them to the kwargs description As well)", "@thom Is-it what you meant?", "Yes and there is a dynamically generated doc string in the metric script KWARGS DESCRIPTION" ]
2020-05-14T09:37:52Z
2020-05-14T23:21:35Z
2020-05-14T23:21:34Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/100.diff", "html_url": "https://github.com/huggingface/datasets/pull/100", "merged_at": "2020-05-14T23:21:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/100.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/100...
This PR add a bit more detail in the seqeval metric. Now the usage and output are: ```python import nlp met = nlp.load_metric('metrics/seqeval') references = [['O', 'O', 'O', 'B-MISC', 'I-MISC', 'I-MISC', 'O'], ['B-PER', 'I-PER', 'O']] predictions = [['O', 'O', 'B-MISC', 'I-MISC', 'I-MISC', 'I-MISC', 'O'], ['B-...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/100/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/99
99
[Cmrc 2018] fix cmrc2018
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-14T08:22:03Z
2020-05-14T08:49:42Z
2020-05-14T08:49:41Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/99.diff", "html_url": "https://github.com/huggingface/datasets/pull/99", "merged_at": "2020-05-14T08:49:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/99.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/99" }
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/99/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/98
98
Webis tl-dr
{ "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "events_url": "https://api.github.com/users/jplu/events{/privacy}", "followers_url": "https://api.github.com/users/jplu/followers", "following_url": "https://api.github.com/users/jplu/following{/other_user}", "gists_url": "https://api.githu...
[]
closed
false
[ "Should that rather be in an organization scope, @thomwolf @patrickvonplaten ?", "> Should that rather be in an organization scope, @thomwolf @patrickvonplaten ?\r\n\r\nI'm a bit indifferent - both would be fine for me!", "@jplu - if creating the dummy_data is too tedious, I can do it as well :-) ", "There is...
2020-05-14T06:22:18Z
2020-09-03T10:00:21Z
2020-05-14T20:54:16Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/98.diff", "html_url": "https://github.com/huggingface/datasets/pull/98", "merged_at": "2020-05-14T20:54:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/98.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/98" }
Add the Webid TL:DR dataset.
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/98/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/97
97
[Csv] add tests for csv dataset script
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[ "@thomwolf - can you check and merge if ok? " ]
2020-05-13T23:06:11Z
2020-05-13T23:23:16Z
2020-05-13T23:23:15Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/97.diff", "html_url": "https://github.com/huggingface/datasets/pull/97", "merged_at": "2020-05-13T23:23:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/97.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/97" }
Adds dummy data tests for csv.
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/97/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/96
96
lm1b
{ "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "events_url": "https://api.github.com/users/jplu/events{/privacy}", "followers_url": "https://api.github.com/users/jplu/followers", "following_url": "https://api.github.com/users/jplu/following{/other_user}", "gists_url": "https://api.githu...
[]
closed
false
[ "I might have a different version of `isort` than others. It seems like I'm always reordering the imports of others. But isn't really a problem..." ]
2020-05-13T20:38:44Z
2020-05-14T14:13:30Z
2020-05-14T14:13:29Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/96.diff", "html_url": "https://github.com/huggingface/datasets/pull/96", "merged_at": "2020-05-14T14:13:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/96.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/96" }
Add lm1b dataset.
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/96/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/95
95
Replace checksums files by Dataset infos json
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "Great! LGTM :-) ", "> Ok, really clean!\r\n> I like the logic (not a huge fan of using `_asdict_inner` but it makes sense).\r\n> I think it's a nice improvement!\r\n> \r\n> How should we update the files in the repo? Run a big job on a server or on somebody's computer who has most of the datasets already downloa...
2020-05-13T19:36:16Z
2020-05-14T08:58:43Z
2020-05-14T08:58:42Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/95.diff", "html_url": "https://github.com/huggingface/datasets/pull/95", "merged_at": "2020-05-14T08:58:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/95.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/95" }
### Better verifications when loading a dataset I replaced the `urls_checksums` directory that used to contain `checksums.txt` and `cached_sizes.txt`, by a single file `dataset_infos.json`. It's just a dict `config_name` -> `DatasetInfo`. It simplifies and improves how verifications of checksums and splits sizes ...
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/95/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/94
94
Librispeech
{ "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "events_url": "https://api.github.com/users/jplu/events{/privacy}", "followers_url": "https://api.github.com/users/jplu/followers", "following_url": "https://api.github.com/users/jplu/following{/other_user}", "gists_url": "https://api.githu...
[]
closed
false
[ "@jplu - I changed this weird archieve - iter method to something simpler. It's only one file to download anyways so I don't see the point of using weird iter methods...It's a huge file though :D 30 million lines of text. Took me quite some time to download :D " ]
2020-05-13T16:04:14Z
2020-05-13T21:29:03Z
2020-05-13T21:29:02Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/94.diff", "html_url": "https://github.com/huggingface/datasets/pull/94", "merged_at": "2020-05-13T21:29:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/94.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/94" }
Add librispeech dataset and remove some useless content.
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/94/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/93
93
Cleanup notebooks and various fixes
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
[]
2020-05-13T14:58:58Z
2020-05-13T15:01:48Z
2020-05-13T15:01:47Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/93.diff", "html_url": "https://github.com/huggingface/datasets/pull/93", "merged_at": "2020-05-13T15:01:47Z", "patch_url": "https://github.com/huggingface/datasets/pull/93.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/93" }
Fixes on dataset (more flexible) metrics (fix) and general clean ups
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/93/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/92
92
[WIP] add wmt14
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-13T10:42:03Z
2020-05-16T11:17:38Z
2020-05-16T11:17:37Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/92.diff", "html_url": "https://github.com/huggingface/datasets/pull/92", "merged_at": "2020-05-16T11:17:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/92.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/92" }
WMT14 takes forever to download :-/ - WMT is the first dataset that uses an abstract class IMO, so I had to modify the `load_dataset_module` a bit.
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/92/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/91
91
[Paracrawl] add paracrawl
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-13T10:39:00Z
2020-05-13T10:40:15Z
2020-05-13T10:40:14Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/91.diff", "html_url": "https://github.com/huggingface/datasets/pull/91", "merged_at": "2020-05-13T10:40:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/91.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/91" }
- Huge dataset - took ~1h to download - Also this PR reformats all dataset scripts and adds `datasets` to `make style`
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/91/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/90
90
Add download gg drive
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "awesome - so no manual downloaded needed here? ", "Yes exactly. It works like a standard download" ]
2020-05-13T09:56:02Z
2020-05-13T12:46:28Z
2020-05-13T10:05:31Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/90.diff", "html_url": "https://github.com/huggingface/datasets/pull/90", "merged_at": "2020-05-13T10:05:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/90.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/90" }
We can now add datasets that download from google drive
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/90/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/89
89
Add list and inspect methods - cleanup hf_api
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
[]
2020-05-13T09:30:15Z
2020-05-13T14:05:00Z
2020-05-13T09:33:10Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/89.diff", "html_url": "https://github.com/huggingface/datasets/pull/89", "merged_at": "2020-05-13T09:33:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/89.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/89" }
Add a bunch of methods to easily list and inspect the processing scripts up-loaded on S3: ```python nlp.list_datasets() nlp.list_metrics() # Copy and prepare the scripts at `local_path` for easy inspection/modification. nlp.inspect_dataset(path, local_path) # Copy and prepare the scripts at `local_path` for easy...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/89/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/88
88
Add wiki40b
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "Looks good to me. I have not really looked too much into the Beam Datasets yet though - so I think you can merge whenever you think is good for Beam datasets :-) " ]
2020-05-13T09:16:01Z
2020-05-13T12:31:55Z
2020-05-13T12:31:54Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/88.diff", "html_url": "https://github.com/huggingface/datasets/pull/88", "merged_at": "2020-05-13T12:31:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/88.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/88" }
This one is a beam dataset that downloads files using tensorflow. I tested it on a small config and it works fine
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/88/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/87
87
Add Flores
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-13T08:51:29Z
2020-05-13T09:23:34Z
2020-05-13T09:23:33Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/87.diff", "html_url": "https://github.com/huggingface/datasets/pull/87", "merged_at": "2020-05-13T09:23:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/87.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/87" }
Beautiful language for sure!
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/87/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/86
86
[Load => load_dataset] change naming
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-13T08:43:00Z
2020-05-13T08:50:58Z
2020-05-13T08:50:57Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/86.diff", "html_url": "https://github.com/huggingface/datasets/pull/86", "merged_at": "2020-05-13T08:50:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/86.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/86" }
Rename leftovers @thomwolf
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/86/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/85
85
Add boolq
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "Awesome :-) Thanks for adding the function to the Mock DL Manager" ]
2020-05-13T08:32:27Z
2020-05-13T09:09:39Z
2020-05-13T09:09:38Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/85.diff", "html_url": "https://github.com/huggingface/datasets/pull/85", "merged_at": "2020-05-13T09:09:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/85.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/85" }
I just added the dummy data for this dataset. This one was uses `tf.io.gfile.copy` to download the data but I added the support for custom download in the mock_download_manager. I also had to add a `tensorflow` dependency for tests.
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/85/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/84
84
[TedHrLr] add left dummy data
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-13T08:27:20Z
2020-05-13T08:29:22Z
2020-05-13T08:29:21Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/84.diff", "html_url": "https://github.com/huggingface/datasets/pull/84", "merged_at": "2020-05-13T08:29:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/84.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/84" }
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/84/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/83
83
New datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
[]
2020-05-12T18:22:27Z
2020-05-12T18:22:47Z
2020-05-12T18:22:45Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/83.diff", "html_url": "https://github.com/huggingface/datasets/pull/83", "merged_at": "2020-05-12T18:22:45Z", "patch_url": "https://github.com/huggingface/datasets/pull/83.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/83" }
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/83/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/82
82
[Datasets] add ted_hrlr
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-12T16:46:50Z
2020-05-13T07:52:54Z
2020-05-13T07:52:53Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/82.diff", "html_url": "https://github.com/huggingface/datasets/pull/82", "merged_at": "2020-05-13T07:52:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/82.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/82" }
@thomwolf - After looking at `xnli` I think it's better to leave the translation features and add a `translation` key to make them work in our framework. The result looks like this: ![Screenshot from 2020-05-12 18-34-43](https://user-images.githubusercontent.com/23423619/81721933-ee1faf00-9480-11ea-9e95-d6557cbd0c...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/82/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/81
81
add tests
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[]
2020-05-12T16:28:19Z
2020-05-13T07:43:57Z
2020-05-13T07:43:56Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/81.diff", "html_url": "https://github.com/huggingface/datasets/pull/81", "merged_at": "2020-05-13T07:43:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/81.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/81" }
Tests for py_utils functions and for the BaseReader used to read from arrow and parquet. I also removed unused utils functions.
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/81/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/80
80
Add nbytes + nexamples check
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "Looks good to me! Should we hard code those numbers in the config classes and make sure that when loading a dataset that the numbers match? " ]
2020-05-12T16:18:43Z
2020-05-13T07:52:34Z
2020-05-13T07:52:33Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/80.diff", "html_url": "https://github.com/huggingface/datasets/pull/80", "merged_at": "2020-05-13T07:52:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/80.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/80" }
### Save size and number of examples Now when you do `save_checksums`, it also create `cached_sizes.txt` right next to the checksum file. This new file stores the bytes sizes and the number of examples of each split that has been prepared and stored in the cache. Example: ``` # Cached sizes: <full_config_name> <n...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/80/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/79
79
[Convert] add new pattern
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-12T16:16:51Z
2020-05-12T16:17:10Z
2020-05-12T16:17:09Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/79.diff", "html_url": "https://github.com/huggingface/datasets/pull/79", "merged_at": "2020-05-12T16:17:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/79.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/79" }
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/79/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/78
78
[Tests] skip beam dataset tests for now
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[ "@lhoestq - I moved the wkipedia file to the \"correct\" folder. ", "Nice thanks !" ]
2020-05-12T16:00:58Z
2020-05-12T16:16:24Z
2020-05-12T16:16:22Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/78.diff", "html_url": "https://github.com/huggingface/datasets/pull/78", "merged_at": "2020-05-12T16:16:22Z", "patch_url": "https://github.com/huggingface/datasets/pull/78.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/78" }
For now we will skip tests for Beam Datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/78/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/77
77
New datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
[]
closed
false
[]
2020-05-12T13:51:59Z
2020-05-12T14:02:16Z
2020-05-12T14:02:15Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/77.diff", "html_url": "https://github.com/huggingface/datasets/pull/77", "merged_at": "2020-05-12T14:02:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/77.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/77" }
{ "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "events_url": "https://api.github.com/users/mariamabarham/events{/privacy}", "followers_url": "https://api.github.com/users/mariamabarham/followers", "following_url": "https://api.github.com/users/mariamabarham/following{/other_user}", "g...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/77/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/76
76
pin flake 8
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-12T11:25:29Z
2020-05-12T11:27:35Z
2020-05-12T11:27:34Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/76.diff", "html_url": "https://github.com/huggingface/datasets/pull/76", "merged_at": "2020-05-12T11:27:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/76.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/76" }
Flake 8's new version does not like our format. Pinning the version for now.
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/76/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/75
75
WIP adding metrics
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
[]
closed
false
[ "It's all about my metric stuff so I'll probably merge it unless you want to have a look.\r\n\r\nTook the occasion to remove the old doc and requirements.txt" ]
2020-05-12T09:52:00Z
2020-05-13T07:44:12Z
2020-05-13T07:44:10Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/75.diff", "html_url": "https://github.com/huggingface/datasets/pull/75", "merged_at": "2020-05-13T07:44:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/75.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/75" }
Adding the following metrics as identified by @mariamabarham: 1. BLEU: BiLingual Evaluation Understudy: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py, https://github.com/chakki-works/sumeval/blob/master/sumeval/metrics/bleu.py (multilingual) 2. GLEU: Google-BLEU: https://github.com/cnap/gec-...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/75/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/74
74
fix overflow check
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[]
2020-05-12T09:38:01Z
2020-05-12T10:04:39Z
2020-05-12T10:04:38Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/74.diff", "html_url": "https://github.com/huggingface/datasets/pull/74", "merged_at": "2020-05-12T10:04:37Z", "patch_url": "https://github.com/huggingface/datasets/pull/74.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/74" }
I did some tests and unfortunately the test ``` pa_array.nbytes > MAX_BATCH_BYTES ``` doesn't work. Indeed for a StructArray, `nbytes` can be less 2GB even if there is an overflow (it loops...). I don't think we can do a proper overflow test for the limit of 2GB... For now I replaced it with a sanity check on...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/74/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/73
73
JSON script
{ "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "events_url": "https://api.github.com/users/jplu/events{/privacy}", "followers_url": "https://api.github.com/users/jplu/followers", "following_url": "https://api.github.com/users/jplu/following{/other_user}", "gists_url": "https://api.githu...
[]
closed
false
[ "The tests for the Wikipedia dataset do not pass anymore with the error:\r\n```\r\nTo be able to use this dataset, you need to install the following dependencies ['mwparserfromhell'] using 'pip install mwparserfromhell' for instance'\r\n```", "This was an issue on master. You can just rebase from master.", "Per...
2020-05-12T07:11:22Z
2020-05-18T06:50:37Z
2020-05-18T06:50:36Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/73.diff", "html_url": "https://github.com/huggingface/datasets/pull/73", "merged_at": "2020-05-18T06:50:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/73.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/73" }
Add a JSONS script to read JSON datasets from files.
{ "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "events_url": "https://api.github.com/users/jplu/events{/privacy}", "followers_url": "https://api.github.com/users/jplu/followers", "following_url": "https://api.github.com/users/jplu/following{/other_user}", "gists_url": "https://api.githu...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/73/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/72
72
[README dummy data tests] README to better understand how the dummy data structure works
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
[]
closed
false
[]
2020-05-11T22:19:03Z
2020-05-11T22:26:03Z
2020-05-11T22:26:01Z
CONTRIBUTOR
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/72.diff", "html_url": "https://github.com/huggingface/datasets/pull/72", "merged_at": "2020-05-11T22:26:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/72.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/72" }
In this PR a README.md is added to tests to shine more light on how the dummy data structure works. I try to explain the different possible cases. IMO the best way to understand the logic is to checkout the dummy data structure of the different datasets I mention in the README.md since those are the "edge cases". @...
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_use...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/72/reactions" }
null
null
null
true
https://github.com/huggingface/datasets/pull/71
71
Fix arrow writer for big datasets using writer_batch_size
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https:...
[]
closed
false
[ "After a quick chat with Yacine : the 2Go test may not be sufficient actually, as I'm looking at the size of the array and not the size of the current_rows. If the test doesn't do the job I think I'll remove it and lower the batch size a bit to be sure that it never exceeds 2Go. I'll do more tests later" ]
2020-05-11T14:45:36Z
2020-05-11T20:09:47Z
2020-05-11T20:00:38Z
MEMBER
false
{ "diff_url": "https://github.com/huggingface/datasets/pull/71.diff", "html_url": "https://github.com/huggingface/datasets/pull/71", "merged_at": "2020-05-11T20:00:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/71.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/71" }
This PR fixes Yacine's bug. According to [this](https://github.com/apache/arrow/blob/master/docs/source/cpp/arrays.rst#size-limitations-and-recommendations), it is not recommended to have pyarrow arrays bigger than 2Go. Therefore I set a default batch size of 100 000 examples per batch. In general it shouldn't exce...
{ "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "events_url": "https://api.github.com/users/thomwolf/events{/privacy}", "followers_url": "https://api.github.com/users/thomwolf/followers", "following_url": "https://api.github.com/users/thomwolf/following{/other_user}", "gists_url": "http...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/71/reactions" }
null
null
null
true