url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1.11B
node_id
stringlengths
18
32
number
int64
1
3.59k
title
stringlengths
1
276
user
dict
labels
list
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
list
milestone
dict
comments
list
created_at
int64
1,587B
1,642B
updated_at
int64
1,587B
1,642B
closed_at
int64
1,587B
1,642B
author_association
stringclasses
3 values
active_lock_reason
null
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
draft
bool
2 classes
pull_request
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/548
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/548/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/548/comments
https://api.github.com/repos/huggingface/datasets/issues/548/events
https://github.com/huggingface/datasets/pull/548
689,285,996
MDExOlB1bGxSZXF1ZXN0NDc2MzYzMjU1
548
[Breaking] Switch text loading to multi-threaded PyArrow loading
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "Awesome !\r\nAlso I was wondering if we should try to make the hashing of the `data_files` faster (it is used to build the cache directory of datasets like `text` or `json`). Right now it reads each file and hashes all of its data. We could simply hash the path and some metadata including the `time last modified` ...
1,598,886,941,000
1,599,560,398,000
1,599,560,397,000
MEMBER
null
Test if we can get better performances for large-scale text datasets by using multi-threaded text file loading based on Apache Arrow multi-threaded CSV loader. If it works ok, it would fix #546. **Breaking change**: The text lines now do not include final line-breaks anymore.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/548/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/548/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/548", "html_url": "https://github.com/huggingface/datasets/pull/548", "diff_url": "https://github.com/huggingface/datasets/pull/548.diff", "patch_url": "https://github.com/huggingface/datasets/pull/548.patch", "merged_at": 1599560397000 }
true
https://api.github.com/repos/huggingface/datasets/issues/547
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/547/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/547/comments
https://api.github.com/repos/huggingface/datasets/issues/547/events
https://github.com/huggingface/datasets/pull/547
689,268,589
MDExOlB1bGxSZXF1ZXN0NDc2MzQ4OTk5
547
[Distributed] Making loading distributed datasets a bit safer
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,598,885,494,000
1,598,886,990,000
1,598,886,989,000
MEMBER
null
Add some file-locks during dataset loading
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/547/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/547/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/547", "html_url": "https://github.com/huggingface/datasets/pull/547", "diff_url": "https://github.com/huggingface/datasets/pull/547.diff", "patch_url": "https://github.com/huggingface/datasets/pull/547.patch", "merged_at": 1598886989000 }
true
https://api.github.com/repos/huggingface/datasets/issues/546
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/546/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/546/comments
https://api.github.com/repos/huggingface/datasets/issues/546/events
https://github.com/huggingface/datasets/issues/546
689,186,526
MDU6SXNzdWU2ODkxODY1MjY=
546
Very slow data loading on large dataset
{ "login": "agemagician", "id": 6087313, "node_id": "MDQ6VXNlcjYwODczMTM=", "avatar_url": "https://avatars.githubusercontent.com/u/6087313?v=4", "gravatar_id": "", "url": "https://api.github.com/users/agemagician", "html_url": "https://github.com/agemagician", "followers_url": "https://api.github.com/us...
[]
closed
false
null
[]
null
[ "When you load a text file for the first time with `nlp`, the file is converted into Apache Arrow format. Arrow allows to use memory-mapping, which means that you can load an arbitrary large dataset.\r\n\r\nNote that as soon as the conversion has been done once, the next time you'll load the dataset it will be much...
1,598,878,643,000
1,633,705,937,000
1,599,560,397,000
NONE
null
I made a simple python script to check the NLP library speed, which loads 1.1 TB of textual data. It has been 8 hours and still, it is on the loading steps. It does work when the text dataset size is small about 1 GB, but it doesn't scale. It also uses a single thread during the data loading step. ``` train_fil...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/546/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/546/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/545
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/545/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/545/comments
https://api.github.com/repos/huggingface/datasets/issues/545/events
https://github.com/huggingface/datasets/issues/545
689,138,878
MDU6SXNzdWU2ODkxMzg4Nzg=
545
New release coming up for this library
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "Update: release is planed mid-next week." ]
1,598,873,858,000
1,610,535,544,000
1,610,535,544,000
MEMBER
null
Hi all, A few words on the roadmap for this library. The next release will be a big one and is planed at the end of this week. In addition to the support for indexed datasets (useful for non-parametric models like REALM, RAG, DPR, knn-LM and many other fast dataset retrieval technics), it will: - have support f...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/545/reactions", "total_count": 4, "+1": 0, "-1": 0, "laugh": 0, "hooray": 4, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/545/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/544
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/544/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/544/comments
https://api.github.com/repos/huggingface/datasets/issues/544/events
https://github.com/huggingface/datasets/pull/544
689,062,519
MDExOlB1bGxSZXF1ZXN0NDc2MTc4MDM2
544
[Distributed] Fix load_dataset error when multiprocessing + add test
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,598,866,210,000
1,598,872,511,000
1,598,872,510,000
MEMBER
null
Fix #543 + add test
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/544/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/544/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/544", "html_url": "https://github.com/huggingface/datasets/pull/544", "diff_url": "https://github.com/huggingface/datasets/pull/544.diff", "patch_url": "https://github.com/huggingface/datasets/pull/544.patch", "merged_at": 1598872510000 }
true
https://api.github.com/repos/huggingface/datasets/issues/543
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/543/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/543/comments
https://api.github.com/repos/huggingface/datasets/issues/543/events
https://github.com/huggingface/datasets/issues/543
688,644,407
MDU6SXNzdWU2ODg2NDQ0MDc=
543
nlp.load_dataset is not safe for multi processes when loading from local files
{ "login": "luyug", "id": 55288513, "node_id": "MDQ6VXNlcjU1Mjg4NTEz", "avatar_url": "https://avatars.githubusercontent.com/u/55288513?v=4", "gravatar_id": "", "url": "https://api.github.com/users/luyug", "html_url": "https://github.com/luyug", "followers_url": "https://api.github.com/users/luyug/follow...
[]
closed
false
null
[]
null
[ "I'll take a look!" ]
1,598,757,634,000
1,598,872,510,000
1,598,872,510,000
NONE
null
Loading from local files, e.g., `dataset = nlp.load_dataset('csv', data_files=['file_1.csv', 'file_2.csv'])` concurrently from multiple processes, will raise `FileExistsError` from builder's line 430, https://github.com/huggingface/nlp/blob/6655008c738cb613c522deb3bd18e35a67b2a7e5/src/nlp/builder.py#L423-L438 Likel...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/543/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/543/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/542
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/542/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/542/comments
https://api.github.com/repos/huggingface/datasets/issues/542/events
https://github.com/huggingface/datasets/pull/542
688,555,036
MDExOlB1bGxSZXF1ZXN0NDc1NzkyNTY0
542
Add TensorFlow example
{ "login": "jplu", "id": 959590, "node_id": "MDQ6VXNlcjk1OTU5MA==", "avatar_url": "https://avatars.githubusercontent.com/u/959590?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jplu", "html_url": "https://github.com/jplu", "followers_url": "https://api.github.com/users/jplu/followers", ...
[]
closed
false
null
[]
null
[]
1,598,715,567,000
1,598,867,360,000
1,598,867,359,000
CONTRIBUTOR
null
Update the Quick Tour documentation in order to add the TensorFlow equivalent source code for the classification example. Now it is possible to select either the code in PyTorch or in TensorFlow in the Quick tour.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/542/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/542/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/542", "html_url": "https://github.com/huggingface/datasets/pull/542", "diff_url": "https://github.com/huggingface/datasets/pull/542.diff", "patch_url": "https://github.com/huggingface/datasets/pull/542.patch", "merged_at": 1598867359000 }
true
https://api.github.com/repos/huggingface/datasets/issues/541
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/541/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/541/comments
https://api.github.com/repos/huggingface/datasets/issues/541/events
https://github.com/huggingface/datasets/issues/541
688,521,224
MDU6SXNzdWU2ODg1MjEyMjQ=
541
Best practices for training tokenizers with nlp
{ "login": "moskomule", "id": 11806234, "node_id": "MDQ6VXNlcjExODA2MjM0", "avatar_url": "https://avatars.githubusercontent.com/u/11806234?v=4", "gravatar_id": "", "url": "https://api.github.com/users/moskomule", "html_url": "https://github.com/moskomule", "followers_url": "https://api.github.com/users/...
[]
open
false
null
[]
null
[]
1,598,702,809,000
1,598,702,820,000
null
NONE
null
Hi, thank you for developing this library. What do you think are the best practices for training tokenizers using `nlp`? In the document and examples, I could only find pre-trained tokenizers used.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/541/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/541/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/540
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/540/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/540/comments
https://api.github.com/repos/huggingface/datasets/issues/540/events
https://github.com/huggingface/datasets/pull/540
688,475,884
MDExOlB1bGxSZXF1ZXN0NDc1NzMzNzMz
540
[BUGFIX] Fix Race Dataset Checksum bug
{ "login": "abarbosa94", "id": 6608232, "node_id": "MDQ6VXNlcjY2MDgyMzI=", "avatar_url": "https://avatars.githubusercontent.com/u/6608232?v=4", "gravatar_id": "", "url": "https://api.github.com/users/abarbosa94", "html_url": "https://github.com/abarbosa94", "followers_url": "https://api.github.com/users...
[]
closed
false
null
[]
null
[ "I'm not sure this would fix #537 .\r\nHowever your point about the missing `middle` data is right and we probably want to include these data as well.\r\nDo you think it would we worth having different configurations for this dataset for users who want to only load part of it (`high school` or `middle` or `all`) ?"...
1,598,684,410,000
1,600,429,340,000
1,600,429,340,000
CONTRIBUTOR
null
In #537 I noticed that there was a bug in checksum checking when I have tried to download the race dataset. The reason for this is that the current preprocessing was just considering the `high school` data and it was ignoring the `middle` one. This PR just fixes it :) Moreover, I have added some descriptions.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/540/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/540/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/540", "html_url": "https://github.com/huggingface/datasets/pull/540", "diff_url": "https://github.com/huggingface/datasets/pull/540.diff", "patch_url": "https://github.com/huggingface/datasets/pull/540.patch", "merged_at": 1600429340000 }
true
https://api.github.com/repos/huggingface/datasets/issues/539
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/539/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/539/comments
https://api.github.com/repos/huggingface/datasets/issues/539/events
https://github.com/huggingface/datasets/issues/539
688,323,602
MDU6SXNzdWU2ODgzMjM2MDI=
539
[Dataset] `NonMatchingChecksumError` due to an update in the LinCE benchmark data
{ "login": "gaguilar", "id": 5833357, "node_id": "MDQ6VXNlcjU4MzMzNTc=", "avatar_url": "https://avatars.githubusercontent.com/u/5833357?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gaguilar", "html_url": "https://github.com/gaguilar", "followers_url": "https://api.github.com/users/gagui...
[]
closed
false
null
[]
null
[ "Hi @gaguilar \r\n\r\nIf you want to take care of this, it very simple, you just need to regenerate the `dataset_infos.json` file as indicated [in the doc](https://huggingface.co/nlp/share_dataset.html#adding-metadata) by [installing from source](https://huggingface.co/nlp/installation.html#installing-from-source) ...
1,598,644,551,000
1,599,150,842,000
1,599,150,841,000
CONTRIBUTOR
null
Hi, There is a `NonMatchingChecksumError` error for the `lid_msaea` (language identification for Modern Standard Arabic - Egyptian Arabic) dataset from the LinCE benchmark due to a minor update on that dataset. How can I update the checksum of the library to solve this issue? The error is below and it also appea...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/539/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/539/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/538
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/538/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/538/comments
https://api.github.com/repos/huggingface/datasets/issues/538/events
https://github.com/huggingface/datasets/pull/538
688,015,912
MDExOlB1bGxSZXF1ZXN0NDc1MzU3MjY2
538
[logging] Add centralized logging - Bump-up cache loads to warnings
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,598,614,949,000
1,598,874,171,000
1,598,874,171,000
MEMBER
null
Add a `nlp.logging` module to set the global logging level easily. The verbosity level also controls the tqdm bars (disabled when set higher than INFO). You can use: ``` nlp.logging.set_verbosity(verbosity: int) nlp.logging.set_verbosity_info() nlp.logging.set_verbosity_warning() nlp.logging.set_verbosity_debug...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/538/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/538/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/538", "html_url": "https://github.com/huggingface/datasets/pull/538", "diff_url": "https://github.com/huggingface/datasets/pull/538.diff", "patch_url": "https://github.com/huggingface/datasets/pull/538.patch", "merged_at": 1598874170000 }
true
https://api.github.com/repos/huggingface/datasets/issues/537
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/537/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/537/comments
https://api.github.com/repos/huggingface/datasets/issues/537/events
https://github.com/huggingface/datasets/issues/537
687,614,699
MDU6SXNzdWU2ODc2MTQ2OTk=
537
[Dataset] RACE dataset Checksums error
{ "login": "abarbosa94", "id": 6608232, "node_id": "MDQ6VXNlcjY2MDgyMzI=", "avatar_url": "https://avatars.githubusercontent.com/u/6608232?v=4", "gravatar_id": "", "url": "https://api.github.com/users/abarbosa94", "html_url": "https://github.com/abarbosa94", "followers_url": "https://api.github.com/users...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
null
[]
null
[ "`NonMatchingChecksumError` means that the checksum of the downloaded file is not the expected one.\r\nEither the file you downloaded was corrupted along the way, or the host updated the file.\r\nCould you try to clear your cache and run `load_dataset` again ? If the error is still there, it means that there was an...
1,598,572,696,000
1,600,430,824,000
1,600,430,824,000
CONTRIBUTOR
null
Hi there, I just would like to use this awesome lib to perform a dataset fine-tuning on RACE dataset. I have performed the following steps: ``` dataset = nlp.load_dataset("race") len(dataset["train"]), len(dataset["validation"]) ``` But then I got the following error: ``` ----------------------------------...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/537/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/537/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/536
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/536/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/536/comments
https://api.github.com/repos/huggingface/datasets/issues/536/events
https://github.com/huggingface/datasets/pull/536
687,378,332
MDExOlB1bGxSZXF1ZXN0NDc0ODE0NzY1
536
Fingerprint
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "I changed the way I implemented fingerprint updates to use decorator functions.\r\n\r\nI also added a new attribute called `_inplace_history` that stores the in-place history of transforms (like cast_, rename_columns, etc.). This history is useful to replay the changes that were done in-place when unpickling a dat...
1,598,545,629,000
1,598,883,640,000
1,598,883,639,000
MEMBER
null
This PR is a continuation of #513 , in which many in-place functions were introduced or updated (cast_, flatten_) etc. However the caching didn't handle these changes. Indeed the caching took into account only the previous cache file name of the table, and not the possible in-place transforms of the table. To fix t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/536/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/536/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/536", "html_url": "https://github.com/huggingface/datasets/pull/536", "diff_url": "https://github.com/huggingface/datasets/pull/536.diff", "patch_url": "https://github.com/huggingface/datasets/pull/536.patch", "merged_at": 1598883639000 }
true
https://api.github.com/repos/huggingface/datasets/issues/535
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/535/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/535/comments
https://api.github.com/repos/huggingface/datasets/issues/535/events
https://github.com/huggingface/datasets/pull/535
686,238,315
MDExOlB1bGxSZXF1ZXN0NDczODM3Njg0
535
Benchmarks
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,598,440,886,000
1,598,517,600,000
1,598,517,599,000
MEMBER
null
Adding some benchmarks with DVC/CML To add a new tracked benchmark: - create a new python benchmarking script in `./benchmarks/`. The script can use the utilities in `./benchmarks/utils.py` and should output a JSON file with results in `./benchmarks/results/`. - add a new pipeline stage in [dvc.yaml](./dvc.yaml) w...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/535/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/535/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/535", "html_url": "https://github.com/huggingface/datasets/pull/535", "diff_url": "https://github.com/huggingface/datasets/pull/535.diff", "patch_url": "https://github.com/huggingface/datasets/pull/535.patch", "merged_at": 1598517599000 }
true
https://api.github.com/repos/huggingface/datasets/issues/534
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/534/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/534/comments
https://api.github.com/repos/huggingface/datasets/issues/534/events
https://github.com/huggingface/datasets/issues/534
686,115,912
MDU6SXNzdWU2ODYxMTU5MTI=
534
`list_datasets()` is broken.
{ "login": "ashutosh-dwivedi-e3502", "id": 314169, "node_id": "MDQ6VXNlcjMxNDE2OQ==", "avatar_url": "https://avatars.githubusercontent.com/u/314169?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ashutosh-dwivedi-e3502", "html_url": "https://github.com/ashutosh-dwivedi-e3502", "followers_u...
[]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nThis has been fixed in #475 and the fix will be available in the next release", "What you can do instead to get the list of the datasets is call\r\n\r\n```python\r\nprint([dataset.id for dataset in nlp.list_datasets()])\r\n```", "Thanks @lhoestq . " ]
1,598,429,941,000
1,598,509,871,000
1,598,509,871,000
NONE
null
version = '0.4.0' `list_datasets()` is broken. It results in the following error : ``` In [3]: nlp.list_datasets() Out[3]: --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) ~/.virtualenvs/san-lgUCsFg_/lib/py...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/534/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/534/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/533
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/533/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/533/comments
https://api.github.com/repos/huggingface/datasets/issues/533/events
https://github.com/huggingface/datasets/pull/533
685,585,914
MDExOlB1bGxSZXF1ZXN0NDczMjg4OTgx
533
Fix ArrayXD for pyarrow 0.17.1 by using non fixed length list arrays
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,598,369,564,000
1,598,428,944,000
1,598,428,943,000
MEMBER
null
It should fix the CI problems in #513
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/533/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/533/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/533", "html_url": "https://github.com/huggingface/datasets/pull/533", "diff_url": "https://github.com/huggingface/datasets/pull/533.diff", "patch_url": "https://github.com/huggingface/datasets/pull/533.patch", "merged_at": 1598428943000 }
true
https://api.github.com/repos/huggingface/datasets/issues/532
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/532/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/532/comments
https://api.github.com/repos/huggingface/datasets/issues/532/events
https://github.com/huggingface/datasets/issues/532
685,540,614
MDU6SXNzdWU2ODU1NDA2MTQ=
532
File exists error when used with TPU
{ "login": "go-inoue", "id": 20531705, "node_id": "MDQ6VXNlcjIwNTMxNzA1", "avatar_url": "https://avatars.githubusercontent.com/u/20531705?v=4", "gravatar_id": "", "url": "https://api.github.com/users/go-inoue", "html_url": "https://github.com/go-inoue", "followers_url": "https://api.github.com/users/go-...
[]
open
false
null
[]
null
[ "I am facing probably facing similar issues with \r\n\r\n`wiki40b_en_100_0`", "Could you try to run `dataset = load_dataset(\"text\", data_files=file_path, split=\"train\")` once before calling the script ?\r\n\r\nIt looks like several processes try to create the dataset in arrow format at the same time. If the d...
1,598,366,198,000
1,598,962,496,000
null
NONE
null
Hi, I'm getting a "File exists" error when I use [text dataset](https://github.com/huggingface/nlp/tree/master/datasets/text) for pre-training a RoBERTa model using `transformers` (3.0.2) and `nlp`(0.4.0) on a VM with TPU (v3-8). I modified [line 131 in the original `run_language_modeling.py`](https://github.com/...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/532/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/532/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/531
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/531/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/531/comments
https://api.github.com/repos/huggingface/datasets/issues/531/events
https://github.com/huggingface/datasets/pull/531
685,291,036
MDExOlB1bGxSZXF1ZXN0NDczMDM4ODc4
531
add concatenate_datasets to the docs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,598,344,805,000
1,598,346,140,000
1,598,346,139,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/531/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/531/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/531", "html_url": "https://github.com/huggingface/datasets/pull/531", "diff_url": "https://github.com/huggingface/datasets/pull/531.diff", "patch_url": "https://github.com/huggingface/datasets/pull/531.patch", "merged_at": 1598346139000 }
true
https://api.github.com/repos/huggingface/datasets/issues/530
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/530/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/530/comments
https://api.github.com/repos/huggingface/datasets/issues/530/events
https://github.com/huggingface/datasets/pull/530
684,825,612
MDExOlB1bGxSZXF1ZXN0NDcyNjQ5NTk2
530
use ragged tensor by default
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "Yes I agree. Maybe something that lets specify different format depending on the column ? Especially to better control dtype and shape (and ragged for tf)\r\n\r\nOh and I forgot: this one should also fix the second issue found in #477 for the next release", "I am running into the same issue with the error messag...
1,598,288,775,000
1,634,931,520,000
1,598,296,945,000
MEMBER
null
I think it's better if it's clear whether the returned tensor is ragged or not when the type is set to tensorflow. Previously it was a tensor (not ragged) if numpy could stack the output (which can change depending on the batch of example you take), which make things difficult to handle, as it may sometimes return a r...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/530/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/530/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/530", "html_url": "https://github.com/huggingface/datasets/pull/530", "diff_url": "https://github.com/huggingface/datasets/pull/530.diff", "patch_url": "https://github.com/huggingface/datasets/pull/530.patch", "merged_at": 1598296945000 }
true
https://api.github.com/repos/huggingface/datasets/issues/529
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/529/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/529/comments
https://api.github.com/repos/huggingface/datasets/issues/529/events
https://github.com/huggingface/datasets/pull/529
684,797,157
MDExOlB1bGxSZXF1ZXN0NDcyNjI2MDY4
529
Add MLSUM
{ "login": "RachelKer", "id": 36986299, "node_id": "MDQ6VXNlcjM2OTg2Mjk5", "avatar_url": "https://avatars.githubusercontent.com/u/36986299?v=4", "gravatar_id": "", "url": "https://api.github.com/users/RachelKer", "html_url": "https://github.com/RachelKer", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Could you test to run the test using the changes in #527 and let me know if it fixes the issue ? If so I'll merge it and we'll be good to go :)", "Hello, it does work on the fixing real dataset branch. Merci Quentin :)", "Nice, glad to hear that :)\r\nde rien !" ]
1,598,285,915,000
1,598,429,051,000
1,598,429,051,000
CONTRIBUTOR
null
Hello (again :) !), So, I started a new branch because of a [rebase issue](https://github.com/huggingface/nlp/pull/463), sorry for the mess. However, the command `pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_mlsum` still fails because there is no default language dataset : the s...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/529/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/529/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/529", "html_url": "https://github.com/huggingface/datasets/pull/529", "diff_url": "https://github.com/huggingface/datasets/pull/529.diff", "patch_url": "https://github.com/huggingface/datasets/pull/529.patch", "merged_at": 1598429050000 }
true
https://api.github.com/repos/huggingface/datasets/issues/528
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/528/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/528/comments
https://api.github.com/repos/huggingface/datasets/issues/528/events
https://github.com/huggingface/datasets/pull/528
684,673,673
MDExOlB1bGxSZXF1ZXN0NDcyNTIzNDI1
528
fix missing variable names in docs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "The problem came from `default: ` that is rendered differently and hides the parameter names. I changed `default: ...` to `defaults to ...`" ]
1,598,275,908,000
1,598,346,244,000
1,598,346,243,000
MEMBER
null
fix #524
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/528/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/528/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/528", "html_url": "https://github.com/huggingface/datasets/pull/528", "diff_url": "https://github.com/huggingface/datasets/pull/528.diff", "patch_url": "https://github.com/huggingface/datasets/pull/528.patch", "merged_at": 1598346243000 }
true
https://api.github.com/repos/huggingface/datasets/issues/527
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/527/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/527/comments
https://api.github.com/repos/huggingface/datasets/issues/527/events
https://github.com/huggingface/datasets/pull/527
684,632,930
MDExOlB1bGxSZXF1ZXN0NDcyNDg4MzUy
527
Fix config used for slow test on real dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,598,272,774,000
1,598,347,245,000
1,598,347,244,000
MEMBER
null
As noticed in #470, #474, #476, #504 , the slow test `test_load_real_dataset` couldn't run on datasets that require config parameters. To fix that I replaced it with one test with the first config of BUILDER_CONFIGS `test_load_real_dataset`, and another test that runs all of the configs in BUILDER_CONFIGS `test_load...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/527/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/527/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/527", "html_url": "https://github.com/huggingface/datasets/pull/527", "diff_url": "https://github.com/huggingface/datasets/pull/527.diff", "patch_url": "https://github.com/huggingface/datasets/pull/527.patch", "merged_at": 1598347244000 }
true
https://api.github.com/repos/huggingface/datasets/issues/526
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/526/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/526/comments
https://api.github.com/repos/huggingface/datasets/issues/526/events
https://github.com/huggingface/datasets/pull/526
684,615,455
MDExOlB1bGxSZXF1ZXN0NDcyNDczNjcw
526
Returning None instead of "python" if dataset is unformatted
{ "login": "TevenLeScao", "id": 26709476, "node_id": "MDQ6VXNlcjI2NzA5NDc2", "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TevenLeScao", "html_url": "https://github.com/TevenLeScao", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[ "We have to change the tests to expect `None` instead of `python` then", "Merging!" ]
1,598,271,035,000
1,598,273,443,000
1,598,273,442,000
MEMBER
null
Following the discussion on Slack, this small fix ensures that calling `dataset.set_format(type=dataset.format["type"])` works properly. Slightly breaking as calling `dataset.format` when the dataset is unformatted will return `None` instead of `python`.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/526/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/526/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/526", "html_url": "https://github.com/huggingface/datasets/pull/526", "diff_url": "https://github.com/huggingface/datasets/pull/526.diff", "patch_url": "https://github.com/huggingface/datasets/pull/526.patch", "merged_at": 1598273442000 }
true
https://api.github.com/repos/huggingface/datasets/issues/525
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/525/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/525/comments
https://api.github.com/repos/huggingface/datasets/issues/525/events
https://github.com/huggingface/datasets/issues/525
683,875,483
MDU6SXNzdWU2ODM4NzU0ODM=
525
wmt download speed example
{ "login": "sshleifer", "id": 6045025, "node_id": "MDQ6VXNlcjYwNDUwMjU=", "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sshleifer", "html_url": "https://github.com/sshleifer", "followers_url": "https://api.github.com/users/ss...
[]
open
false
null
[]
null
[ "Thanks for creating the issue :)\r\nThe download link for wmt-en-de raw looks like a mirror. We should use that instead of the current url.\r\nIs this mirror official ?\r\n\r\nAlso it looks like for `ro-en` it tried to download other languages. If we manage to only download the one that is asked it'd be cool\r\n\r...
1,598,052,546,000
1,613,664,967,000
null
CONTRIBUTOR
null
Continuing from the slack 1.0 roadmap thread w @lhoestq , I realized the slow downloads is only a thing sometimes. Here are a few examples, I suspect there are multiple issues. All commands were run from the same gcp us-central-1f machine. ``` import nlp nlp.load_dataset('wmt16', 'de-en') ``` Downloads at 49.1 K...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/525/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/525/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/524
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/524/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/524/comments
https://api.github.com/repos/huggingface/datasets/issues/524/events
https://github.com/huggingface/datasets/issues/524
683,686,359
MDU6SXNzdWU2ODM2ODYzNTk=
524
Some docs are missing parameter names
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "Indeed, good catch!" ]
1,598,028,454,000
1,598,346,243,000
1,598,346,243,000
CONTRIBUTOR
null
See https://huggingface.co/nlp/master/package_reference/main_classes.html#nlp.Dataset.map. I believe this is because the parameter names are enclosed in backticks in the docstrings, maybe it's an old docstring format that doesn't work with the current Sphinx version.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/524/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/524/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/523
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/523/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/523/comments
https://api.github.com/repos/huggingface/datasets/issues/523/events
https://github.com/huggingface/datasets/pull/523
682,573,232
MDExOlB1bGxSZXF1ZXN0NDcwNzkxMjA1
523
Speed up Tokenization by optimizing cast_to_python_objects
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "I took your comments into account and added tests for `cast_to_python_objects`" ]
1,597,916,522,000
1,598,259,255,000
1,598,259,254,000
MEMBER
null
I changed how `cast_to_python_objects` works to make it faster. It is used to cast numpy/pytorch/tensorflow/pandas objects to python lists, and it works recursively. To avoid iterating over possibly long lists, it first checks if the first element that is not None has to be casted. If the first element needs to be...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/523/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/523/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/523", "html_url": "https://github.com/huggingface/datasets/pull/523", "diff_url": "https://github.com/huggingface/datasets/pull/523.diff", "patch_url": "https://github.com/huggingface/datasets/pull/523.patch", "merged_at": 1598259254000 }
true
https://api.github.com/repos/huggingface/datasets/issues/522
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/522/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/522/comments
https://api.github.com/repos/huggingface/datasets/issues/522/events
https://github.com/huggingface/datasets/issues/522
682,478,833
MDU6SXNzdWU2ODI0Nzg4MzM=
522
dictionnary typo in docs
{ "login": "yonigottesman", "id": 4004127, "node_id": "MDQ6VXNlcjQwMDQxMjc=", "avatar_url": "https://avatars.githubusercontent.com/u/4004127?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yonigottesman", "html_url": "https://github.com/yonigottesman", "followers_url": "https://api.github....
[]
closed
false
null
[]
null
[ "Thanks!" ]
1,597,907,465,000
1,597,909,934,000
1,597,909,933,000
CONTRIBUTOR
null
Many places dictionary is spelled dictionnary, not sure if its on purpose or not. Fixed in this pr: https://github.com/huggingface/nlp/pull/521
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/522/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/522/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/521
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/521/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/521/comments
https://api.github.com/repos/huggingface/datasets/issues/521/events
https://github.com/huggingface/datasets/pull/521
682,477,648
MDExOlB1bGxSZXF1ZXN0NDcwNzEyNzgz
521
Fix dictionnary (dictionary) typo
{ "login": "yonigottesman", "id": 4004127, "node_id": "MDQ6VXNlcjQwMDQxMjc=", "avatar_url": "https://avatars.githubusercontent.com/u/4004127?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yonigottesman", "html_url": "https://github.com/yonigottesman", "followers_url": "https://api.github....
[]
closed
false
null
[]
null
[ "Hahah thanks Yonatan. It was not on purpose, we are just not very good at spelling :)" ]
1,597,907,342,000
1,597,909,924,000
1,597,909,924,000
CONTRIBUTOR
null
This error happens many times I'm thinking maybe its spelled like this on purpose?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/521/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/521/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/521", "html_url": "https://github.com/huggingface/datasets/pull/521", "diff_url": "https://github.com/huggingface/datasets/pull/521.diff", "patch_url": "https://github.com/huggingface/datasets/pull/521.patch", "merged_at": 1597909924000 }
true
https://api.github.com/repos/huggingface/datasets/issues/520
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/520/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/520/comments
https://api.github.com/repos/huggingface/datasets/issues/520/events
https://github.com/huggingface/datasets/pull/520
682,264,839
MDExOlB1bGxSZXF1ZXN0NDcwNTI4MDE0
520
Transform references for sacrebleu
{ "login": "jbragg", "id": 2238344, "node_id": "MDQ6VXNlcjIyMzgzNDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2238344?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jbragg", "html_url": "https://github.com/jbragg", "followers_url": "https://api.github.com/users/jbragg/foll...
[]
closed
false
null
[]
null
[ "I think I agree @lhoestq so I pushed a change.\r\nThanks for your work on the library!" ]
1,597,883,215,000
1,597,915,854,000
1,597,915,853,000
CONTRIBUTOR
null
Currently it is impossible to use sacrebleu when len(predictions) != the number of references per prediction (very uncommon), due to a strange format expected by sacrebleu. If one passes in the data to `nlp.metric.compute()` in sacrebleu format, `nlp` throws an error due to mismatching lengths between predictions and r...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/520/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/520/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/520", "html_url": "https://github.com/huggingface/datasets/pull/520", "diff_url": "https://github.com/huggingface/datasets/pull/520.diff", "patch_url": "https://github.com/huggingface/datasets/pull/520.patch", "merged_at": 1597915853000 }
true
https://api.github.com/repos/huggingface/datasets/issues/519
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/519/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/519/comments
https://api.github.com/repos/huggingface/datasets/issues/519/events
https://github.com/huggingface/datasets/issues/519
682,193,882
MDU6SXNzdWU2ODIxOTM4ODI=
519
[BUG] Metrics throwing new error on master since 0.4.0
{ "login": "jbragg", "id": 2238344, "node_id": "MDQ6VXNlcjIyMzgzNDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2238344?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jbragg", "html_url": "https://github.com/jbragg", "followers_url": "https://api.github.com/users/jbragg/foll...
[]
closed
false
null
[]
null
[ "Update - maybe this is only failing on bleu because I was not tokenizing inputs to the metric", "Closing - seems to be just forgetting to tokenize. And found the helpful discussion in #137 " ]
1,597,872,555,000
1,597,874,680,000
1,597,874,680,000
CONTRIBUTOR
null
The following error occurs when passing in references of type `List[List[str]]` to metrics like bleu. Wasn't happening on 0.4.0 but happening now on master. ``` File "/usr/local/lib/python3.7/site-packages/nlp/metric.py", line 226, in compute self.add_batch(predictions=predictions, references=references) ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/519/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/519/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/518
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/518/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/518/comments
https://api.github.com/repos/huggingface/datasets/issues/518/events
https://github.com/huggingface/datasets/pull/518
682,131,165
MDExOlB1bGxSZXF1ZXN0NDcwNDE0ODE1
518
[METRICS, breaking] Refactor caching behavior, pickle/cloudpickle metrics and dataset, add tests on metrics
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "(test failure is unrelated)", "As discussed with @thomwolf merging since the hyperparameter-search has been merged in transformers." ]
1,597,866,188,000
1,598,284,900,000
1,598,284,899,000
MEMBER
null
Move the acquisition of the filelock at a later stage during metrics processing so it can be pickled/cloudpickled after instantiation. Also add some tests on pickling, concurrent but separate metric instances and concurrent and distributed metric instances. Changes significantly the caching behavior for the metri...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/518/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/518/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/518", "html_url": "https://github.com/huggingface/datasets/pull/518", "diff_url": "https://github.com/huggingface/datasets/pull/518.diff", "patch_url": "https://github.com/huggingface/datasets/pull/518.patch", "merged_at": 1598284899000 }
true
https://api.github.com/repos/huggingface/datasets/issues/517
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/517/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/517/comments
https://api.github.com/repos/huggingface/datasets/issues/517/events
https://github.com/huggingface/datasets/issues/517
681,896,944
MDU6SXNzdWU2ODE4OTY5NDQ=
517
add MLDoc dataset
{ "login": "jxmorris12", "id": 13238952, "node_id": "MDQ6VXNlcjEzMjM4OTUy", "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jxmorris12", "html_url": "https://github.com/jxmorris12", "followers_url": "https://api.github.com/use...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[ "Any updates on this?", "This request is still an open issue waiting to be addressed by any community member, @GuillemGSubies." ]
1,597,848,119,000
1,627,970,373,000
null
CONTRIBUTOR
null
Hi, I am recommending that someone add MLDoc, a multilingual news topic classification dataset. - Here's a link to the Github: https://github.com/facebookresearch/MLDoc - and the paper: http://www.lrec-conf.org/proceedings/lrec2018/pdf/658.pdf Looks like the dataset contains news stories in multiple languages...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/517/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/517/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/516
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/516/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/516/comments
https://api.github.com/repos/huggingface/datasets/issues/516/events
https://github.com/huggingface/datasets/pull/516
681,846,032
MDExOlB1bGxSZXF1ZXN0NDcwMTY5NTA0
516
[Breaking] Rename formated to formatted
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,597,844,123,000
1,597,912,877,000
1,597,912,876,000
MEMBER
null
`formated` is not correct but `formatted` is
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/516/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/516/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/516", "html_url": "https://github.com/huggingface/datasets/pull/516", "diff_url": "https://github.com/huggingface/datasets/pull/516.diff", "patch_url": "https://github.com/huggingface/datasets/pull/516.patch", "merged_at": 1597912876000 }
true
https://api.github.com/repos/huggingface/datasets/issues/515
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/515/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/515/comments
https://api.github.com/repos/huggingface/datasets/issues/515/events
https://github.com/huggingface/datasets/pull/515
681,845,619
MDExOlB1bGxSZXF1ZXN0NDcwMTY5MTQ0
515
Fix batched map for formatted dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,597,844,090,000
1,597,955,443,000
1,597,955,442,000
MEMBER
null
If you had a dataset formatted as numpy for example, and tried to do a batched map, then it would crash because one of the elements from the inputs was missing for unchanged columns (ex: batch of length 999 instead of 1000). The happened during the creation of the `pa.Table`, since columns had different lengths.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/515/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/515/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/515", "html_url": "https://github.com/huggingface/datasets/pull/515", "diff_url": "https://github.com/huggingface/datasets/pull/515.diff", "patch_url": "https://github.com/huggingface/datasets/pull/515.patch", "merged_at": 1597955442000 }
true
https://api.github.com/repos/huggingface/datasets/issues/514
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/514/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/514/comments
https://api.github.com/repos/huggingface/datasets/issues/514/events
https://github.com/huggingface/datasets/issues/514
681,256,348
MDU6SXNzdWU2ODEyNTYzNDg=
514
dataset.shuffle(keep_in_memory=True) is never allowed
{ "login": "vegarab", "id": 24683907, "node_id": "MDQ6VXNlcjI0NjgzOTA3", "avatar_url": "https://avatars.githubusercontent.com/u/24683907?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vegarab", "html_url": "https://github.com/vegarab", "followers_url": "https://api.github.com/users/vegara...
[]
open
false
null
[]
null
[ "This seems to be fixed in #513 for the filter function, replacing `cache_file_name` with `indices_cache_file_name` in the assert. Although not for the `map()` function @thomwolf ", "Maybe I'm a bit tired but I fail to see the issue here.\r\n\r\nSince `cache_file_name` is `None` by default, if you set `keep_in_me...
1,597,776,460,000
1,627,063,631,000
null
CONTRIBUTOR
null
As of commit ef4aac2, the usage of the parameter `keep_in_memory=True` is never possible: `dataset.select(keep_in_memory=True)` The commit added the lines ```python # lines 994-996 in src/nlp/arrow_dataset.py assert ( not keep_in_memory or cache_file_name is None ), "Please use either...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/514/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/514/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/513
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/513/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/513/comments
https://api.github.com/repos/huggingface/datasets/issues/513/events
https://github.com/huggingface/datasets/pull/513
681,215,612
MDExOlB1bGxSZXF1ZXN0NDY5NjQxMjg1
513
[speedup] Use indices mappings instead of deepcopy for all the samples reordering methods
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "Ok I fixed `concatenate_datasets` and added tests\r\nFeel free to merge if it's good for you @thomwolf ", "Ok, adding some benchmarks for map/filters and then I'll merge", "Warning from pytorch that we should maybe consider at some point @lhoestq:\r\n```\r\n/__w/nlp/nlp/src/nlp/arrow_dataset.py:648: UserWarnin...
1,597,772,162,000
1,598,604,111,000
1,598,604,110,000
MEMBER
null
Use an indices mapping instead of rewriting the dataset for all the samples re-ordering/selection methods (`select`, `sort`, `shuffle`, `shard`, `train_test_split`). Added a `flatten_indices` method which copy the dataset to a new table to remove the indices mapping with tests. All the samples re-ordering/selecti...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/513/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 1, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/513/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/513", "html_url": "https://github.com/huggingface/datasets/pull/513", "diff_url": "https://github.com/huggingface/datasets/pull/513.diff", "patch_url": "https://github.com/huggingface/datasets/pull/513.patch", "merged_at": 1598604110000 }
true
https://api.github.com/repos/huggingface/datasets/issues/512
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/512/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/512/comments
https://api.github.com/repos/huggingface/datasets/issues/512/events
https://github.com/huggingface/datasets/pull/512
681,137,164
MDExOlB1bGxSZXF1ZXN0NDY5NTc2NzE3
512
Delete CONTRIBUTING.md
{ "login": "ChenZehong13", "id": 56394989, "node_id": "MDQ6VXNlcjU2Mzk0OTg5", "avatar_url": "https://avatars.githubusercontent.com/u/56394989?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ChenZehong13", "html_url": "https://github.com/ChenZehong13", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "😱", "Yeah, this is spammy behavior. I've reported the user handle." ]
1,597,764,805,000
1,597,765,701,000
1,597,765,147,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/512/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/512/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/512", "html_url": "https://github.com/huggingface/datasets/pull/512", "diff_url": "https://github.com/huggingface/datasets/pull/512.diff", "patch_url": "https://github.com/huggingface/datasets/pull/512.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/511
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/511/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/511/comments
https://api.github.com/repos/huggingface/datasets/issues/511/events
https://github.com/huggingface/datasets/issues/511
681,055,553
MDU6SXNzdWU2ODEwNTU1NTM=
511
dataset.shuffle() and select() resets format. Intended?
{ "login": "vegarab", "id": 24683907, "node_id": "MDQ6VXNlcjI0NjgzOTA3", "avatar_url": "https://avatars.githubusercontent.com/u/24683907?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vegarab", "html_url": "https://github.com/vegarab", "followers_url": "https://api.github.com/users/vegara...
[]
closed
false
null
[]
null
[ "Hi @vegarab yes feel free to open a discussion here.\r\n\r\nThis design choice was not very much thought about.\r\n\r\nSince `dataset.select()` (like all the method without a trailing underscore) is non-destructive and returns a new dataset it has most of its properties initialized from scratch (except the table a...
1,597,758,361,000
1,600,073,138,000
1,600,073,138,000
CONTRIBUTOR
null
Calling `dataset.shuffle()` or `dataset.select()` on a dataset resets its format set by `dataset.set_format()`. Is this intended or an oversight? When working on quite large datasets that require a lot of preprocessing I find it convenient to save the processed dataset to file using `torch.save("dataset.pt")`. Later...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/511/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/511/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/510
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/510/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/510/comments
https://api.github.com/repos/huggingface/datasets/issues/510/events
https://github.com/huggingface/datasets/issues/510
680,823,644
MDU6SXNzdWU2ODA4MjM2NDQ=
510
Version of numpy to use the library
{ "login": "isspek", "id": 6966175, "node_id": "MDQ6VXNlcjY5NjYxNzU=", "avatar_url": "https://avatars.githubusercontent.com/u/6966175?v=4", "gravatar_id": "", "url": "https://api.github.com/users/isspek", "html_url": "https://github.com/isspek", "followers_url": "https://api.github.com/users/isspek/foll...
[]
closed
false
null
[]
null
[ "Seems like this method was added in 1.17. I'll add a requirement on this.", "Thank you so much. After upgrading the numpy library, it worked." ]
1,597,741,153,000
1,597,862,156,000
1,597,862,156,000
NONE
null
Thank you so much for your excellent work! I would like to use nlp library in my project. While importing nlp, I am receiving the following error `AttributeError: module 'numpy.random' has no attribute 'Generator'` Numpy version in my project is 1.16.0. May I learn which numpy version is used for the nlp library. Th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/510/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/510/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/509
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/509/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/509/comments
https://api.github.com/repos/huggingface/datasets/issues/509/events
https://github.com/huggingface/datasets/issues/509
679,711,585
MDU6SXNzdWU2Nzk3MTE1ODU=
509
Converting TensorFlow dataset example
{ "login": "saareliad", "id": 22762845, "node_id": "MDQ6VXNlcjIyNzYyODQ1", "avatar_url": "https://avatars.githubusercontent.com/u/22762845?v=4", "gravatar_id": "", "url": "https://api.github.com/users/saareliad", "html_url": "https://github.com/saareliad", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Do you want to convert a dataset script to the tfds format ?\r\nIf so, we currently have a comversion script nlp/commands/convert.py but it is a conversion script that goes from tfds to nlp.\r\nI think it shouldn't be too hard to do the changes in reverse (at some manual adjustments).\r\nIf you manage to make it w...
1,597,565,120,000
1,627,970,478,000
1,627,970,477,000
NONE
null
Hi, I want to use TensorFlow datasets with this repo, I noticed you made some conversion script, can you give a simple example of using it? Thanks
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/509/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/509/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/508
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/508/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/508/comments
https://api.github.com/repos/huggingface/datasets/issues/508/events
https://github.com/huggingface/datasets/issues/508
679,705,734
MDU6SXNzdWU2Nzk3MDU3MzQ=
508
TypeError: Receiver() takes no arguments
{ "login": "sebastiantomac", "id": 1225851, "node_id": "MDQ6VXNlcjEyMjU4NTE=", "avatar_url": "https://avatars.githubusercontent.com/u/1225851?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sebastiantomac", "html_url": "https://github.com/sebastiantomac", "followers_url": "https://api.gith...
[]
closed
false
null
[]
null
[ "Which version of Apache Beam do you have (can you copy your full environment info here)?", "apache-beam==2.23.0\r\nnlp==0.4.0\r\n\r\nFor me this was resolved by running the same python script on Linux (or really WSL). ", "Do you manage to run a dummy beam pipeline with python on windows ? \r\nYou can test a du...
1,597,562,296,000
1,598,972,013,000
1,598,971,743,000
NONE
null
I am trying to load a wikipedia data set ``` import nlp from nlp import load_dataset dataset = load_dataset("wikipedia", "20200501.en", split="train", cache_dir=data_path, beam_runner='DirectRunner') #dataset = load_dataset('wikipedia', '20200501.sv', cache_dir=data_path, beam_runner='DirectRunner') ``` Th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/508/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/508/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/507
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/507/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/507/comments
https://api.github.com/repos/huggingface/datasets/issues/507/events
https://github.com/huggingface/datasets/issues/507
679,400,683
MDU6SXNzdWU2Nzk0MDA2ODM=
507
Errors when I use
{ "login": "mchari", "id": 30506151, "node_id": "MDQ6VXNlcjMwNTA2MTUx", "avatar_url": "https://avatars.githubusercontent.com/u/30506151?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchari", "html_url": "https://github.com/mchari", "followers_url": "https://api.github.com/users/mchari/fo...
[]
closed
false
null
[]
null
[ "Looks like an issue with 3.0.2 transformers version. Works fine when I use \"master\" version of transformers." ]
1,597,439,037,000
1,597,441,150,000
1,597,441,150,000
NONE
null
I tried the following example code from https://huggingface.co/deepset/roberta-base-squad2 and got errors I am using **transformers 3.0.2** code . from transformers.pipelines import pipeline from transformers.modeling_auto import AutoModelForQuestionAnswering from transformers.tokenization_auto import AutoToke...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/507/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/507/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/506
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/506/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/506/comments
https://api.github.com/repos/huggingface/datasets/issues/506/events
https://github.com/huggingface/datasets/pull/506
679,164,788
MDExOlB1bGxSZXF1ZXN0NDY3OTkwNjc2
506
fix dataset.map for function without outputs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,597,412,422,000
1,597,663,479,000
1,597,663,478,000
MEMBER
null
As noticed in #505 , giving a function that doesn't return anything in `.map` raises an error because of an unreferenced variable. I fixed that and added tests. Thanks @avloss for reporting
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/506/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/506/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/506", "html_url": "https://github.com/huggingface/datasets/pull/506", "diff_url": "https://github.com/huggingface/datasets/pull/506.diff", "patch_url": "https://github.com/huggingface/datasets/pull/506.patch", "merged_at": 1597663478000 }
true
https://api.github.com/repos/huggingface/datasets/issues/505
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/505/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/505/comments
https://api.github.com/repos/huggingface/datasets/issues/505/events
https://github.com/huggingface/datasets/pull/505
678,791,400
MDExOlB1bGxSZXF1ZXN0NDY3NjgxMjY4
505
tmp_file referenced before assignment
{ "login": "avloss", "id": 17853685, "node_id": "MDQ6VXNlcjE3ODUzNjg1", "avatar_url": "https://avatars.githubusercontent.com/u/17853685?v=4", "gravatar_id": "", "url": "https://api.github.com/users/avloss", "html_url": "https://github.com/avloss", "followers_url": "https://api.github.com/users/avloss/fo...
[]
closed
false
null
[]
null
[ "Thanks for reporting the issue ! I'm creating a new PR to fix it and add tests.\r\n(I'm doing a new PR because I know there's some other place where it needs to be fixed)", "I'm closing this one as I created the other PR." ]
1,597,361,253,000
1,597,412,566,000
1,597,412,566,000
NONE
null
Just learning about this library - so might've not set up all the flags correctly, but was getting this error about "tmp_file".
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/505/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/505/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/505", "html_url": "https://github.com/huggingface/datasets/pull/505", "diff_url": "https://github.com/huggingface/datasets/pull/505.diff", "patch_url": "https://github.com/huggingface/datasets/pull/505.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/504
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/504/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/504/comments
https://api.github.com/repos/huggingface/datasets/issues/504/events
https://github.com/huggingface/datasets/pull/504
678,756,211
MDExOlB1bGxSZXF1ZXN0NDY3NjUxOTA5
504
Added downloading to Hyperpartisan news detection
{ "login": "ghomasHudson", "id": 13795113, "node_id": "MDQ6VXNlcjEzNzk1MTEz", "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghomasHudson", "html_url": "https://github.com/ghomasHudson", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Thank you @ghomasHudson for making our dataset available! This is great!", "The test passes since #527 :)" ]
1,597,355,626,000
1,598,516,321,000
1,598,516,321,000
CONTRIBUTOR
null
Following the discussion on Slack and #349, I've updated the hyperpartisan dataset to pull directly from Zenodo rather than manual install, which should make this dataset much more accessible. Many thanks to @johanneskiesel ! Currently doesn't pass `test_load_real_dataset` - I'm using `self.config.name` which is `de...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/504/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/504/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/504", "html_url": "https://github.com/huggingface/datasets/pull/504", "diff_url": "https://github.com/huggingface/datasets/pull/504.diff", "patch_url": "https://github.com/huggingface/datasets/pull/504.patch", "merged_at": 1598516321000 }
true
https://api.github.com/repos/huggingface/datasets/issues/503
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/503/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/503/comments
https://api.github.com/repos/huggingface/datasets/issues/503/events
https://github.com/huggingface/datasets/pull/503
678,726,538
MDExOlB1bGxSZXF1ZXN0NDY3NjI3MTEw
503
CompGuessWhat?! 0.2.0
{ "login": "aleSuglia", "id": 1479733, "node_id": "MDQ6VXNlcjE0Nzk3MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/1479733?v=4", "gravatar_id": "", "url": "https://api.github.com/users/aleSuglia", "html_url": "https://github.com/aleSuglia", "followers_url": "https://api.github.com/users/al...
[]
closed
false
null
[]
null
[ "I don't see any significant change in the dataset script (except the version value update), can you check that again please ?", "Hi @aleSuglia , can you check that all the changes you wanted to do are in the dataset script ?", "Hey sorry but I'm in the middle of a conference deadline. I'll let you know asap!",...
1,597,351,886,000
1,603,263,269,000
1,603,263,269,000
CONTRIBUTOR
null
We updated some metadata information associated with the dataset. In addition, we've updated the `create_dummy_data.py` script to generate data samples for the dataset.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/503/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/503/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/503", "html_url": "https://github.com/huggingface/datasets/pull/503", "diff_url": "https://github.com/huggingface/datasets/pull/503.diff", "patch_url": "https://github.com/huggingface/datasets/pull/503.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/502
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/502/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/502/comments
https://api.github.com/repos/huggingface/datasets/issues/502/events
https://github.com/huggingface/datasets/pull/502
678,546,070
MDExOlB1bGxSZXF1ZXN0NDY3NDc1MDg0
502
Fix tokenizers caching
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "This should fix #501 and also the issue you sent me on slack @sgugger ." ]
1,597,334,017,000
1,597,844,239,000
1,597,844,238,000
MEMBER
null
I've found some cases where the caching didn't work properly for tokenizers: 1. if a tokenizer has a regex pattern, then the caching would be inconsistent across sessions 2. if a tokenizer has a cache attribute that changes after some calls, the the caching would not work after cache updates 3. if a tokenizer is u...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/502/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/502/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/502", "html_url": "https://github.com/huggingface/datasets/pull/502", "diff_url": "https://github.com/huggingface/datasets/pull/502.diff", "patch_url": "https://github.com/huggingface/datasets/pull/502.patch", "merged_at": 1597844237000 }
true
https://api.github.com/repos/huggingface/datasets/issues/501
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/501/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/501/comments
https://api.github.com/repos/huggingface/datasets/issues/501/events
https://github.com/huggingface/datasets/issues/501
677,952,893
MDU6SXNzdWU2Nzc5NTI4OTM=
501
Caching doesn't work for map (non-deterministic)
{ "login": "wulu473", "id": 8149933, "node_id": "MDQ6VXNlcjgxNDk5MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/8149933?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wulu473", "html_url": "https://github.com/wulu473", "followers_url": "https://api.github.com/users/wulu473/...
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "Thanks for reporting !\r\n\r\nTo store the cache file, we compute a hash of the function given in `.map`, using our own hashing function.\r\nThe hash doesn't seem to stay the same over sessions for the tokenizer.\r\nApparently this is because of the regex at `tokenizer.pat` is not well supported by our hashing fun...
1,597,263,607,000
1,598,286,900,000
1,598,286,875,000
NONE
null
The caching functionality doesn't work reliably when tokenizing a dataset. Here's a small example to reproduce it. ```python import nlp import transformers def main(): ds = nlp.load_dataset("reddit", split="train[:500]") tokenizer = transformers.AutoTokenizer.from_pretrained("gpt2") def conv...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/501/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/501/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/500
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/500/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/500/comments
https://api.github.com/repos/huggingface/datasets/issues/500/events
https://github.com/huggingface/datasets/pull/500
677,841,708
MDExOlB1bGxSZXF1ZXN0NDY2ODk0NTk0
500
Use hnsw in wiki_dpr
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,597,251,487,000
1,597,910,359,000
1,597,910,358,000
MEMBER
null
The HNSW faiss index is much faster that regular Flat index.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/500/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/500/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/500", "html_url": "https://github.com/huggingface/datasets/pull/500", "diff_url": "https://github.com/huggingface/datasets/pull/500.diff", "patch_url": "https://github.com/huggingface/datasets/pull/500.patch", "merged_at": 1597910358000 }
true
https://api.github.com/repos/huggingface/datasets/issues/499
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/499/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/499/comments
https://api.github.com/repos/huggingface/datasets/issues/499/events
https://github.com/huggingface/datasets/pull/499
677,709,938
MDExOlB1bGxSZXF1ZXN0NDY2Nzg1MjAy
499
Narrativeqa (with full text)
{ "login": "ghomasHudson", "id": 13795113, "node_id": "MDQ6VXNlcjEzNzk1MTEz", "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghomasHudson", "html_url": "https://github.com/ghomasHudson", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "I took a look at the dummy data creation for this dataset.\r\n\r\nMaybe it didn't work on your side might be because `master.zip` and `narrativeqa_full_text.zip` are supposed to be directories and not acutal zip files in the dummy data folder.\r\n\r\nI managed to make it work with this `dummy_data.zip` file:\r\nht...
1,597,240,183,000
1,607,512,862,000
1,607,512,862,000
CONTRIBUTOR
null
Following the uploading of the full text data in #309, I've added the full text to the narrativeqa dataset. Few notes: - Had some encoding issues using the default `open` so am using `open(encoding="latin-1"...` which seems to fix it. Looks fine. - Can't get the dummy data to work. Currently putting stuff at: ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/499/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/499/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/499", "html_url": "https://github.com/huggingface/datasets/pull/499", "diff_url": "https://github.com/huggingface/datasets/pull/499.diff", "patch_url": "https://github.com/huggingface/datasets/pull/499.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/498
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/498/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/498/comments
https://api.github.com/repos/huggingface/datasets/issues/498/events
https://github.com/huggingface/datasets/pull/498
677,597,479
MDExOlB1bGxSZXF1ZXN0NDY2Njg5NTcy
498
dont use beam fs to save info for local cache dir
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,597,230,000,000
1,597,411,041,000
1,597,411,040,000
MEMBER
null
If the cache dir is local, then we shouldn't use beam's filesystem to save the dataset info Fix #490
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/498/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/498/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/498", "html_url": "https://github.com/huggingface/datasets/pull/498", "diff_url": "https://github.com/huggingface/datasets/pull/498.diff", "patch_url": "https://github.com/huggingface/datasets/pull/498.patch", "merged_at": 1597411040000 }
true
https://api.github.com/repos/huggingface/datasets/issues/497
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/497/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/497/comments
https://api.github.com/repos/huggingface/datasets/issues/497/events
https://github.com/huggingface/datasets/pull/497
677,057,116
MDExOlB1bGxSZXF1ZXN0NDY2MjQ2NDQ3
497
skip header in PAWS-X
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,597,166,785,000
1,597,830,602,000
1,597,830,601,000
MEMBER
null
This should fix #485 I also updated the `dataset_infos.json` file that is used to verify the integrity of the generated splits (the number of examples was reduced by one). Note that there are new fields in `dataset_infos.json` introduced in the latest release 0.4.0 corresponding to post processing info. I remove...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/497/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/497/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/497", "html_url": "https://github.com/huggingface/datasets/pull/497", "diff_url": "https://github.com/huggingface/datasets/pull/497.diff", "patch_url": "https://github.com/huggingface/datasets/pull/497.patch", "merged_at": 1597830601000 }
true
https://api.github.com/repos/huggingface/datasets/issues/496
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/496/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/496/comments
https://api.github.com/repos/huggingface/datasets/issues/496/events
https://github.com/huggingface/datasets/pull/496
677,016,998
MDExOlB1bGxSZXF1ZXN0NDY2MjE1Mjg1
496
fix bad type in overflow check
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,597,163,098,000
1,597,411,775,000
1,597,411,774,000
MEMBER
null
When writing an arrow file and inferring the features, the overflow check could fail if the first example had a `null` field. This is because we were not using the inferred features to do this check, and we could end up with arrays that don't match because of a type mismatch (`null` vs `string` for example). This s...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/496/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/496/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/496", "html_url": "https://github.com/huggingface/datasets/pull/496", "diff_url": "https://github.com/huggingface/datasets/pull/496.diff", "patch_url": "https://github.com/huggingface/datasets/pull/496.patch", "merged_at": 1597411774000 }
true
https://api.github.com/repos/huggingface/datasets/issues/495
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/495/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/495/comments
https://api.github.com/repos/huggingface/datasets/issues/495/events
https://github.com/huggingface/datasets/pull/495
676,959,289
MDExOlB1bGxSZXF1ZXN0NDY2MTY5MTA3
495
stack vectors in pytorch and tensorflow
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,597,158,773,000
1,597,224,649,000
1,597,224,648,000
MEMBER
null
When the format of a dataset is set to pytorch or tensorflow, and if the dataset has vectors in it, they were not stacked together as tensors when calling `dataset[i:i + batch_size][column]` or `dataset[column]`. I added support for stacked tensors for both pytorch and tensorflow. For ragged tensors, they are stack...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/495/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/495/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/495", "html_url": "https://github.com/huggingface/datasets/pull/495", "diff_url": "https://github.com/huggingface/datasets/pull/495.diff", "patch_url": "https://github.com/huggingface/datasets/pull/495.patch", "merged_at": 1597224648000 }
true
https://api.github.com/repos/huggingface/datasets/issues/494
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/494/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/494/comments
https://api.github.com/repos/huggingface/datasets/issues/494/events
https://github.com/huggingface/datasets/pull/494
676,886,955
MDExOlB1bGxSZXF1ZXN0NDY2MTExOTQz
494
Fix numpy stacking
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "This PR also fixed a bug where numpy arrays were returned instead of pytorch tensors when getting with a clumn as a key." ]
1,597,153,230,000
1,597,157,810,000
1,597,153,792,000
MEMBER
null
When getting items using a column name as a key, numpy arrays were not stacked. I fixed that and added some tests. There is another issue that still needs to be fixed though: when getting items using a column name as a key, pytorch tensors are not stacked (it outputs a list of tensors). This PR should help with the...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/494/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/494/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/494", "html_url": "https://github.com/huggingface/datasets/pull/494", "diff_url": "https://github.com/huggingface/datasets/pull/494.diff", "patch_url": "https://github.com/huggingface/datasets/pull/494.patch", "merged_at": 1597153792000 }
true
https://api.github.com/repos/huggingface/datasets/issues/493
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/493/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/493/comments
https://api.github.com/repos/huggingface/datasets/issues/493/events
https://github.com/huggingface/datasets/pull/493
676,527,351
MDExOlB1bGxSZXF1ZXN0NDY1ODIxOTA0
493
Fix wmt zh-en url
{ "login": "sshleifer", "id": 6045025, "node_id": "MDQ6VXNlcjYwNDUwMjU=", "avatar_url": "https://avatars.githubusercontent.com/u/6045025?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sshleifer", "html_url": "https://github.com/sshleifer", "followers_url": "https://api.github.com/users/ss...
[]
closed
false
null
[]
null
[ "this doesn't work. I can decompress the file after download locally." ]
1,597,112,092,000
1,597,112,548,000
1,597,112,532,000
CONTRIBUTOR
null
I verified that ``` wget https://stuncorpusprod.blob.core.windows.net/corpusfiles/UNv1.0.en-zh.tar.gz.00 ``` runs in 2 minutes.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/493/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/493/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/493", "html_url": "https://github.com/huggingface/datasets/pull/493", "diff_url": "https://github.com/huggingface/datasets/pull/493.diff", "patch_url": "https://github.com/huggingface/datasets/pull/493.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/492
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/492/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/492/comments
https://api.github.com/repos/huggingface/datasets/issues/492/events
https://github.com/huggingface/datasets/issues/492
676,495,064
MDU6SXNzdWU2NzY0OTUwNjQ=
492
nlp.Features does not distinguish between nullable and non-nullable types in PyArrow schema
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "In 0.4.0, the assertion in `concatenate_datasets ` is on the features, and not the schema.\r\nCould you try to update `nlp` ?\r\n\r\nAlso, since 0.4.0, you can use `dset_wikipedia.cast_(dset_books.features)` to avoid the schema cast hack.", "Or maybe the assertion comes from elsewhere ?", "I'm using the master...
1,597,105,666,000
1,598,458,639,000
1,598,458,639,000
CONTRIBUTOR
null
Here's the code I'm trying to run: ```python dset_wikipedia = nlp.load_dataset("wikipedia", "20200501.en", split="train", cache_dir=args.cache_dir) dset_wikipedia.drop(columns=["title"]) dset_wikipedia.features.pop("title") dset_books = nlp.load_dataset("bookcorpus", split="train", cache_dir=args.cache_dir) dse...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/492/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/492/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/491
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/491/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/491/comments
https://api.github.com/repos/huggingface/datasets/issues/491/events
https://github.com/huggingface/datasets/issues/491
676,486,275
MDU6SXNzdWU2NzY0ODYyNzU=
491
No 0.4.0 release on GitHub
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "I did the release on github, and updated the doc :)\r\nSorry for the delay", "Thanks!" ]
1,597,103,997,000
1,597,164,607,000
1,597,164,607,000
CONTRIBUTOR
null
0.4.0 was released on PyPi, but not on GitHub. This means [the documentation](https://huggingface.co/nlp/) is still displaying from 0.3.0, and that there's no tag to easily clone the 0.4.0 version of the repo.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/491/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/491/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/490
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/490/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/490/comments
https://api.github.com/repos/huggingface/datasets/issues/490/events
https://github.com/huggingface/datasets/issues/490
676,482,242
MDU6SXNzdWU2NzY0ODIyNDI=
490
Loading preprocessed Wikipedia dataset requires apache_beam
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[]
1,597,103,210,000
1,597,411,040,000
1,597,411,040,000
CONTRIBUTOR
null
Running `nlp.load_dataset("wikipedia", "20200501.en", split="train", dir="/tmp/wikipedia")` gives an error if apache_beam is not installed, stemming from https://github.com/huggingface/nlp/blob/38eb2413de54ee804b0be81781bd65ac4a748ced/src/nlp/builder.py#L981-L988 This succeeded without the dependency in ve...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/490/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/490/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/489
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/489/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/489/comments
https://api.github.com/repos/huggingface/datasets/issues/489/events
https://github.com/huggingface/datasets/issues/489
676,456,257
MDU6SXNzdWU2NzY0NTYyNTc=
489
ug
{ "login": "timothyjlaurent", "id": 2000204, "node_id": "MDQ6VXNlcjIwMDAyMDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/timothyjlaurent", "html_url": "https://github.com/timothyjlaurent", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "whoops", "please delete this" ]
1,597,098,783,000
1,597,100,114,000
1,597,098,820,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/489/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/489/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/488
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/488/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/488/comments
https://api.github.com/repos/huggingface/datasets/issues/488/events
https://github.com/huggingface/datasets/issues/488
676,299,993
MDU6SXNzdWU2NzYyOTk5OTM=
488
issues with downloading datasets for wmt16 and wmt19
{ "login": "stas00", "id": 10676103, "node_id": "MDQ6VXNlcjEwNjc2MTAz", "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stas00", "html_url": "https://github.com/stas00", "followers_url": "https://api.github.com/users/stas00/fo...
[]
open
false
null
[]
null
[ "I found `UNv1.0.en-ru.tar.gz` here: https://conferences.unite.un.org/uncorpus/en/downloadoverview, so it can be reconstructed with:\r\n```\r\nwget -c https://stuncorpusprod.blob.core.windows.net/corpusfiles/UNv1.0.en-ru.tar.gz.00\r\nwget -c https://stuncorpusprod.blob.core.windows.net/corpusfiles/UNv1.0.en-ru.tar....
1,597,080,771,000
1,597,122,454,000
null
CONTRIBUTOR
null
I have encountered multiple issues while trying to: ``` import nlp dataset = nlp.load_dataset('wmt16', 'ru-en') metric = nlp.load_metric('wmt16') ``` 1. I had to do `pip install -e ".[dev]" ` on master, currently released nlp didn't work (sorry, didn't save the error) - I went back to the released version and no...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/488/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/488/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/487
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/487/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/487/comments
https://api.github.com/repos/huggingface/datasets/issues/487/events
https://github.com/huggingface/datasets/pull/487
676,143,029
MDExOlB1bGxSZXF1ZXN0NDY1NTA1NjQy
487
Fix elasticsearch result ids returning as strings
{ "login": "sai-prasanna", "id": 3595526, "node_id": "MDQ6VXNlcjM1OTU1MjY=", "avatar_url": "https://avatars.githubusercontent.com/u/3595526?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sai-prasanna", "html_url": "https://github.com/sai-prasanna", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "It looks like you need to rebase from master to fix the CI. Could you do that please ?" ]
1,597,066,631,000
1,598,870,566,000
1,598,870,566,000
CONTRIBUTOR
null
I am using the latest elasticsearch binary and master of nlp. For me elasticsearch searches failed because the resultant "id_" returned for searches are strings, but our library assumes them to be integers.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/487/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/487/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/487", "html_url": "https://github.com/huggingface/datasets/pull/487", "diff_url": "https://github.com/huggingface/datasets/pull/487.diff", "patch_url": "https://github.com/huggingface/datasets/pull/487.patch", "merged_at": 1598870566000 }
true
https://api.github.com/repos/huggingface/datasets/issues/486
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/486/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/486/comments
https://api.github.com/repos/huggingface/datasets/issues/486/events
https://github.com/huggingface/datasets/issues/486
675,649,034
MDU6SXNzdWU2NzU2NDkwMzQ=
486
Bookcorpus data contains pretokenized text
{ "login": "orsharir", "id": 99543, "node_id": "MDQ6VXNlcjk5NTQz", "avatar_url": "https://avatars.githubusercontent.com/u/99543?v=4", "gravatar_id": "", "url": "https://api.github.com/users/orsharir", "html_url": "https://github.com/orsharir", "followers_url": "https://api.github.com/users/orsharir/foll...
[]
open
false
null
[]
null
[ "Yes indeed it looks like some `'` and spaces are missing (for example in `dont` or `didnt`).\r\nDo you know if there exist some copies without this issue ?\r\nHow would you fix this issue on the current data exactly ? I can see that the data is raw text (not tokenized) so I'm not sure I understand how you would do...
1,596,956,004,000
1,601,467,264,000
null
CONTRIBUTOR
null
It seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece tokenizer works. For example, "didn't" becomes "did" + "n't", and double quotes are changed to `` and '' for start and end q...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/486/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/486/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/485
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/485/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/485/comments
https://api.github.com/repos/huggingface/datasets/issues/485/events
https://github.com/huggingface/datasets/issues/485
675,595,393
MDU6SXNzdWU2NzU1OTUzOTM=
485
PAWS dataset first item is header
{ "login": "jxmorris12", "id": 13238952, "node_id": "MDQ6VXNlcjEzMjM4OTUy", "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jxmorris12", "html_url": "https://github.com/jxmorris12", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[]
1,596,924,325,000
1,597,830,601,000
1,597,830,601,000
CONTRIBUTOR
null
``` import nlp dataset = nlp.load_dataset('xtreme', 'PAWS-X.en') dataset['test'][0] ``` prints the following ``` {'label': 'label', 'sentence1': 'sentence1', 'sentence2': 'sentence2'} ``` dataset['test'][0] should probably be the first item in the dataset, not just a dictionary mapping the column names t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/485/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/485/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/484
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/484/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/484/comments
https://api.github.com/repos/huggingface/datasets/issues/484/events
https://github.com/huggingface/datasets/pull/484
675,088,983
MDExOlB1bGxSZXF1ZXN0NDY0NjY1NTU4
484
update mirror for RT dataset
{ "login": "jxmorris12", "id": 13238952, "node_id": "MDQ6VXNlcjEzMjM4OTUy", "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jxmorris12", "html_url": "https://github.com/jxmorris12", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "Thanks for adding this mirror link :)\r\n\r\nCould you run the following command to update the json file `dataset_infos.json` used to verify the integrity of the downloaded file ?\r\n\r\n```\r\nnlp-cli test ./datasets/rotten_tomatoes --save_infos --ignore_verifications\r\n```", "done! @lhoestq ", "the build_do...
1,596,813,945,000
1,598,276,017,000
1,598,276,017,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/484/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/484/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/484", "html_url": "https://github.com/huggingface/datasets/pull/484", "diff_url": "https://github.com/huggingface/datasets/pull/484.diff", "patch_url": "https://github.com/huggingface/datasets/pull/484.patch", "merged_at": 1598276017000 }
true
https://api.github.com/repos/huggingface/datasets/issues/483
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/483/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/483/comments
https://api.github.com/repos/huggingface/datasets/issues/483/events
https://github.com/huggingface/datasets/issues/483
675,080,694
MDU6SXNzdWU2NzUwODA2OTQ=
483
rotten tomatoes movie review dataset taken down
{ "login": "jxmorris12", "id": 13238952, "node_id": "MDQ6VXNlcjEzMjM4OTUy", "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jxmorris12", "html_url": "https://github.com/jxmorris12", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "found a mirror: https://storage.googleapis.com/seldon-datasets/sentence_polarity_v1/rt-polaritydata.tar.gz", "fixed in #484 ", "Closing this one. Thanks again @jxmorris12 for taking care of this :)" ]
1,596,813,121,000
1,599,557,794,000
1,599,557,793,000
CONTRIBUTOR
null
In an interesting twist of events, the individual who created the movie review seems to have left Cornell, and their webpage has been removed, along with the movie review dataset (http://www.cs.cornell.edu/people/pabo/movie-review-data/rt-polaritydata.tar.gz). It's not downloadable anymore.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/483/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/483/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/482
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/482/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/482/comments
https://api.github.com/repos/huggingface/datasets/issues/482/events
https://github.com/huggingface/datasets/issues/482
674,851,147
MDU6SXNzdWU2NzQ4NTExNDc=
482
Bugs : dataset.map() is frozen on ELI5
{ "login": "ratthachat", "id": 56621342, "node_id": "MDQ6VXNlcjU2NjIxMzQy", "avatar_url": "https://avatars.githubusercontent.com/u/56621342?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ratthachat", "html_url": "https://github.com/ratthachat", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "This comes from an overflow in pyarrow's array.\r\nIt is stuck inside the loop that reduces the batch size to avoid the overflow.\r\nI'll take a look", "I created a PR to fix the issue.\r\nIt was due to an overflow check that handled badly an empty list.\r\n\r\nYou can try the changes by using \r\n```\r\n!pip in...
1,596,788,615,000
1,597,241,626,000
1,597,190,115,000
NONE
null
Hi Huggingface Team! Thank you guys once again for this amazing repo. I have tried to prepare ELI5 to train with T5, based on [this wonderful notebook of Suraj Patil](https://github.com/patil-suraj/exploring-T5/blob/master/T5_on_TPU.ipynb) However, when I run `dataset.map()` on ELI5 to prepare `input_text, ta...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/482/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/482/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/481
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/481/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/481/comments
https://api.github.com/repos/huggingface/datasets/issues/481/events
https://github.com/huggingface/datasets/pull/481
674,567,389
MDExOlB1bGxSZXF1ZXN0NDY0MjM2MTA1
481
Apply utf-8 encoding to all datasets
{ "login": "lewtun", "id": 26859204, "node_id": "MDQ6VXNlcjI2ODU5MjA0", "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lewtun", "html_url": "https://github.com/lewtun", "followers_url": "https://api.github.com/users/lewtun/fo...
[]
closed
false
null
[]
null
[ "Not sure why the AWS test is failing - perhaps I made too many concurrent CI builds 😢. Can someone please rerun the CI to check the error is not on my end?", "I pushed an improved docstring and the unit tests now pass, which suggests the previous failure on AWS was simply a timeout error. \r\n\r\nFor some reaso...
1,596,744,129,000
1,597,911,368,000
1,597,911,368,000
MEMBER
null
## Description This PR applies utf-8 encoding for all instances of `with open(...) as f` to all Python files in `datasets/`. As suggested by @thomwolf in #468 , we use regular expressions and the following function ```python def apply_encoding_on_file_open(filepath: str): """Apply UTF-8 encoding for all insta...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/481/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/481/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/481", "html_url": "https://github.com/huggingface/datasets/pull/481", "diff_url": "https://github.com/huggingface/datasets/pull/481.diff", "patch_url": "https://github.com/huggingface/datasets/pull/481.patch", "merged_at": 1597911368000 }
true
https://api.github.com/repos/huggingface/datasets/issues/480
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/480/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/480/comments
https://api.github.com/repos/huggingface/datasets/issues/480/events
https://github.com/huggingface/datasets/pull/480
674,245,959
MDExOlB1bGxSZXF1ZXN0NDYzOTcwNjQ2
480
Column indexing hotfix
{ "login": "TevenLeScao", "id": 26709476, "node_id": "MDQ6VXNlcjI2NzA5NDc2", "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TevenLeScao", "html_url": "https://github.com/TevenLeScao", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[ "Looks good to me as well but we'll want to add a test indeed.\r\nYou can add one if you have time @TevenLeScao.\r\nOtherwise, we'll do it when we are back with Quentin. ", "I fixed it in #494 " ]
1,596,713,825,000
1,597,221,370,000
1,597,221,370,000
MEMBER
null
As observed for example in #469 , currently `__getitem__` does not convert the data to the dataset format when indexing by column. This is a hotfix that imitates functional 0.3.0. code. In the future it'd probably be nice to have a test there.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/480/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/480/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/480", "html_url": "https://github.com/huggingface/datasets/pull/480", "diff_url": "https://github.com/huggingface/datasets/pull/480.diff", "patch_url": "https://github.com/huggingface/datasets/pull/480.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/479
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/479/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/479/comments
https://api.github.com/repos/huggingface/datasets/issues/479/events
https://github.com/huggingface/datasets/pull/479
673,905,407
MDExOlB1bGxSZXF1ZXN0NDYzNjkxMjA0
479
add METEOR metric
{ "login": "vegarab", "id": 24683907, "node_id": "MDQ6VXNlcjI0NjgzOTA3", "avatar_url": "https://avatars.githubusercontent.com/u/24683907?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vegarab", "html_url": "https://github.com/vegarab", "followers_url": "https://api.github.com/users/vegara...
[]
closed
false
null
[]
null
[ "Really nice !\r\nThanks for adding this one.\r\n\r\nI noticed that there are some '-' that are left in the description in the middle of some workds. It migh come from copy-pasting the pdf paper. ex: `im-provement`. Could you fix that please ?", "@lhoestq \r\nLinebreaks have been removed! Note that there are stil...
1,596,669,180,000
1,597,844,349,000
1,597,844,349,000
CONTRIBUTOR
null
Added the METEOR metric. Can be used like this: ```python import nlp meteor = nlp.load_metric('metrics/meteor') meteor.compute(["some string", "some string"], ["some string", "some similar string"]) # {'meteor': 0.6411637931034483} meteor.add("some string", "some string") meteor.add('some string", "some simila...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/479/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/479/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/479", "html_url": "https://github.com/huggingface/datasets/pull/479", "diff_url": "https://github.com/huggingface/datasets/pull/479.diff", "patch_url": "https://github.com/huggingface/datasets/pull/479.patch", "merged_at": 1597844349000 }
true
https://api.github.com/repos/huggingface/datasets/issues/478
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/478/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/478/comments
https://api.github.com/repos/huggingface/datasets/issues/478/events
https://github.com/huggingface/datasets/issues/478
673,178,317
MDU6SXNzdWU2NzMxNzgzMTc=
478
Export TFRecord to GCP bucket
{ "login": "astariul", "id": 43774355, "node_id": "MDQ6VXNlcjQzNzc0MzU1", "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "gravatar_id": "", "url": "https://api.github.com/users/astariul", "html_url": "https://github.com/astariul", "followers_url": "https://api.github.com/users/ast...
[]
closed
false
null
[]
null
[ "Nevermind, I restarted my python session and it worked fine...\r\n\r\n---\r\n\r\nI had an authentification error, and I authenticated from another terminal. After that, no more error but it was not working. Restarting the sessions makes it work :)" ]
1,596,589,712,000
1,596,590,497,000
1,596,590,496,000
NONE
null
Previously, I was writing TFRecords manually to GCP bucket with : `with tf.io.TFRecordWriter('gs://my_bucket/x.tfrecord')` Since `0.4.0` is out with the `export()` function, I tried it. But it seems TFRecords cannot be directly written to GCP bucket. `dataset.export('local.tfrecord')` works fine, but `dataset....
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/478/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/478/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/477
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/477/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/477/comments
https://api.github.com/repos/huggingface/datasets/issues/477/events
https://github.com/huggingface/datasets/issues/477
673,142,143
MDU6SXNzdWU2NzMxNDIxNDM=
477
Overview.ipynb throws exceptions with nlp 0.4.0
{ "login": "mandy-li", "id": 23109219, "node_id": "MDQ6VXNlcjIzMTA5MjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23109219?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mandy-li", "html_url": "https://github.com/mandy-li", "followers_url": "https://api.github.com/users/man...
[]
closed
false
null
[]
null
[ "Thanks for reporting this issue\r\n\r\nThere was a bug where numpy arrays would get returned instead of tensorflow tensors.\r\nThis is fixed on master.\r\n\r\nI tried to re-run the colab and encountered this error instead:\r\n\r\n```\r\nAttributeError: 'tensorflow.python.framework.ops.EagerTensor' object has no at...
1,596,583,095,000
1,627,970,535,000
1,627,970,535,000
NONE
null
with nlp 0.4.0, the TensorFlow example in Overview.ipynb throws the following exceptions: --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) <ipython-input-5-48907f2ad433> in <module> ----> 1 features = {x: trai...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/477/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/477/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/476
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/476/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/476/comments
https://api.github.com/repos/huggingface/datasets/issues/476/events
https://github.com/huggingface/datasets/pull/476
672,991,854
MDExOlB1bGxSZXF1ZXN0NDYyOTMyMTgx
476
CheckList
{ "login": "marcotcr", "id": 698010, "node_id": "MDQ6VXNlcjY5ODAxMA==", "avatar_url": "https://avatars.githubusercontent.com/u/698010?v=4", "gravatar_id": "", "url": "https://api.github.com/users/marcotcr", "html_url": "https://github.com/marcotcr", "followers_url": "https://api.github.com/users/marcotc...
[]
open
false
null
[]
null
[ "> Also, a little out of my depth there, but would there be a way to have the default pip install checklist command not require mysql and mariadb to be installed? Feels like that might be a source of confusion for users.\r\n\r\nI removed the pattern dependency, mysql is not a requirement anymore. I'm not sure where...
1,596,565,925,000
1,599,161,648,000
null
NONE
null
Sorry for the large pull request. - Added checklists as datasets. I can't run `test_load_real_dataset` (see #474), but I can load the datasets successfully as shown in the example notebook - Added a checklist wrapper
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/476/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/476/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/476", "html_url": "https://github.com/huggingface/datasets/pull/476", "diff_url": "https://github.com/huggingface/datasets/pull/476.diff", "patch_url": "https://github.com/huggingface/datasets/pull/476.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/475
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/475/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/475/comments
https://api.github.com/repos/huggingface/datasets/issues/475/events
https://github.com/huggingface/datasets/pull/475
672,884,595
MDExOlB1bGxSZXF1ZXN0NDYyODQzMzQz
475
misc. bugs and quality of life
{ "login": "joeddav", "id": 9353833, "node_id": "MDQ6VXNlcjkzNTM4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joeddav", "html_url": "https://github.com/joeddav", "followers_url": "https://api.github.com/users/joeddav/...
[]
closed
false
null
[]
null
[ "Cool thanks, I made those changes. LMK if you think it's ready for merge.", "Ok to merge for me" ]
1,596,555,149,000
1,597,698,848,000
1,597,698,847,000
CONTRIBUTOR
null
A few misc. bugs and QOL improvements that I've come across in using the library. Let me know if you don't like any of them and I can adjust/remove them. 1. Printing datasets without a description field throws an error when formatting the `single_line_description`. This fixes that, and also adds some formatting to t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/475/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/475/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/475", "html_url": "https://github.com/huggingface/datasets/pull/475", "diff_url": "https://github.com/huggingface/datasets/pull/475.diff", "patch_url": "https://github.com/huggingface/datasets/pull/475.patch", "merged_at": 1597698847000 }
true
https://api.github.com/repos/huggingface/datasets/issues/474
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/474/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/474/comments
https://api.github.com/repos/huggingface/datasets/issues/474/events
https://github.com/huggingface/datasets/issues/474
672,407,330
MDU6SXNzdWU2NzI0MDczMzA=
474
test_load_real_dataset when config has BUILDER_CONFIGS that matter
{ "login": "marcotcr", "id": 698010, "node_id": "MDQ6VXNlcjY5ODAxMA==", "avatar_url": "https://avatars.githubusercontent.com/u/698010?v=4", "gravatar_id": "", "url": "https://api.github.com/users/marcotcr", "html_url": "https://github.com/marcotcr", "followers_url": "https://api.github.com/users/marcotc...
[]
closed
false
null
[]
null
[ "The `data_dir` parameter has been removed. Now the error is `ValueError: Config name is missing`\r\n\r\nAs mentioned in #470 I think we can have one test with the first config of BUILDER_CONFIGS, and another test that runs all of the configs in BUILDER_CONFIGS", "This was fixed in #527 \r\n\r\nClosing this one, ...
1,596,498,396,000
1,599,490,393,000
1,599,490,393,000
NONE
null
It a dataset has custom `BUILDER_CONFIGS` with non-keyword arguments (or keyword arguments with non default values), the config is not loaded during the test and causes an error. I think the problem is that `test_load_real_dataset` calls `load_dataset` with `data_dir=temp_data_dir` ([here](https://github.com/huggingfa...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/474/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/474/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/473
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/473/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/473/comments
https://api.github.com/repos/huggingface/datasets/issues/473/events
https://github.com/huggingface/datasets/pull/473
672,007,247
MDExOlB1bGxSZXF1ZXN0NDYyMTIwNzU4
473
add DoQA dataset (ACL 2020)
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,596,454,012,000
1,599,758,351,000
1,599,133,455,000
CONTRIBUTOR
null
add DoQA dataset (ACL 2020) http://ixa.eus/node/12931
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/473/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/473/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/473", "html_url": "https://github.com/huggingface/datasets/pull/473", "diff_url": "https://github.com/huggingface/datasets/pull/473.diff", "patch_url": "https://github.com/huggingface/datasets/pull/473.patch", "merged_at": 1599133454000 }
true
https://api.github.com/repos/huggingface/datasets/issues/472
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/472/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/472/comments
https://api.github.com/repos/huggingface/datasets/issues/472/events
https://github.com/huggingface/datasets/pull/472
672,000,745
MDExOlB1bGxSZXF1ZXN0NDYyMTE1MjA4
472
add crd3 dataset
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "This PR was already approved by @lhoestq in #456 . This one just make style to remove some typos" ]
1,596,453,302,000
1,596,453,730,000
1,596,453,729,000
CONTRIBUTOR
null
opening new PR for CRD3 dataset (ACL2020) to fix the circle CI problems
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/472/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/472/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/472", "html_url": "https://github.com/huggingface/datasets/pull/472", "diff_url": "https://github.com/huggingface/datasets/pull/472.diff", "patch_url": "https://github.com/huggingface/datasets/pull/472.patch", "merged_at": 1596453729000 }
true
https://api.github.com/repos/huggingface/datasets/issues/471
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/471/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/471/comments
https://api.github.com/repos/huggingface/datasets/issues/471/events
https://github.com/huggingface/datasets/pull/471
671,996,423
MDExOlB1bGxSZXF1ZXN0NDYyMTExNTU1
471
add reuters21578 dataset
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,596,452,834,000
1,599,127,683,000
1,599,127,130,000
CONTRIBUTOR
null
new PR to add the reuters21578 dataset and fix the circle CI problems.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/471/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/471/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/471", "html_url": "https://github.com/huggingface/datasets/pull/471", "diff_url": "https://github.com/huggingface/datasets/pull/471.diff", "patch_url": "https://github.com/huggingface/datasets/pull/471.patch", "merged_at": 1599127130000 }
true
https://api.github.com/repos/huggingface/datasets/issues/470
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/470/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/470/comments
https://api.github.com/repos/huggingface/datasets/issues/470/events
https://github.com/huggingface/datasets/pull/470
671,952,276
MDExOlB1bGxSZXF1ZXN0NDYyMDc0MzQ0
470
Adding IWSLT 2017 dataset.
{ "login": "Narsil", "id": 204321, "node_id": "MDQ6VXNlcjIwNDMyMQ==", "avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Narsil", "html_url": "https://github.com/Narsil", "followers_url": "https://api.github.com/users/Narsil/follow...
[]
closed
false
null
[]
null
[ "Ok I tried to add the dummy dataset (I actually modified the dummy_data command to generate them for me because it was too painful to do that manually).\r\n\r\nThe dummy_data test seems to work:\r\n```bash\r\nRUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_configs_iwslt2017\...
1,596,448,359,000
1,599,482,010,000
1,599,482,010,000
CONTRIBUTOR
null
Created a [IWSLT 2017](https://sites.google.com/site/iwsltevaluation2017/TED-tasks) dataset script for the *multilingual data*. ``` Bilingual data: {Arabic, German, French, Japanese, Korean, Chinese} <-> English Multilingual data: German, English, Italian, Dutch, Romanian. (Any pair) ``` I'm unsure how to h...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/470/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/470/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/470", "html_url": "https://github.com/huggingface/datasets/pull/470", "diff_url": "https://github.com/huggingface/datasets/pull/470.diff", "patch_url": "https://github.com/huggingface/datasets/pull/470.patch", "merged_at": 1599482010000 }
true
https://api.github.com/repos/huggingface/datasets/issues/469
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/469/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/469/comments
https://api.github.com/repos/huggingface/datasets/issues/469/events
https://github.com/huggingface/datasets/issues/469
671,876,963
MDU6SXNzdWU2NzE4NzY5NjM=
469
invalid data type 'str' at _convert_outputs in arrow_dataset.py
{ "login": "Murgates", "id": 30617486, "node_id": "MDQ6VXNlcjMwNjE3NDg2", "avatar_url": "https://avatars.githubusercontent.com/u/30617486?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Murgates", "html_url": "https://github.com/Murgates", "followers_url": "https://api.github.com/users/Mur...
[]
open
false
null
[]
null
[ "Hi ! Did you try to set the output format to pytorch ? (or tensorflow if you're using tensorflow)\r\nIt can be done with `dataset.set_format(\"torch\", columns=columns)` (or \"tensorflow\").\r\n\r\nNote that for pytorch, string columns can't be converted to `torch.Tensor`, so you have to specify in `columns=` the...
1,596,440,909,000
1,603,357,466,000
null
NONE
null
I trying to build multi label text classifier model using Transformers lib. I'm using Transformers NLP to load the data set, while calling trainer.train() method. It throws the following error File "C:\***\arrow_dataset.py", line 343, in _convert_outputs v = command(v) TypeError: new(): invalid data type ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/469/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/469/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/468
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/468/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/468/comments
https://api.github.com/repos/huggingface/datasets/issues/468/events
https://github.com/huggingface/datasets/issues/468
671,622,441
MDU6SXNzdWU2NzE2MjI0NDE=
468
UnicodeDecodeError while loading PAN-X task of XTREME dataset
{ "login": "lewtun", "id": 26859204, "node_id": "MDQ6VXNlcjI2ODU5MjA0", "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lewtun", "html_url": "https://github.com/lewtun", "followers_url": "https://api.github.com/users/lewtun/fo...
[]
closed
false
null
[]
null
[ "Indeed. Solution 1 is the simplest.\r\n\r\nThis is actually a recurring problem.\r\nI think we should scan all the datasets with regexpr to fix the use of `open()` without encodings.\r\nAnd probably add a test in the CI to forbid using this in the future.", "I'm happy to tackle the broader problem - will open a ...
1,596,377,110,000
1,597,911,368,000
1,597,911,368,000
MEMBER
null
Hi 🤗 team! ## Description of the problem I'm running into a `UnicodeDecodeError` while trying to load the PAN-X subset the XTREME dataset: ``` --------------------------------------------------------------------------- UnicodeDecodeError Traceback (most recent call last) <ipython-inp...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/468/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/468/timeline
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/467
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/467/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/467/comments
https://api.github.com/repos/huggingface/datasets/issues/467/events
https://github.com/huggingface/datasets/pull/467
671,580,010
MDExOlB1bGxSZXF1ZXN0NDYxNzgwMzUy
467
DOCS: Fix typo
{ "login": "Bharat123rox", "id": 13381361, "node_id": "MDQ6VXNlcjEzMzgxMzYx", "avatar_url": "https://avatars.githubusercontent.com/u/13381361?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Bharat123rox", "html_url": "https://github.com/Bharat123rox", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Thanks!" ]
1,596,358,777,000
1,596,376,347,000
1,596,359,934,000
CONTRIBUTOR
null
Fix typo from dictionnary -> dictionary
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/467/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/467/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/467", "html_url": "https://github.com/huggingface/datasets/pull/467", "diff_url": "https://github.com/huggingface/datasets/pull/467.diff", "patch_url": "https://github.com/huggingface/datasets/pull/467.patch", "merged_at": 1596359934000 }
true
https://api.github.com/repos/huggingface/datasets/issues/466
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/466/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/466/comments
https://api.github.com/repos/huggingface/datasets/issues/466/events
https://github.com/huggingface/datasets/pull/466
670,766,891
MDExOlB1bGxSZXF1ZXN0NDYxMDEzOTM0
466
[METRICS] Various improvements on metrics
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "The cast function is now called inside `features.encode_example`.\r\nI also added `encode_batch` that was missing.\r\n\r\nMoreover I used the cast function in `Dataset.map` to support torch/tensorflow tensors or numpy arrays inputs.\r\n\r\nThere are tests for tensors inputs in metrics and in .map", "I think we c...
1,596,279,825,000
1,597,677,300,000
1,597,677,299,000
MEMBER
null
- Disallow the use of positional arguments to avoid `predictions` vs `references` mistakes - Allow to directly feed numpy/pytorch/tensorflow/pandas objects in metrics
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/466/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/466/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/466", "html_url": "https://github.com/huggingface/datasets/pull/466", "diff_url": "https://github.com/huggingface/datasets/pull/466.diff", "patch_url": "https://github.com/huggingface/datasets/pull/466.patch", "merged_at": 1597677299000 }
true
https://api.github.com/repos/huggingface/datasets/issues/465
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/465/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/465/comments
https://api.github.com/repos/huggingface/datasets/issues/465/events
https://github.com/huggingface/datasets/pull/465
669,889,779
MDExOlB1bGxSZXF1ZXN0NDYwMjEwODYw
465
Keep features after transform
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "One note on features inference:\r\n\r\nif an arrow type is `struct of items` where each item is a `list`, then we return a `dict` in which each item is a `Sequence`.\r\nIt means that we don't use the Sequence <-> dict swap when we infer features.\r\n\r\nIt's fine because the swap is generally used in dataset scrip...
1,596,206,601,000
1,596,220,053,000
1,596,220,052,000
MEMBER
null
When applying a transform like `map`, some features were lost (and inferred features were used). It was the case for ClassLabel, Translation, etc. To fix that, I did some modifications in the `ArrowWriter`: - added the `update_features` parameter. When it's `True`, then the features specified by the user (if any...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/465/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/465/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/465", "html_url": "https://github.com/huggingface/datasets/pull/465", "diff_url": "https://github.com/huggingface/datasets/pull/465.diff", "patch_url": "https://github.com/huggingface/datasets/pull/465.patch", "merged_at": 1596220052000 }
true
https://api.github.com/repos/huggingface/datasets/issues/464
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/464/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/464/comments
https://api.github.com/repos/huggingface/datasets/issues/464/events
https://github.com/huggingface/datasets/pull/464
669,767,381
MDExOlB1bGxSZXF1ZXN0NDYwMTAxNDYz
464
Add rename, remove and cast in-place operations
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,596,198,621,000
1,596,210,602,000
1,596,210,600,000
MEMBER
null
Add a bunch of in-place operation leveraging the Arrow back-end to rename and remove columns and cast to new features without using the more expensive `map` method. These methods are added to `Dataset` as well as `DatasetDict`. Added tests for these new methods and add the methods to the doc. Naming follows th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/464/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/464/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/464", "html_url": "https://github.com/huggingface/datasets/pull/464", "diff_url": "https://github.com/huggingface/datasets/pull/464.diff", "patch_url": "https://github.com/huggingface/datasets/pull/464.patch", "merged_at": 1596210600000 }
true
https://api.github.com/repos/huggingface/datasets/issues/463
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/463/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/463/comments
https://api.github.com/repos/huggingface/datasets/issues/463/events
https://github.com/huggingface/datasets/pull/463
669,735,455
MDExOlB1bGxSZXF1ZXN0NDYwMDcyNjQ1
463
Add dataset/mlsum
{ "login": "RachelKer", "id": 36986299, "node_id": "MDQ6VXNlcjM2OTg2Mjk5", "avatar_url": "https://avatars.githubusercontent.com/u/36986299?v=4", "gravatar_id": "", "url": "https://api.github.com/users/RachelKer", "html_url": "https://github.com/RachelKer", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "I think the problem is related to `wiki_dpr` dataset which is making the circle CI failed as you can see:\r\n```\r\nFAILED tests/test_dataset_common.py::AWSDatasetTest::test_load_dataset_wiki_dpr\r\nFAILED tests/test_hf_gcp.py::TestDatasetOnHfGcp::test_script_synced_with_s3_wiki_dpr/dummy_psgs_w100_no_embeddings\r...
1,596,196,252,000
1,598,280,882,000
1,598,280,882,000
CONTRIBUTOR
null
New pull request that should correct the previous errors. The load_real_data stills fails because it is looking for a default dataset URL that does not exists, this does not happen when loading the dataset with load_dataset
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/463/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/463/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/463", "html_url": "https://github.com/huggingface/datasets/pull/463", "diff_url": "https://github.com/huggingface/datasets/pull/463.diff", "patch_url": "https://github.com/huggingface/datasets/pull/463.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/462
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/462/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/462/comments
https://api.github.com/repos/huggingface/datasets/issues/462/events
https://github.com/huggingface/datasets/pull/462
669,715,547
MDExOlB1bGxSZXF1ZXN0NDYwMDU0NDgz
462
add DoQA (ACL 2020) dataset
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,596,194,756,000
1,596,454,107,000
1,596,454,107,000
CONTRIBUTOR
null
adds DoQA (ACL 2020) dataset
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/462/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/462/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/462", "html_url": "https://github.com/huggingface/datasets/pull/462", "diff_url": "https://github.com/huggingface/datasets/pull/462.diff", "patch_url": "https://github.com/huggingface/datasets/pull/462.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/461
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/461/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/461/comments
https://api.github.com/repos/huggingface/datasets/issues/461/events
https://github.com/huggingface/datasets/pull/461
669,703,508
MDExOlB1bGxSZXF1ZXN0NDYwMDQzNDY5
461
Doqa
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,596,193,872,000
1,596,193,995,000
1,596,193,995,000
CONTRIBUTOR
null
add DoQA (ACL 2020) dataset
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/461/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/461/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/461", "html_url": "https://github.com/huggingface/datasets/pull/461", "diff_url": "https://github.com/huggingface/datasets/pull/461.diff", "patch_url": "https://github.com/huggingface/datasets/pull/461.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/460
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/460/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/460/comments
https://api.github.com/repos/huggingface/datasets/issues/460/events
https://github.com/huggingface/datasets/pull/460
669,585,256
MDExOlB1bGxSZXF1ZXN0NDU5OTM2OTU2
460
Fix KeyboardInterrupt in map and bad indices in select
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "Thanks @TevenLeScao for finding this issue", "Thanks @lhoestq for catching this ❤️" ]
1,596,185,835,000
1,596,195,139,000
1,596,195,138,000
MEMBER
null
If you interrupted a map function while it was writing, the cached file was not discarded. Therefore the next time you called map, it was loading an incomplete arrow file. We had the same issue with select if there was a bad indice at one point. To fix that I used temporary files that are renamed once everything...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/460/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/460/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/460", "html_url": "https://github.com/huggingface/datasets/pull/460", "diff_url": "https://github.com/huggingface/datasets/pull/460.diff", "patch_url": "https://github.com/huggingface/datasets/pull/460.patch", "merged_at": 1596195138000 }
true
https://api.github.com/repos/huggingface/datasets/issues/459
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/459/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/459/comments
https://api.github.com/repos/huggingface/datasets/issues/459/events
https://github.com/huggingface/datasets/pull/459
669,545,437
MDExOlB1bGxSZXF1ZXN0NDU5OTAxMjEy
459
[Breaking] Update Dataset and DatasetDict API
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,596,183,093,000
1,598,430,516,000
1,598,430,515,000
MEMBER
null
This PR contains a few breaking changes so it's probably good to keep it for the next (major) release: - rename the `flatten`, `drop` and `dictionary_encode_column` methods in `flatten_`, `drop_` and `dictionary_encode_column_` to indicate that these methods have in-place effects as discussed in #166. From now on we s...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/459/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/459/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/459", "html_url": "https://github.com/huggingface/datasets/pull/459", "diff_url": "https://github.com/huggingface/datasets/pull/459.diff", "patch_url": "https://github.com/huggingface/datasets/pull/459.patch", "merged_at": 1598430515000 }
true
https://api.github.com/repos/huggingface/datasets/issues/458
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/458/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/458/comments
https://api.github.com/repos/huggingface/datasets/issues/458/events
https://github.com/huggingface/datasets/pull/458
668,972,666
MDExOlB1bGxSZXF1ZXN0NDU5Mzk5ODg2
458
Install CoVal metric from github
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[]
closed
false
null
[]
null
[]
1,596,128,365,000
1,596,203,793,000
1,596,203,793,000
MEMBER
null
Changed the import statements in `coval.py` to direct the user to install the original package from github if it's not already installed (the warning will only display properly after merging [PR455](https://github.com/huggingface/nlp/pull/455)) Also changed the function call to use named rather than positional argum...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/458/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/458/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/458", "html_url": "https://github.com/huggingface/datasets/pull/458", "diff_url": "https://github.com/huggingface/datasets/pull/458.diff", "patch_url": "https://github.com/huggingface/datasets/pull/458.patch", "merged_at": 1596203793000 }
true
https://api.github.com/repos/huggingface/datasets/issues/457
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/457/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/457/comments
https://api.github.com/repos/huggingface/datasets/issues/457/events
https://github.com/huggingface/datasets/pull/457
668,898,386
MDExOlB1bGxSZXF1ZXN0NDU5MzMyOTM1
457
add set_format to DatasetDict + tests
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,596,124,400,000
1,596,130,476,000
1,596,130,474,000
MEMBER
null
Add the `set_format` and `formated_as` and `reset_format` to `DatasetDict`. Add tests to these for `Dataset` and `DatasetDict`. Fix some bugs uncovered by the tests for `pandas` formating.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/457/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/457/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/457", "html_url": "https://github.com/huggingface/datasets/pull/457", "diff_url": "https://github.com/huggingface/datasets/pull/457.diff", "patch_url": "https://github.com/huggingface/datasets/pull/457.patch", "merged_at": 1596130474000 }
true
https://api.github.com/repos/huggingface/datasets/issues/456
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/456/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/456/comments
https://api.github.com/repos/huggingface/datasets/issues/456/events
https://github.com/huggingface/datasets/pull/456
668,723,785
MDExOlB1bGxSZXF1ZXN0NDU5MTc1MTY0
456
add crd3(ACL 2020) dataset
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,596,115,715,000
1,596,454,132,000
1,596,454,132,000
CONTRIBUTOR
null
This PR adds the **Critical Role Dungeons and Dragons Dataset** published at ACL 2020
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/456/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/456/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/456", "html_url": "https://github.com/huggingface/datasets/pull/456", "diff_url": "https://github.com/huggingface/datasets/pull/456.diff", "patch_url": "https://github.com/huggingface/datasets/pull/456.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/455
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/455/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/455/comments
https://api.github.com/repos/huggingface/datasets/issues/455/events
https://github.com/huggingface/datasets/pull/455
668,037,965
MDExOlB1bGxSZXF1ZXN0NDU4NTk4NTUw
455
Add bleurt
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[]
closed
false
null
[]
null
[ "Sorry one nit: Could we use named arguments for the call to BLEURT?\r\n\r\ni.e. \r\n scores = self.scorer.score(references=references, candidates=predictions)\r\n\r\n(i.e. so it is less bug prone)\r\n", "Following up on Ankur's comment---we are going to drop support for\npositional (not named) arguments i...
1,596,046,112,000
1,596,203,774,000
1,596,203,774,000
MEMBER
null
This PR adds the BLEURT metric to the library. The BLEURT `Metric` downloads a TF checkpoint corresponding to its `config_name` at creation (in the `_info` function). Default is set to `bleurt-base-128`. Note that the default in the original package is `bleurt-tiny-128`, but they throw a warning and recommend usi...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/455/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/455/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/455", "html_url": "https://github.com/huggingface/datasets/pull/455", "diff_url": "https://github.com/huggingface/datasets/pull/455.diff", "patch_url": "https://github.com/huggingface/datasets/pull/455.patch", "merged_at": 1596203774000 }
true
https://api.github.com/repos/huggingface/datasets/issues/454
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/454/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/454/comments
https://api.github.com/repos/huggingface/datasets/issues/454/events
https://github.com/huggingface/datasets/pull/454
668,011,577
MDExOlB1bGxSZXF1ZXN0NDU4NTc3MzA3
454
Create SECURITY.md
{ "login": "ChenZehong13", "id": 56394989, "node_id": "MDQ6VXNlcjU2Mzk0OTg5", "avatar_url": "https://avatars.githubusercontent.com/u/56394989?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ChenZehong13", "html_url": "https://github.com/ChenZehong13", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[]
1,596,043,414,000
1,596,059,152,000
1,596,059,152,000
NONE
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/454/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/454/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/454", "html_url": "https://github.com/huggingface/datasets/pull/454", "diff_url": "https://github.com/huggingface/datasets/pull/454.diff", "patch_url": "https://github.com/huggingface/datasets/pull/454.patch", "merged_at": null }
true
https://api.github.com/repos/huggingface/datasets/issues/453
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/453/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/453/comments
https://api.github.com/repos/huggingface/datasets/issues/453/events
https://github.com/huggingface/datasets/pull/453
667,728,247
MDExOlB1bGxSZXF1ZXN0NDU4MzQwNzky
453
add builder tests
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,596,018,127,000
1,596,021,246,000
1,596,021,245,000
MEMBER
null
I added `as_dataset` and `download_and_prepare` to the tests
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/453/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/453/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/453", "html_url": "https://github.com/huggingface/datasets/pull/453", "diff_url": "https://github.com/huggingface/datasets/pull/453.diff", "patch_url": "https://github.com/huggingface/datasets/pull/453.patch", "merged_at": 1596021245000 }
true
https://api.github.com/repos/huggingface/datasets/issues/452
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/452/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/452/comments
https://api.github.com/repos/huggingface/datasets/issues/452/events
https://github.com/huggingface/datasets/pull/452
667,498,295
MDExOlB1bGxSZXF1ZXN0NDU4MTUzNjQy
452
Guardian authorship dataset
{ "login": "malikaltakrori", "id": 25109412, "node_id": "MDQ6VXNlcjI1MTA5NDEy", "avatar_url": "https://avatars.githubusercontent.com/u/25109412?v=4", "gravatar_id": "", "url": "https://api.github.com/users/malikaltakrori", "html_url": "https://github.com/malikaltakrori", "followers_url": "https://api.gi...
[]
closed
false
null
[]
null
[ "Hi ! Glad you managed to fix the version issue.\r\n\r\nThe command `\r\npython nlp-cli dummy_data datasets/guardian_authorship --save_infos --all_configs` is supposed to generate a json file `dataset_infos.json` next to your dataset script, but I can't see it in the PR.\r\nCan you make sure you have the json file ...
1,595,989,437,000
1,597,936,197,000
1,597,936,076,000
CONTRIBUTOR
null
A new dataset: Guardian news articles for authorship attribution **tests passed:** python nlp-cli dummy_data datasets/guardian_authorship --save_infos --all_configs RUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_dataset_all_configs_guardian_authorship **Tests failed:** Real data:...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/452/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/452/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/452", "html_url": "https://github.com/huggingface/datasets/pull/452", "diff_url": "https://github.com/huggingface/datasets/pull/452.diff", "patch_url": "https://github.com/huggingface/datasets/pull/452.patch", "merged_at": 1597936075000 }
true
https://api.github.com/repos/huggingface/datasets/issues/451
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/451/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/451/comments
https://api.github.com/repos/huggingface/datasets/issues/451/events
https://github.com/huggingface/datasets/pull/451
667,210,468
MDExOlB1bGxSZXF1ZXN0NDU3OTIxNDMx
451
Fix csv/json/txt cache dir
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "I think this is the way to go but I’m afraid this might be a little slow. I was thinking that we could use a high quality very fast non crypto hash like xxhash for these stuff (hashing data files)", "Yep good idea, I'll take a look", "I tested the hashing speed [here](https://colab.research.google.com/drive/1h...
1,595,953,851,000
1,596,031,043,000
1,596,031,042,000
MEMBER
null
The cache dir for csv/json/txt datasets was always the same. This is an issue because it should be different depending on the data files provided by the user. To fix that, I added a line that use the hash of the data files provided by the user to define the cache dir. This should fix #444
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/451/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/451/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/451", "html_url": "https://github.com/huggingface/datasets/pull/451", "diff_url": "https://github.com/huggingface/datasets/pull/451.diff", "patch_url": "https://github.com/huggingface/datasets/pull/451.patch", "merged_at": 1596031042000 }
true
https://api.github.com/repos/huggingface/datasets/issues/450
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/450/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/450/comments
https://api.github.com/repos/huggingface/datasets/issues/450/events
https://github.com/huggingface/datasets/pull/450
667,074,120
MDExOlB1bGxSZXF1ZXN0NDU3ODA5ODA2
450
add sogou_news
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,595,942,950,000
1,596,029,418,000
1,596,029,417,000
CONTRIBUTOR
null
This PR adds the sogou news dataset #353
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/450/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/450/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/450", "html_url": "https://github.com/huggingface/datasets/pull/450", "diff_url": "https://github.com/huggingface/datasets/pull/450.diff", "patch_url": "https://github.com/huggingface/datasets/pull/450.patch", "merged_at": 1596029417000 }
true
https://api.github.com/repos/huggingface/datasets/issues/449
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/449/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/449/comments
https://api.github.com/repos/huggingface/datasets/issues/449/events
https://github.com/huggingface/datasets/pull/449
666,898,923
MDExOlB1bGxSZXF1ZXN0NDU3NjY0NjYx
449
add reuters21578 dataset
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "> Awesome !\r\n> Good job on parsing these files :O\r\n> \r\n> Do you think it would be hard to get the two other split configurations ?\r\n\r\nIt shouldn't be that hard, I think I can consider different config names for each split ", "> > Awesome !\r\n> > Good job on parsing these files :O\r\n> > Do you think i...
1,595,926,692,000
1,596,453,031,000
1,596,453,031,000
CONTRIBUTOR
null
This PR adds the `Reuters_21578` dataset https://kdd.ics.uci.edu/databases/reuters21578/reuters21578.html #353 The datasets is a lit of `.sgm` files which are a bit different from xml file indeed `xml.etree` couldn't be used to read files. I consider them as text file (to avoid using external library) and read ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/449/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/449/timeline
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/449", "html_url": "https://github.com/huggingface/datasets/pull/449", "diff_url": "https://github.com/huggingface/datasets/pull/449.diff", "patch_url": "https://github.com/huggingface/datasets/pull/449.patch", "merged_at": null }
true