id
int64
1.14B
2.23B
labels_url
stringlengths
75
75
body
stringlengths
2
33.9k
updated_at
stringlengths
20
20
number
int64
3.76k
6.79k
milestone
dict
repository_url
stringclasses
1 value
draft
bool
2 classes
labels
listlengths
0
4
created_at
stringlengths
20
20
comments_url
stringlengths
70
70
assignee
dict
timeline_url
stringlengths
70
70
title
stringlengths
1
290
events_url
stringlengths
68
68
active_lock_reason
null
user
dict
assignees
listlengths
0
3
performed_via_github_app
null
state_reason
stringclasses
3 values
author_association
stringclasses
3 values
closed_at
stringlengths
20
20
pull_request
dict
node_id
stringlengths
18
19
comments
listlengths
0
30
reactions
dict
state
stringclasses
2 values
locked
bool
1 class
url
stringlengths
61
61
html_url
stringlengths
49
51
is_pull_request
bool
2 classes
1,607,054,110
https://api.github.com/repos/huggingface/datasets/issues/5602/labels{/name}
This PR introduces new logic to `to_tf_dataset` affecting the returned data structure, enabling a dictionary structure to be returned, even if only one feature column is selected. If the passed in `columns` or `label_cols` to `to_tf_dataset` are a list, they are returned as a dictionary, respectively. If they are a string, the tensor is returned. An outline of the behaviour: ``` dataset,to_tf_dataset(columns=["col_1"], label_cols="col_2") # ({'col_1': col_1}, col_2} dataset,to_tf_dataset(columns="col1", label_cols="col_2") # (col1, col2) dataset,to_tf_dataset(columns="col1") # col1 dataset,to_tf_dataset(columns=["col_1"], labels=["col_2"]) # ({'col1': tensor}, {'col2': tensor}} dataset,to_tf_dataset(columns="col_1", labels=["col_2"]) # (col1, {'col2': tensor}} ``` ## Motivation Currently, when calling `to_tf_dataset`, the returned dataset will always return a tuple structure if a single feature column is used. This can cause issues when calling `model.fit` on models which train without labels e.g. [TFVitMAEForPreTraining](https://github.com/huggingface/transformers/blob/b6f47b539377ac1fd845c7adb4ccaa5eb514e126/src/transformers/models/vit_mae/modeling_vit_mae.py#L849). Specifically, [this line](https://github.com/huggingface/transformers/blob/d9e28d91a8b2d09b51a33155d3a03ad9fcfcbd1f/src/transformers/modeling_tf_utils.py#L1521) where it's assumed the input `x` is a dictionary if there is no label. ## Example Previous behaviour ```python In [1]: import tensorflow as tf ...: from datasets import load_dataset ...: ...: ...: def transform(batch): ...: def _transform_img(img): ...: img = img.convert("RGB") ...: img = tf.keras.utils.img_to_array(img) ...: img = tf.image.resize(img, (224, 224)) ...: img /= 255.0 ...: img = tf.transpose(img, perm=[2, 0, 1]) ...: return img ...: batch['pixel_values'] = [_transform_img(pil_img) for pil_img in batch['img']] ...: return batch ...: ...: ...: def collate_fn(examples): ...: pixel_values = tf.stack([example["pixel_values"] for example in examples]) ...: return {"pixel_values": pixel_values} ...: ...: ...: dataset = load_dataset('cifar10')['train'] ...: dataset = dataset.with_transform(transform) ...: dataset.to_tf_dataset(batch_size=8, columns=['pixel_values'], collate_fn=collate_fn) Out[1]: <PrefetchDataset element_spec=TensorSpec(shape=(None, 3, 224, 224), dtype=tf.float32, name=None)> ``` New behaviour ```python In [1]: import tensorflow as tf ...: from datasets import load_dataset ...: ...: ...: def transform(batch): ...: def _transform_img(img): ...: img = img.convert("RGB") ...: img = tf.keras.utils.img_to_array(img) ...: img = tf.image.resize(img, (224, 224)) ...: img /= 255.0 ...: img = tf.transpose(img, perm=[2, 0, 1]) ...: return img ...: batch['pixel_values'] = [_transform_img(pil_img) for pil_img in batch['img']] ...: return batch ...: ...: ...: def collate_fn(examples): ...: pixel_values = tf.stack([example["pixel_values"] for example in examples]) ...: return {"pixel_values": pixel_values} ...: ...: ...: dataset = load_dataset('cifar10')['train'] ...: dataset = dataset.with_transform(transform) ...: dataset.to_tf_dataset(batch_size=8, columns=['pixel_values'], collate_fn=collate_fn) Out[1]: <PrefetchDataset element_spec={'pixel_values': TensorSpec(shape=(None, 3, 224, 224), dtype=tf.float32, name=None)}> ```
2023-04-12T15:54:53Z
5,602
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-03-02T15:51:12Z
https://api.github.com/repos/huggingface/datasets/issues/5602/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5602/timeline
Return dict structure if columns are lists - to_tf_dataset
https://api.github.com/repos/huggingface/datasets/issues/5602/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/22614925?v=4", "events_url": "https://api.github.com/users/amyeroberts/events{/privacy}", "followers_url": "https://api.github.com/users/amyeroberts/followers", "following_url": "https://api.github.com/users/amyeroberts/following{/other_user}", "gists_url": "https://api.github.com/users/amyeroberts/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/amyeroberts", "id": 22614925, "login": "amyeroberts", "node_id": "MDQ6VXNlcjIyNjE0OTI1", "organizations_url": "https://api.github.com/users/amyeroberts/orgs", "received_events_url": "https://api.github.com/users/amyeroberts/received_events", "repos_url": "https://api.github.com/users/amyeroberts/repos", "site_admin": false, "starred_url": "https://api.github.com/users/amyeroberts/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amyeroberts/subscriptions", "type": "User", "url": "https://api.github.com/users/amyeroberts" }
[]
null
null
CONTRIBUTOR
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/5602.diff", "html_url": "https://github.com/huggingface/datasets/pull/5602", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5602.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5602" }
PR_kwDODunzps5LJGfa
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5602). All of your documentation changes will be reflected on that endpoint.", "This is a great PR! Thinking about the UX though, maybe we could do it without the extra argument? Before this PR, the logic in `to_tf_dataset` was...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5602/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5602
https://github.com/huggingface/datasets/pull/5602
true
1,606,685,976
https://api.github.com/repos/huggingface/datasets/issues/5601/labels{/name}
### Describe the bug Get `Authorization error` when try to push data into hugginface datasets hub. ### Steps to reproduce the bug I did all steps in the [tutorial](https://huggingface.co/docs/datasets/share), 1. `huggingface-cli login` with WRITE token 2. `git lfs install` 3. `git clone https://huggingface.co/datasets/namespace/your_dataset_name` 4. ``` cp /somewhere/data/*.json . git lfs track *.json git add .gitattributes git add *.json git commit -m "add json files" ``` but when I execute `git push` I got the error: ``` Uploading LFS objects: 0% (0/1), 0 B | 0 B/s, done. batch response: Authorization error. error: failed to push some refs to 'https://huggingface.co/datasets/zeusfsx/ukrainian-news' ``` Size of data ~100Gb. I have five json files - different parts. ### Expected behavior All my data pushed into hub ### Environment info - `datasets` version: 2.10.1 - Platform: macOS-13.2.1-arm64-arm-64bit - Python version: 3.10.10 - PyArrow version: 11.0.0 - Pandas version: 1.5.3
2023-03-14T16:55:35Z
5,601
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-03-02T12:08:39Z
https://api.github.com/repos/huggingface/datasets/issues/5601/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5601/timeline
Authorization error
https://api.github.com/repos/huggingface/datasets/issues/5601/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/107404835?v=4", "events_url": "https://api.github.com/users/OleksandrKorovii/events{/privacy}", "followers_url": "https://api.github.com/users/OleksandrKorovii/followers", "following_url": "https://api.github.com/users/OleksandrKorovii/following{/other_user}", "gists_url": "https://api.github.com/users/OleksandrKorovii/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/OleksandrKorovii", "id": 107404835, "login": "OleksandrKorovii", "node_id": "U_kgDOBmbeIw", "organizations_url": "https://api.github.com/users/OleksandrKorovii/orgs", "received_events_url": "https://api.github.com/users/OleksandrKorovii/received_events", "repos_url": "https://api.github.com/users/OleksandrKorovii/repos", "site_admin": false, "starred_url": "https://api.github.com/users/OleksandrKorovii/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/OleksandrKorovii/subscriptions", "type": "User", "url": "https://api.github.com/users/OleksandrKorovii" }
[]
null
completed
NONE
2023-03-14T16:55:34Z
null
I_kwDODunzps5fxBUY
[ "Hi! \r\n\r\nIt's better to report this kind of issue in the `huggingface_hub` repo, so if you still haven't resolved it, I suggest you open an issue there.", "Yeah, I solved it. Problem was in osxkeychain. When I do `hugginface-cli login` it's add token with default account (username)`hg_user` but my repo cont...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5601/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5601
https://github.com/huggingface/datasets/issues/5601
false
1,606,585,596
https://api.github.com/repos/huggingface/datasets/issues/5600/labels{/name}
### Describe the bug Dataloader getitem is not working as before (see example of [DreamboothDatasets](https://github.com/huggingface/peft/blob/main/examples/lora_dreambooth/train_dreambooth.py#L451C14-L529)) moving Datasets to 2.8.0 solved the issue. ### Steps to reproduce the bug 1- using DreamBoothDataset to load some images 2- error after loading when trying to visualise the images ### Expected behavior I was expecting a numpy array of the image ### Environment info - Platform: Linux-5.10.147+-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 9.0.0 - Pandas version: 1.3.5
2023-03-13T17:59:35Z
5,600
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-03-02T11:00:27Z
https://api.github.com/repos/huggingface/datasets/issues/5600/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5600/timeline
Dataloader getitem not working for DreamboothDatasets
https://api.github.com/repos/huggingface/datasets/issues/5600/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/76955987?v=4", "events_url": "https://api.github.com/users/salahiguiliz/events{/privacy}", "followers_url": "https://api.github.com/users/salahiguiliz/followers", "following_url": "https://api.github.com/users/salahiguiliz/following{/other_user}", "gists_url": "https://api.github.com/users/salahiguiliz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/salahiguiliz", "id": 76955987, "login": "salahiguiliz", "node_id": "MDQ6VXNlcjc2OTU1OTg3", "organizations_url": "https://api.github.com/users/salahiguiliz/orgs", "received_events_url": "https://api.github.com/users/salahiguiliz/received_events", "repos_url": "https://api.github.com/users/salahiguiliz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/salahiguiliz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/salahiguiliz/subscriptions", "type": "User", "url": "https://api.github.com/users/salahiguiliz" }
[]
null
completed
NONE
2023-03-13T17:59:35Z
null
I_kwDODunzps5fwoz8
[ "Hi! \r\n\r\n> (see example of DreamboothDatasets)\r\n\r\n\r\nCould you please provide a link to it? If you are referring to the example in the `diffusers` repo, your issue is unrelated to `datasets` as that example uses `Dataset` from PyTorch to load data." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5600/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5600
https://github.com/huggingface/datasets/issues/5600
false
1,605,018,478
https://api.github.com/repos/huggingface/datasets/issues/5598/labels{/name}
As reported in https://github.com/vijaydwivedi75/lrgb/issues/10, `push_to_hub` fails if the remote repository already exists and has a README.md without `dataset_info` in the YAML tags cc @clefourrier
2023-03-02T13:47:13Z
5,598
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-03-01T13:54:06Z
https://api.github.com/repos/huggingface/datasets/issues/5598/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5598/timeline
Fix push_to_hub with no dataset_infos
https://api.github.com/repos/huggingface/datasets/issues/5598/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-03-02T13:40:17Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5598.diff", "html_url": "https://github.com/huggingface/datasets/pull/5598", "merged_at": "2023-03-02T13:40:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/5598.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5598" }
PR_kwDODunzps5LCMiX
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5598/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5598
https://github.com/huggingface/datasets/pull/5598
true
1,604,928,721
https://api.github.com/repos/huggingface/datasets/issues/5597/labels{/name}
### Motivation For the circumstance that I creat an empty `Dataset` and keep appending new rows into it, I found that it leads to creating a new dataset at each call. It looks quite memory-consuming. I just wonder if there is any more efficient way to do this. ```python from datasets import Dataset ds = Dataset.from_list([]) ds.add_item({'a': [1, 2, 3], 'b': 4}) print(ds) >>> Dataset({ >>> features: [], >>> num_rows: 0 >>> }) ds = ds.add_item({'a': [1, 2, 3], 'b': 4}) print(ds) >>> Dataset({ >>> features: ['a', 'b'], >>> num_rows: 1 >>> }) ``` ### Feature request Call for in-place dataset update functions, that update the existing `Dataset` in place without creating a new copy. The interface is supposed to keep the same style as PyTorch, such as the in-place version of a `function` is named `function_`. For example, the in-pace version of `add_item`, i.e., `add_item_`, immediately updates the `Dataset`. ```python from datasets import Dataset ds = Dataset.from_list([]) ds.add_item({'a': [1, 2, 3], 'b': 4}) print(ds) >>> Dataset({ >>> features: [], >>> num_rows: 0 >>> }) ds.add_item_({'a': [1, 2, 3], 'b': 4}) print(ds) >>> Dataset({ >>> features: ['a', 'b'], >>> num_rows: 1 >>> }) ``` ### Related Functions * `.map` * `.filter` * `.add_item`
2023-03-02T13:30:41Z
5,597
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "ffffff", "default": true, "description": "This will not be worked on", "id": 1935892913, "name": "wontfix", "node_id": "MDU6TGFiZWwxOTM1ODkyOTEz", "url": "https://api.github.com/repos/huggingface/datasets/labels/wontfix" } ]
2023-03-01T12:58:18Z
https://api.github.com/repos/huggingface/datasets/issues/5597/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5597/timeline
in-place dataset update
https://api.github.com/repos/huggingface/datasets/issues/5597/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3585459?v=4", "events_url": "https://api.github.com/users/speedcell4/events{/privacy}", "followers_url": "https://api.github.com/users/speedcell4/followers", "following_url": "https://api.github.com/users/speedcell4/following{/other_user}", "gists_url": "https://api.github.com/users/speedcell4/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/speedcell4", "id": 3585459, "login": "speedcell4", "node_id": "MDQ6VXNlcjM1ODU0NTk=", "organizations_url": "https://api.github.com/users/speedcell4/orgs", "received_events_url": "https://api.github.com/users/speedcell4/received_events", "repos_url": "https://api.github.com/users/speedcell4/repos", "site_admin": false, "starred_url": "https://api.github.com/users/speedcell4/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/speedcell4/subscriptions", "type": "User", "url": "https://api.github.com/users/speedcell4" }
[]
null
completed
NONE
2023-03-02T03:47:00Z
null
I_kwDODunzps5fqUTR
[ "We won't support in-place modifications since `datasets` is based on the Apache Arrow format which doesn't support in-place modifications.\r\n\r\nIn your case the old dataset is garbage collected pretty quickly so you won't have memory issues.\r\n\r\nNote that datasets loaded from disk (memory mapped) are not load...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5597/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5597
https://github.com/huggingface/datasets/issues/5597
false
1,604,919,993
https://api.github.com/repos/huggingface/datasets/issues/5596/labels{/name}
### Describe the bug I'm trying to load this [dataset](https://huggingface.co/datasets/bigcode-data/the-stack-gh-issues) which consists of jsonl files and I get the following error: ``` casted_values = _c(array.values, feature[0]) File "/opt/conda/lib/python3.7/site-packages/datasets/table.py", line 1839, in wrapper return func(array, *args, **kwargs) File "/opt/conda/lib/python3.7/site-packages/datasets/table.py", line 2132, in cast_array_to_feature raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{feature}") TypeError: Couldn't cast array of type struct<type: string, action: string, datetime: timestamp[s], author: string, title: string, description: string, comment_id: int64, comment: string, labels: list<item: string>> to {'type': Value(dtype='string', id=None), 'action': Value(dtype='string', id=None), 'datetime': Value(dtype='timestamp[s]', id=None), 'author': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'description': Value(dtype='string', id=None), 'comment_id': Value(dtype='int64', id=None), 'comment': Value(dtype='string', id=None)} ``` But I can succesfully load a subset of the dataset, for example this works: ```python ds = load_dataset('bigcode-data/the-stack-gh-issues', split="train", data_files=[f"data/data-{x}.jsonl" for x in range(10)]) ``` and `ds.features` returns: ``` {'repo': Value(dtype='string', id=None), 'org': Value(dtype='string', id=None), 'issue_id': Value(dtype='int64', id=None), 'issue_number': Value(dtype='int64', id=None), 'pull_request': {'user_login': Value(dtype='string', id=None), 'repo': Value(dtype='string', id=None), 'number': Value(dtype='int64', id=None)}, 'events': [{'type': Value(dtype='string', id=None), 'action': Value(dtype='string', id=None), 'datetime': Value(dtype='timestamp[s]', id=None), 'author': Value(dtype='string', id=None), 'title': Value(dtype='string', id=None), 'description': Value(dtype='string', id=None), 'comment_id': Value(dtype='int64', id=None), 'comment': Value(dtype='string', id=None)}]} ``` So I'm not sure if there's an issue with just some of the files. Grateful if you have any suggestions to fix the issue. Side note: I saw this related [issue](https://github.com/huggingface/datasets/issues/3637) and tried to write a loading script to have `events` as a `Sequence` and not `list` [here](https://huggingface.co/datasets/bigcode-data/the-stack-gh-issues/blob/main/loading.py) (the script was renamed). It worked with a subset locally but doesn't for the remote dataset it can't find https://huggingface.co/datasets/bigcode-data/the-stack-gh-issues/resolve/main/data. ### Steps to reproduce the bug ```python from datasets import load_dataset ds = load_dataset('bigcode-data/the-stack-gh-issues', split="train") ``` ### Expected behavior Load the entire dataset succesfully. ### Environment info - `datasets` version: 2.10.1 - Platform: Linux-4.19.0-23-cloud-amd64-x86_64-with-debian-10.13 - Python version: 3.7.12 - PyArrow version: 9.0.0 - Pandas version: 1.3.4
2023-12-05T03:22:00Z
5,596
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-03-01T12:53:08Z
https://api.github.com/repos/huggingface/datasets/issues/5596/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5596/timeline
[TypeError: Couldn't cast array of type] Can only load a subset of the dataset
https://api.github.com/repos/huggingface/datasets/issues/5596/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/44069155?v=4", "events_url": "https://api.github.com/users/loubnabnl/events{/privacy}", "followers_url": "https://api.github.com/users/loubnabnl/followers", "following_url": "https://api.github.com/users/loubnabnl/following{/other_user}", "gists_url": "https://api.github.com/users/loubnabnl/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/loubnabnl", "id": 44069155, "login": "loubnabnl", "node_id": "MDQ6VXNlcjQ0MDY5MTU1", "organizations_url": "https://api.github.com/users/loubnabnl/orgs", "received_events_url": "https://api.github.com/users/loubnabnl/received_events", "repos_url": "https://api.github.com/users/loubnabnl/repos", "site_admin": false, "starred_url": "https://api.github.com/users/loubnabnl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/loubnabnl/subscriptions", "type": "User", "url": "https://api.github.com/users/loubnabnl" }
[]
null
completed
NONE
2023-03-02T11:12:11Z
null
I_kwDODunzps5fqSK5
[ "Apparently some JSON objects have a `\"labels\"` field. Since this field is not present in every object, you must specify all the fields types in the README.md\r\n\r\nEDIT: actually specifying the feature types doesn’t solve the issue, it raises an error because “labels” is missing in the data", "We've updated t...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5596/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5596
https://github.com/huggingface/datasets/issues/5596
false
1,604,070,629
https://api.github.com/repos/huggingface/datasets/issues/5595/labels{/name}
Closes #5477
2023-04-04T08:20:19Z
5,595
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-03-01T01:33:45Z
https://api.github.com/repos/huggingface/datasets/issues/5595/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5595/timeline
Unpins sqlAlchemy
https://api.github.com/repos/huggingface/datasets/issues/5595/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/46943923?v=4", "events_url": "https://api.github.com/users/lazarust/events{/privacy}", "followers_url": "https://api.github.com/users/lazarust/followers", "following_url": "https://api.github.com/users/lazarust/following{/other_user}", "gists_url": "https://api.github.com/users/lazarust/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lazarust", "id": 46943923, "login": "lazarust", "node_id": "MDQ6VXNlcjQ2OTQzOTIz", "organizations_url": "https://api.github.com/users/lazarust/orgs", "received_events_url": "https://api.github.com/users/lazarust/received_events", "repos_url": "https://api.github.com/users/lazarust/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lazarust/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lazarust/subscriptions", "type": "User", "url": "https://api.github.com/users/lazarust" }
[]
null
null
NONE
2023-04-04T08:19:14Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5595.diff", "html_url": "https://github.com/huggingface/datasets/pull/5595", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5595.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5595" }
PR_kwDODunzps5K--V9
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5595). All of your documentation changes will be reflected on that endpoint.", "It looks like this issue hasn't been fixed yet, so let's wait a bit more.", "@lazarust thanks for your work, but unfortunately we cannot merge it...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5595/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5595
https://github.com/huggingface/datasets/pull/5595
true
1,603,980,995
https://api.github.com/repos/huggingface/datasets/issues/5594/labels{/name}
### Describe the bug Hi, I am facing an error while downloading the xtreme udpos dataset using load_dataset. I have datasets 2.10.1 installed ```Downloading and preparing dataset xtreme/udpos.Arabic to /compute/tir-1-18/skhanuja/multilingual_ft/cache/data/xtreme/udpos.Arabic/1.0.0/29f5d57a48779f37ccb75cb8708d1095448aad0713b425bdc1ff9a4a128a56e4... Downloading data: 16%|██████████████▏ | 56.9M/355M [03:11<16:43, 297kB/s] Generating train split: 0%| | 0/6075 [00:00<?, ? examples/s]Traceback (most recent call last): File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/site-packages/datasets/builder.py", line 1608, in _prepare_split_single for key, record in generator: File "/home/skhanuja/.cache/huggingface/modules/datasets_modules/datasets/xtreme/29f5d57a48779f37ccb75cb8708d1095448aad0713b425bdc1ff9a4a128a56e4/xtreme.py", line 732, in _generate_examples yield from UdposParser.generate_examples(config=self.config, filepath=filepath, **kwargs) File "/home/skhanuja/.cache/huggingface/modules/datasets_modules/datasets/xtreme/29f5d57a48779f37ccb75cb8708d1095448aad0713b425bdc1ff9a4a128a56e4/xtreme.py", line 921, in generate_examples for path, file in filepath: File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/site-packages/datasets/download/download_manager.py", line 158, in __iter__ yield from self.generator(*self.args, **self.kwargs) File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/site-packages/datasets/download/download_manager.py", line 211, in _iter_from_path yield from cls._iter_tar(f) File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/site-packages/datasets/download/download_manager.py", line 167, in _iter_tar for tarinfo in stream: File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/tarfile.py", line 2475, in __iter__ tarinfo = self.next() File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/tarfile.py", line 2344, in next raise ReadError("unexpected end of data") tarfile.ReadError: unexpected end of data The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/skhanuja/Optimal-Resource-Allocation-for-Multilingual-Finetuning/src/train_al.py", line 855, in <module> main() File "/home/skhanuja/Optimal-Resource-Allocation-for-Multilingual-Finetuning/src/train_al.py", line 487, in main train_dataset = load_dataset(dataset_name, source_language, split="train", cache_dir=args.cache_dir, download_mode="force_redownload") File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/site-packages/datasets/load.py", line 1782, in load_dataset builder_instance.download_and_prepare( File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/site-packages/datasets/builder.py", line 872, in download_and_prepare self._download_and_prepare( File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/site-packages/datasets/builder.py", line 1649, in _download_and_prepare super()._download_and_prepare( File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/site-packages/datasets/builder.py", line 967, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/site-packages/datasets/builder.py", line 1488, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/home/skhanuja/miniconda3/envs/multilingual_ft/lib/python3.10/site-packages/datasets/builder.py", line 1644, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.builder.DatasetGenerationError: An error occurred while generating the dataset ``` ### Steps to reproduce the bug ``` train_dataset = load_dataset('xtreme', 'udpos.English', split="train", cache_dir=args.cache_dir, download_mode="force_redownload") ``` ### Expected behavior Download the udpos dataset ### Environment info - `datasets` version: 2.10.1 - Platform: Linux-3.10.0-957.1.3.el7.x86_64-x86_64-with-glibc2.17 - Python version: 3.10.8 - PyArrow version: 10.0.1 - Pandas version: 1.5.2
2023-11-04T20:45:56Z
5,594
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-28T23:40:53Z
https://api.github.com/repos/huggingface/datasets/issues/5594/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5594/timeline
Error while downloading the xtreme udpos dataset
https://api.github.com/repos/huggingface/datasets/issues/5594/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/24687672?v=4", "events_url": "https://api.github.com/users/simran-khanuja/events{/privacy}", "followers_url": "https://api.github.com/users/simran-khanuja/followers", "following_url": "https://api.github.com/users/simran-khanuja/following{/other_user}", "gists_url": "https://api.github.com/users/simran-khanuja/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/simran-khanuja", "id": 24687672, "login": "simran-khanuja", "node_id": "MDQ6VXNlcjI0Njg3Njcy", "organizations_url": "https://api.github.com/users/simran-khanuja/orgs", "received_events_url": "https://api.github.com/users/simran-khanuja/received_events", "repos_url": "https://api.github.com/users/simran-khanuja/repos", "site_admin": false, "starred_url": "https://api.github.com/users/simran-khanuja/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/simran-khanuja/subscriptions", "type": "User", "url": "https://api.github.com/users/simran-khanuja" }
[]
null
completed
NONE
2023-07-24T14:22:18Z
null
I_kwDODunzps5fms7D
[ "Hi! I cannot reproduce this error on my machine.\r\n\r\nThe raised error could mean that one of the downloaded files is corrupted. To verify this is not the case, you can run `load_dataset` as follows:\r\n```python\r\ntrain_dataset = load_dataset('xtreme', 'udpos.English', split=\"train\", cache_dir=args.cache_dir...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5594/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5594
https://github.com/huggingface/datasets/issues/5594
false
1,603,619,124
https://api.github.com/repos/huggingface/datasets/issues/5592/labels{/name}
Fixes #5581 to use the correct output for the `set_format` method.
2023-02-28T19:26:33Z
5,592
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-28T18:42:37Z
https://api.github.com/repos/huggingface/datasets/issues/5592/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5592/timeline
Fix docstring example
https://api.github.com/repos/huggingface/datasets/issues/5592/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
null
null
MEMBER
2023-02-28T19:19:15Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5592.diff", "html_url": "https://github.com/huggingface/datasets/pull/5592", "merged_at": "2023-02-28T19:19:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/5592.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5592" }
PR_kwDODunzps5K9dWr
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5592/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5592
https://github.com/huggingface/datasets/pull/5592
true
1,603,571,407
https://api.github.com/repos/huggingface/datasets/issues/5591/labels{/name}
null
2023-02-28T18:16:31Z
5,591
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-28T18:09:05Z
https://api.github.com/repos/huggingface/datasets/issues/5591/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5591/timeline
set dev version
https://api.github.com/repos/huggingface/datasets/issues/5591/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-28T18:09:15Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5591.diff", "html_url": "https://github.com/huggingface/datasets/pull/5591", "merged_at": "2023-02-28T18:09:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/5591.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5591" }
PR_kwDODunzps5K9S79
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5591). All of your documentation changes will be reflected on that endpoint.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5591/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5591
https://github.com/huggingface/datasets/pull/5591
true
1,603,549,504
https://api.github.com/repos/huggingface/datasets/issues/5590/labels{/name}
null
2023-02-28T18:16:27Z
5,590
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-28T17:58:11Z
https://api.github.com/repos/huggingface/datasets/issues/5590/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5590/timeline
Release: 2.10.1
https://api.github.com/repos/huggingface/datasets/issues/5590/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-28T18:06:08Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5590.diff", "html_url": "https://github.com/huggingface/datasets/pull/5590", "merged_at": "2023-02-28T18:06:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/5590.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5590" }
PR_kwDODunzps5K9N_H
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5590/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5590
https://github.com/huggingface/datasets/pull/5590
true
1,603,535,704
https://api.github.com/repos/huggingface/datasets/issues/5589/labels{/name}
This reverts commit b91070b9c09673e2e148eec458036ab6a62ac042 (temporarily) It hurts iterable dataset performance a lot (e.g. x4 slower because it encodes+decodes images unnecessarily). I think we need to fix this before re-adding it cc @mariosasko @Hubert-Bonisseur
2023-09-24T10:07:33Z
5,589
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-28T17:52:04Z
https://api.github.com/repos/huggingface/datasets/issues/5589/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5589/timeline
Revert "pass the dataset features to the IterableDataset.from_generator"
https://api.github.com/repos/huggingface/datasets/issues/5589/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-03-21T14:18:18Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5589.diff", "html_url": "https://github.com/huggingface/datasets/pull/5589", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5589.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5589" }
PR_kwDODunzps5K9K1i
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5589/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5589
https://github.com/huggingface/datasets/pull/5589
true
1,603,304,766
https://api.github.com/repos/huggingface/datasets/issues/5588/labels{/name}
Flatten a dataset on the fly in `save_to_disk` instead of doing it with `flatten_indices` to avoid creating an additional cache file. (this is one of the sub-tasks in https://github.com/huggingface/datasets/issues/5507)
2023-02-28T17:28:35Z
5,588
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-28T15:37:46Z
https://api.github.com/repos/huggingface/datasets/issues/5588/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5588/timeline
Flatten dataset on the fly in `save_to_disk`
https://api.github.com/repos/huggingface/datasets/issues/5588/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2023-02-28T17:21:17Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5588.diff", "html_url": "https://github.com/huggingface/datasets/pull/5588", "merged_at": "2023-02-28T17:21:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/5588.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5588" }
PR_kwDODunzps5K8YYz
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5588/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5588
https://github.com/huggingface/datasets/pull/5588
true
1,603,139,420
https://api.github.com/repos/huggingface/datasets/issues/5587/labels{/name}
Fixes the `key` range in the `query_table` call in `sort` to account for an indices mapping Fix #5586
2023-02-28T17:28:57Z
5,587
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-28T14:05:08Z
https://api.github.com/repos/huggingface/datasets/issues/5587/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5587/timeline
Fix `sort` with indices mapping
https://api.github.com/repos/huggingface/datasets/issues/5587/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2023-02-28T17:21:58Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5587.diff", "html_url": "https://github.com/huggingface/datasets/pull/5587", "merged_at": "2023-02-28T17:21:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/5587.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5587" }
PR_kwDODunzps5K70pp
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5587/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5587
https://github.com/huggingface/datasets/pull/5587
true
1,602,961,544
https://api.github.com/repos/huggingface/datasets/issues/5586/labels{/name}
### Describe the bug Hi, thank you for your support! It seems like the addition of multiple key sort (#5502) in 2.10.0 broke the `.sort()` method. After filtering a dataset with `.filter()`, the `.sort()` seems to refer to the query_table index of the previous unfiltered dataset, resulting in an IndexError. This only happens with the 2.10.0 release. ### Steps to reproduce the bug ```Python from datasets import load_dataset # dataset with length of 1104 ds = load_dataset('glue', 'ax')['test'] ds = ds.filter(lambda x: x['idx'] > 1100) ds.sort('premise') print('Done') ``` File "/home/dongkeun/datasets_test/test.py", line 5, in <module> ds.sort('premise') File "/home/dongkeun/miniconda3/envs/datasets_test/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 528, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/dongkeun/miniconda3/envs/datasets_test/lib/python3.9/site-packages/datasets/fingerprint.py", line 511, in wrapper out = func(dataset, *args, **kwargs) File "/home/dongkeun/miniconda3/envs/datasets_test/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3959, in sort sort_table = query_table( File "/home/dongkeun/miniconda3/envs/datasets_test/lib/python3.9/site-packages/datasets/formatting/formatting.py", line 588, in query_table _check_valid_index_key(key, size) File "/home/dongkeun/miniconda3/envs/datasets_test/lib/python3.9/site-packages/datasets/formatting/formatting.py", line 537, in _check_valid_index_key _check_valid_index_key(max(key), size=size) File "/home/dongkeun/miniconda3/envs/datasets_test/lib/python3.9/site-packages/datasets/formatting/formatting.py", line 531, in _check_valid_index_key raise IndexError(f"Invalid key: {key} is out of bounds for size {size}") IndexError: Invalid key: 1103 is out of bounds for size 3 ### Expected behavior It should sort the dataset and print "Done". Which it does on 2.9.0. ### Environment info - `datasets` version: 2.10.0 - Platform: Linux-5.15.0-41-generic-x86_64-with-glibc2.31 - Python version: 3.9.16 - PyArrow version: 11.0.0 - Pandas version: 1.5.3
2023-02-28T18:17:26Z
5,586
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2023-02-28T12:18:09Z
https://api.github.com/repos/huggingface/datasets/issues/5586/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5586/timeline
.sort() is broken when used after .filter(), only in 2.10.0
https://api.github.com/repos/huggingface/datasets/issues/5586/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/57797966?v=4", "events_url": "https://api.github.com/users/MattYoon/events{/privacy}", "followers_url": "https://api.github.com/users/MattYoon/followers", "following_url": "https://api.github.com/users/MattYoon/following{/other_user}", "gists_url": "https://api.github.com/users/MattYoon/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/MattYoon", "id": 57797966, "login": "MattYoon", "node_id": "MDQ6VXNlcjU3Nzk3OTY2", "organizations_url": "https://api.github.com/users/MattYoon/orgs", "received_events_url": "https://api.github.com/users/MattYoon/received_events", "repos_url": "https://api.github.com/users/MattYoon/repos", "site_admin": false, "starred_url": "https://api.github.com/users/MattYoon/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MattYoon/subscriptions", "type": "User", "url": "https://api.github.com/users/MattYoon" }
[]
null
completed
NONE
2023-02-28T17:21:59Z
null
I_kwDODunzps5fi0CI
[ "Thanks for reporting and thanks @mariosasko for fixing ! We just did a patch release `2.10.1` with the fix" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5586/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5586
https://github.com/huggingface/datasets/issues/5586
false
1,602,190,030
https://api.github.com/repos/huggingface/datasets/issues/5585/labels{/name}
### Describe the bug I would like to share cache between two machines (a Windows host machine and a WSL instance). I run most my code in WSL. I have just run out of space in the virtual drive. Rather than expand the drive size, I plan to move to cache to the host Windows machine, thereby sharing the downloads. I'm hoping that I can just copy/paste the cache files, but I notice that a lot of the file names start with the path name, e.g. `_home_davidg_.cache_huggingface_datasets_conll2003_default-451...98.lock` where `home/davidg` is where the cache is in WSL. This seems to suggest that the cache is not portable/cannot be centralised or shared. Is this the case, or are the files that start with path names not integral to the caching mechanism? Because copying the cache files _seems_ to work, but I'm not filled with confidence that something isn't going to break. A related issue, when trying to load a dataset that should come from cache (running in WSL, pointing to cache on the Windows host) it seemed to work fine, but it still uses a WSL directory for `.cache\huggingface\modules\datasets_modules`. I see nothing in the docs about this, or how to point it to a different place. I have asked a related question on the forum: https://discuss.huggingface.co/t/is-datasets-cache-operating-system-agnostic/32656 ### Steps to reproduce the bug View the cache directory in WSL/Windows. ### Expected behavior Cache can be shared between (virtual) machines and be transportable. It would be nice to have a simple way to say "Dear Hugging Face packages, please put ALL your cache in `blah/de/blah`" and have all the Hugging Face packages respect that single location. ### Environment info ``` - `datasets` version: 2.9.0 - Platform: Linux-5.10.102.1-microsoft-standard-WSL2-x86_64-with-glibc2.31 - Python version: 3.10.8 - PyArrow version: 11.0.0 - Pandas version: 1.5.3 - ```
2023-02-28T21:26:52Z
5,585
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-28T00:53:06Z
https://api.github.com/repos/huggingface/datasets/issues/5585/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5585/timeline
Cache is not transportable
https://api.github.com/repos/huggingface/datasets/issues/5585/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/4443482?v=4", "events_url": "https://api.github.com/users/davidgilbertson/events{/privacy}", "followers_url": "https://api.github.com/users/davidgilbertson/followers", "following_url": "https://api.github.com/users/davidgilbertson/following{/other_user}", "gists_url": "https://api.github.com/users/davidgilbertson/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/davidgilbertson", "id": 4443482, "login": "davidgilbertson", "node_id": "MDQ6VXNlcjQ0NDM0ODI=", "organizations_url": "https://api.github.com/users/davidgilbertson/orgs", "received_events_url": "https://api.github.com/users/davidgilbertson/received_events", "repos_url": "https://api.github.com/users/davidgilbertson/repos", "site_admin": false, "starred_url": "https://api.github.com/users/davidgilbertson/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/davidgilbertson/subscriptions", "type": "User", "url": "https://api.github.com/users/davidgilbertson" }
[]
null
completed
NONE
2023-02-28T21:26:52Z
null
I_kwDODunzps5ff3rO
[ "Hi ! No the cache is not transportable in general. It will work on a shared filesystem if you use the same python environment, but not across machines/os/environments.\r\n\r\nIn particular, reloading cached datasets does work, but reloading cached processed datasets (e.g. from `map`) may not work. This is because ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5585/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5585
https://github.com/huggingface/datasets/issues/5585
false
1,601,821,808
https://api.github.com/repos/huggingface/datasets/issues/5584/labels{/name}
### Describe the bug Seeing this error when downloading https://huggingface.co/datasets/kakaobrain/coyo-700m: ```ArrowInvalid: Parquet magic bytes not found in footer. Either the file is corrupted or this is not a parquet file.``` Full stack trace ```Downloading and preparing dataset parquet/kakaobrain--coyo-700m to /root/.cache/huggingface/datasets/kakaobrain___parquet/kakaobrain--coyo-700m-ae729692ae3e0073/0.0.0/2a3b91fbd88a2c90d1dbbb32b460cf621d31bd5b05b934492fdef7d8d6f236ec... Downloading data files: 100% 1/1 [00:00<00:00, 63.35it/s] Extracting data files: 100% 1/1 [00:00<00:00, 5.00it/s] --------------------------------------------------------------------------- ArrowInvalid Traceback (most recent call last) [/usr/local/lib/python3.8/dist-packages/datasets/builder.py](https://localhost:8080/#) in _prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id) 1859 _time = time.time() -> 1860 for _, table in generator: 1861 if max_shard_size is not None and writer._num_bytes > max_shard_size: 9 frames ArrowInvalid: Parquet magic bytes not found in footer. Either the file is corrupted or this is not a parquet file. The above exception was the direct cause of the following exception: DatasetGenerationError Traceback (most recent call last) [/usr/local/lib/python3.8/dist-packages/datasets/builder.py](https://localhost:8080/#) in _prepare_split_single(self, gen_kwargs, fpath, file_format, max_shard_size, job_id) 1890 if isinstance(e, SchemaInferenceError) and e.__context__ is not None: 1891 e = e.__context__ -> 1892 raise DatasetGenerationError("An error occurred while generating the dataset") from e 1893 1894 yield job_id, True, (total_num_examples, total_num_bytes, writer._features, num_shards, shard_lengths) DatasetGenerationError: An error occurred while generating the dataset``` ### Steps to reproduce the bug ``` from datasets import load_dataset hf_dataset = load_dataset("kakaobrain/coyo-700m") ``` ### Expected behavior The above commands load the dataset successfully. Or handles exception and continue loading the remainder. ### Environment info colab. any
2023-02-28T07:27:59Z
5,584
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-27T19:35:03Z
https://api.github.com/repos/huggingface/datasets/issues/5584/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5584/timeline
Unable to load coyo700M dataset
https://api.github.com/repos/huggingface/datasets/issues/5584/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3059998?v=4", "events_url": "https://api.github.com/users/manuaero/events{/privacy}", "followers_url": "https://api.github.com/users/manuaero/followers", "following_url": "https://api.github.com/users/manuaero/following{/other_user}", "gists_url": "https://api.github.com/users/manuaero/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/manuaero", "id": 3059998, "login": "manuaero", "node_id": "MDQ6VXNlcjMwNTk5OTg=", "organizations_url": "https://api.github.com/users/manuaero/orgs", "received_events_url": "https://api.github.com/users/manuaero/received_events", "repos_url": "https://api.github.com/users/manuaero/repos", "site_admin": false, "starred_url": "https://api.github.com/users/manuaero/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/manuaero/subscriptions", "type": "User", "url": "https://api.github.com/users/manuaero" }
[]
null
completed
NONE
2023-02-28T07:27:58Z
null
I_kwDODunzps5fedxw
[ "Hi @manuaero \r\n\r\nThank you for your interest in the COYO dataset.\r\n\r\nOur dataset provides the img-url and alt-text in the form of a parquet, so to utilize the coyo dataset you will need to download it directly.\r\n\r\nWe provide a [guide](https://github.com/kakaobrain/coyo-dataset/blob/main/download/README...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5584/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5584
https://github.com/huggingface/datasets/issues/5584
false
1,601,583,625
https://api.github.com/repos/huggingface/datasets/issues/5583/labels{/name}
Ensures all the writers that use Pandas for conversion (JSON, CSV, SQL) do not export `index` by default (https://github.com/huggingface/datasets/pull/5490 only did this for CSV)
2023-02-28T13:52:15Z
5,583
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-27T17:04:46Z
https://api.github.com/repos/huggingface/datasets/issues/5583/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5583/timeline
Do no write index by default when exporting a dataset
https://api.github.com/repos/huggingface/datasets/issues/5583/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2023-02-28T13:44:04Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5583.diff", "html_url": "https://github.com/huggingface/datasets/pull/5583", "merged_at": "2023-02-28T13:44:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/5583.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5583" }
PR_kwDODunzps5K2mIz
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5583/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5583
https://github.com/huggingface/datasets/pull/5583
true
1,600,932,092
https://api.github.com/repos/huggingface/datasets/issues/5582/labels{/name}
This PR closes #5383 * Add column_names property to IterableDataset * Add multiple tests for this new property
2023-03-13T19:10:22Z
5,582
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-27T10:50:07Z
https://api.github.com/repos/huggingface/datasets/issues/5582/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5582/timeline
Add column_names to IterableDataset
https://api.github.com/repos/huggingface/datasets/issues/5582/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/50772274?v=4", "events_url": "https://api.github.com/users/patrickloeber/events{/privacy}", "followers_url": "https://api.github.com/users/patrickloeber/followers", "following_url": "https://api.github.com/users/patrickloeber/following{/other_user}", "gists_url": "https://api.github.com/users/patrickloeber/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickloeber", "id": 50772274, "login": "patrickloeber", "node_id": "MDQ6VXNlcjUwNzcyMjc0", "organizations_url": "https://api.github.com/users/patrickloeber/orgs", "received_events_url": "https://api.github.com/users/patrickloeber/received_events", "repos_url": "https://api.github.com/users/patrickloeber/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickloeber/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickloeber/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickloeber" }
[]
null
null
CONTRIBUTOR
2023-03-13T19:03:32Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5582.diff", "html_url": "https://github.com/huggingface/datasets/pull/5582", "merged_at": "2023-03-13T19:03:31Z", "patch_url": "https://github.com/huggingface/datasets/pull/5582.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5582" }
PR_kwDODunzps5K0ZcN
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5582/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5582
https://github.com/huggingface/datasets/pull/5582
true
1,600,675,489
https://api.github.com/repos/huggingface/datasets/issues/5581/labels{/name}
### Describe the bug https://huggingface.co/docs/datasets/v2.10.0/en/package_reference/main_classes#datasets.Dataset.set_format <img width="700" alt="image" src="https://user-images.githubusercontent.com/36224762/221506973-ae2e3991-60a7-4d4e-99f8-965c6eb61e59.png"> While actually running it will result in: <img width="1094" alt="image" src="https://user-images.githubusercontent.com/36224762/221507032-007dab82-8781-4319-b21a-e6e4d40d97b3.png"> ### Steps to reproduce the bug _ ### Expected behavior _ ### Environment info - `datasets` version: 2.10.0 - Platform: Linux-5.10.147+-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 9.0.0 - Pandas version: 1.3.5
2023-02-28T19:19:17Z
5,581
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "7057ff", "default": true, "description": "Good for newcomers", "id": 1935892877, "name": "good first issue", "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue" } ]
2023-02-27T08:03:09Z
https://api.github.com/repos/huggingface/datasets/issues/5581/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5581/timeline
[DOC] Mistaken docs on set_format
https://api.github.com/repos/huggingface/datasets/issues/5581/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36224762?v=4", "events_url": "https://api.github.com/users/NightMachinery/events{/privacy}", "followers_url": "https://api.github.com/users/NightMachinery/followers", "following_url": "https://api.github.com/users/NightMachinery/following{/other_user}", "gists_url": "https://api.github.com/users/NightMachinery/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NightMachinery", "id": 36224762, "login": "NightMachinery", "node_id": "MDQ6VXNlcjM2MjI0NzYy", "organizations_url": "https://api.github.com/users/NightMachinery/orgs", "received_events_url": "https://api.github.com/users/NightMachinery/received_events", "repos_url": "https://api.github.com/users/NightMachinery/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NightMachinery/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NightMachinery/subscriptions", "type": "User", "url": "https://api.github.com/users/NightMachinery" }
[]
null
completed
CONTRIBUTOR
2023-02-28T19:19:17Z
null
I_kwDODunzps5faF6h
[ "Thanks for reporting!" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5581/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5581
https://github.com/huggingface/datasets/issues/5581
false
1,600,431,792
https://api.github.com/repos/huggingface/datasets/issues/5580/labels{/name}
Closes https://github.com/huggingface/datasets/issues/5281 This PR uses fsspec to support datasets on cloud storage (tested manually with GCS). ETags are currently unsupported for cloud storage. In general, a much larger refactor could be done to just use fsspec for all schemes (ftp, http/s, s3, gcs) to unify the interfaces here, but I ultimately opted to leave that out of this PR I didn't create a GCS filesystem class in `datasets.filesystems` since the S3 one appears to be a wrapper around `s3fs.S3FileSystem` and mainly used to generate docs.
2023-03-11T01:02:49Z
5,580
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-27T04:06:05Z
https://api.github.com/repos/huggingface/datasets/issues/5580/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5580/timeline
Support cloud storage in load_dataset via fsspec
https://api.github.com/repos/huggingface/datasets/issues/5580/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/2512762?v=4", "events_url": "https://api.github.com/users/dwyatte/events{/privacy}", "followers_url": "https://api.github.com/users/dwyatte/followers", "following_url": "https://api.github.com/users/dwyatte/following{/other_user}", "gists_url": "https://api.github.com/users/dwyatte/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dwyatte", "id": 2512762, "login": "dwyatte", "node_id": "MDQ6VXNlcjI1MTI3NjI=", "organizations_url": "https://api.github.com/users/dwyatte/orgs", "received_events_url": "https://api.github.com/users/dwyatte/received_events", "repos_url": "https://api.github.com/users/dwyatte/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dwyatte/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dwyatte/subscriptions", "type": "User", "url": "https://api.github.com/users/dwyatte" }
[]
null
null
CONTRIBUTOR
2023-03-11T00:55:40Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5580.diff", "html_url": "https://github.com/huggingface/datasets/pull/5580", "merged_at": "2023-03-11T00:55:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/5580.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5580" }
PR_kwDODunzps5Kys1c
[ "_The documentation is not available anymore as the PR was closed or merged._", "> Regarding the tests I think it should be possible to use the mockfs fixture, it allows to play with a dummy fsspec FileSystem with the \"mock://\" protocol.\r\n\r\n> However it requires some storage_options to be passed. Maybe it c...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5580/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5580
https://github.com/huggingface/datasets/pull/5580
true
1,599,732,211
https://api.github.com/repos/huggingface/datasets/issues/5579/labels{/name}
The following adds instructions on how to create a `DataLoader` from the guide on how to use object detection with augmentations (#4710). I am open to hearing any suggestions for improvement !
2023-03-23T19:24:59Z
5,579
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-25T14:53:17Z
https://api.github.com/repos/huggingface/datasets/issues/5579/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5579/timeline
Add instructions to create `DataLoader` from augmented dataset in object detection guide
https://api.github.com/repos/huggingface/datasets/issues/5579/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/21087104?v=4", "events_url": "https://api.github.com/users/Laurent2916/events{/privacy}", "followers_url": "https://api.github.com/users/Laurent2916/followers", "following_url": "https://api.github.com/users/Laurent2916/following{/other_user}", "gists_url": "https://api.github.com/users/Laurent2916/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Laurent2916", "id": 21087104, "login": "Laurent2916", "node_id": "MDQ6VXNlcjIxMDg3MTA0", "organizations_url": "https://api.github.com/users/Laurent2916/orgs", "received_events_url": "https://api.github.com/users/Laurent2916/received_events", "repos_url": "https://api.github.com/users/Laurent2916/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Laurent2916/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Laurent2916/subscriptions", "type": "User", "url": "https://api.github.com/users/Laurent2916" }
[]
null
null
CONTRIBUTOR
2023-03-23T19:24:50Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5579.diff", "html_url": "https://github.com/huggingface/datasets/pull/5579", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5579.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5579" }
PR_kwDODunzps5Kwgo4
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5579). All of your documentation changes will be reflected on that endpoint.", "I'm not sure we need this part as we provide a link to the notebook that shows how to train an object detection model, and this notebook instantiat...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5579/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5579
https://github.com/huggingface/datasets/pull/5579
true
1,598,863,119
https://api.github.com/repos/huggingface/datasets/issues/5578/labels{/name}
Add the `huggingface_hub` version to the `env` command's output.
2023-02-27T17:28:25Z
5,578
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-24T15:37:43Z
https://api.github.com/repos/huggingface/datasets/issues/5578/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5578/timeline
Add `huggingface_hub` version to env cli command
https://api.github.com/repos/huggingface/datasets/issues/5578/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2023-02-27T17:21:09Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5578.diff", "html_url": "https://github.com/huggingface/datasets/pull/5578", "merged_at": "2023-02-27T17:21:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/5578.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5578" }
PR_kwDODunzps5Kto96
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5578/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5578
https://github.com/huggingface/datasets/pull/5578
true
1,598,587,665
https://api.github.com/repos/huggingface/datasets/issues/5577/labels{/name}
### Describe the bug I met the same bug mentioned in #3053 which is never fixed. Because several `reddit_scores` are larger than `int8` even `int16`. https://huggingface.co/datasets/the_pile_openwebtext2/blob/main/the_pile_openwebtext2.py#L62 ### Steps to reproduce the bug ```python3 from datasets import load_dataset dataset = load_dataset("the_pile_openwebtext2") ``` ### Expected behavior load as normal. ### Environment info - `datasets` version: 2.10.0 - Platform: Linux-5.4.143.bsk.7-amd64-x86_64-with-glibc2.31 - Python version: 3.9.2 - PyArrow version: 11.0.0 - Pandas version: 1.5.3
2023-02-24T14:01:09Z
5,577
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-24T13:01:48Z
https://api.github.com/repos/huggingface/datasets/issues/5577/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5577/timeline
Cannot load `the_pile_openwebtext2`
https://api.github.com/repos/huggingface/datasets/issues/5577/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/5126316?v=4", "events_url": "https://api.github.com/users/wjfwzzc/events{/privacy}", "followers_url": "https://api.github.com/users/wjfwzzc/followers", "following_url": "https://api.github.com/users/wjfwzzc/following{/other_user}", "gists_url": "https://api.github.com/users/wjfwzzc/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/wjfwzzc", "id": 5126316, "login": "wjfwzzc", "node_id": "MDQ6VXNlcjUxMjYzMTY=", "organizations_url": "https://api.github.com/users/wjfwzzc/orgs", "received_events_url": "https://api.github.com/users/wjfwzzc/received_events", "repos_url": "https://api.github.com/users/wjfwzzc/repos", "site_admin": false, "starred_url": "https://api.github.com/users/wjfwzzc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wjfwzzc/subscriptions", "type": "User", "url": "https://api.github.com/users/wjfwzzc" }
[]
null
completed
NONE
2023-02-24T14:01:09Z
null
I_kwDODunzps5fSIMR
[ "Hi! I've merged a PR to use `int32` instead of `int8` for `reddit_scores`, so it should work now.\r\n\r\n" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5577/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5577
https://github.com/huggingface/datasets/issues/5577
false
1,598,582,744
https://api.github.com/repos/huggingface/datasets/issues/5576/labels{/name}
I was getting a similar error `pyarrow.lib.ArrowInvalid: Integer value 528 not in range: -128 to 127` - AFAICT, this is because the type specified for `reddit_scores` is `datasets.Sequence(datasets.Value("int8"))`, but the actual values can be well outside the max range for 8-bit integers. I worked around this by downloading the `the_pile_openwebtext2.py` and editing it to use local files and drop reddit scores as a column (not needed for my purposes). _Originally posted by @tc-wolf in https://github.com/huggingface/datasets/issues/3053#issuecomment-1281392422_
2023-02-24T12:58:31Z
5,576
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-24T12:57:49Z
https://api.github.com/repos/huggingface/datasets/issues/5576/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5576/timeline
I was getting a similar error `pyarrow.lib.ArrowInvalid: Integer value 528 not in range: -128 to 127` - AFAICT, this is because the type specified for `reddit_scores` is `datasets.Sequence(datasets.Value("int8"))`, but the actual values can be well outside the max range for 8-bit integers.
https://api.github.com/repos/huggingface/datasets/issues/5576/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/5126316?v=4", "events_url": "https://api.github.com/users/wjfwzzc/events{/privacy}", "followers_url": "https://api.github.com/users/wjfwzzc/followers", "following_url": "https://api.github.com/users/wjfwzzc/following{/other_user}", "gists_url": "https://api.github.com/users/wjfwzzc/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/wjfwzzc", "id": 5126316, "login": "wjfwzzc", "node_id": "MDQ6VXNlcjUxMjYzMTY=", "organizations_url": "https://api.github.com/users/wjfwzzc/orgs", "received_events_url": "https://api.github.com/users/wjfwzzc/received_events", "repos_url": "https://api.github.com/users/wjfwzzc/repos", "site_admin": false, "starred_url": "https://api.github.com/users/wjfwzzc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wjfwzzc/subscriptions", "type": "User", "url": "https://api.github.com/users/wjfwzzc" }
[]
null
not_planned
NONE
2023-02-24T12:58:18Z
null
I_kwDODunzps5fSG_Y
[ "Duplicated issue." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5576/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5576
https://github.com/huggingface/datasets/issues/5576
false
1,598,396,552
https://api.github.com/repos/huggingface/datasets/issues/5575/labels{/name}
### Feature request Being able to put some metadata for each column as a string or any other type. ### Motivation I will bring the motivation by an example, lets say we are experimenting with embedding produced by some image encoder network, and we want to iterate through a couple of preprocessing and see which one works better in our downstream task, here as workaround right now what I do is the compute the hash of the preprocessing that the images went through as part of the new columns name, it would be nice to attach some kinda meta data in these scenarios to the each columns. metadata ### Your contribution Maybe we could map another relational like database as the metadata?
2024-01-05T21:48:35Z
5,575
{ "closed_at": null, "closed_issues": 1, "created_at": "2023-02-13T16:22:42Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }, "description": "Next major release", "due_on": null, "html_url": "https://github.com/huggingface/datasets/milestone/10", "id": 9038583, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/10/labels", "node_id": "MI_kwDODunzps4Aier3", "number": 10, "open_issues": 3, "state": "open", "title": "3.0", "updated_at": "2024-03-19T15:29:43Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/10" }
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-02-24T10:53:44Z
https://api.github.com/repos/huggingface/datasets/issues/5575/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5575/timeline
Metadata for each column
https://api.github.com/repos/huggingface/datasets/issues/5575/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/11356471?v=4", "events_url": "https://api.github.com/users/parsa-ra/events{/privacy}", "followers_url": "https://api.github.com/users/parsa-ra/followers", "following_url": "https://api.github.com/users/parsa-ra/following{/other_user}", "gists_url": "https://api.github.com/users/parsa-ra/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/parsa-ra", "id": 11356471, "login": "parsa-ra", "node_id": "MDQ6VXNlcjExMzU2NDcx", "organizations_url": "https://api.github.com/users/parsa-ra/orgs", "received_events_url": "https://api.github.com/users/parsa-ra/received_events", "repos_url": "https://api.github.com/users/parsa-ra/repos", "site_admin": false, "starred_url": "https://api.github.com/users/parsa-ra/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/parsa-ra/subscriptions", "type": "User", "url": "https://api.github.com/users/parsa-ra" }
[]
null
null
NONE
null
null
I_kwDODunzps5fRZiI
[ "Hi! Indeed it would be useful to support this. PyArrow natively supports schema-level and column-level metadata, so implementing this should be straightforward. The API I have in mind would work as follows:\r\n```python\r\ncol_feature = Value(\"string\", metadata=\"Some column-level metadata\")\r\n\r\nfeatures = F...
{ "+1": 4, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 4, "url": "https://api.github.com/repos/huggingface/datasets/issues/5575/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5575
https://github.com/huggingface/datasets/issues/5575
false
1,598,104,691
https://api.github.com/repos/huggingface/datasets/issues/5574/labels{/name}
### Describe the bug Loading the `c4` dataset in streaming mode with `load_dataset("c4", "en", split="validation", streaming=True)` and then using it fails with a `FileNotFoundException`. ### Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("c4", "en", split="train", streaming=True) next(iter(dataset)) ``` causes a ``` FileNotFoundError: https://huggingface.co/datasets/allenai/c4/resolve/1ddc917116b730e1859edef32896ec5c16be51d0/en/c4-train.00000-of-01024.json.gz ``` I can download this file manually though e.g. by entering this URL in a browser. There is an underlying HTTP 403 status code: ``` aiohttp.client_exceptions.ClientResponseError: 403, message='Forbidden', url=URL('https://cdn-lfs.huggingface.co/datasets/allenai/c4/8ef8d75b0e045dec4aa5123a671b4564466b0707086a7ed1ba8721626dfffbc9?response-content-disposition=attachment%3B+filename*%3DUTF-8''c4-train.00000-of-01024.json.gz%3B+filename%3D%22c4-train.00000-of-01024.json.gz%22%3B&response-content-type=application/gzip&Expires=1677483770&Policy=eyJTdGF0ZW1lbnQiOlt7IlJlc291cmNlIjoiaHR0cHM6Ly9jZG4tbGZzLmh1Z2dpbmdmYWNlLmNvL2RhdGFzZXRzL2FsbGVuYWkvYzQvOGVmOGQ3NWIwZTA0NWRlYzRhYTUxMjNhNjcxYjQ1NjQ0NjZiMDcwNzA4NmE3ZWQxYmE4NzIxNjI2ZGZmZmJjOT9yZXNwb25zZS1jb250ZW50LWRpc3Bvc2l0aW9uPSomcmVzcG9uc2UtY29udGVudC10eXBlPWFwcGxpY2F0aW9uJTJGZ3ppcCIsIkNvbmRpdGlvbiI6eyJEYXRlTGVzc1RoYW4iOnsiQVdTOkVwb2NoVGltZSI6MTY3NzQ4Mzc3MH19fV19&Signature=yjL3UeY72cf2xpnvPvD68eAYOEe2qtaUJV55sB-jnPskBJEMwpMJcBZvg2~GqXZdM3O-GWV-Z3CI~d4u5VCb4YZ-HlmOjr3VBYkvox2EKiXnBIhjMecf2UVUPtxhTa9kBVlWjqu4qKzB9gKXZF2Cwpp5ctLzapEaT2nnqF84RAL-rsqMA3I~M8vWWfivQsbBK63hMfgZqqKMgdWM0iKMaItveDl0ufQ29azMFmsR7qd8V7sU2Z-F1fAeohS8HpN9OOnClW34yi~YJ2AbgZJJBXA~qsylfVA0Qp7Q~yX~q4P8JF1vmJ2BjkiSbGrj3bAXOGugpOVU5msI52DT88yMdA__&Key-Pair-Id=KVTP0A1DKRTAX') ``` ### Expected behavior This should retrieve the first example from the C4 validation set. This worked a few days ago but stopped working now. ### Environment info - `datasets` version: 2.9.0 - Platform: Linux-5.15.0-60-generic-x86_64-with-glibc2.31 - Python version: 3.9.16 - PyArrow version: 11.0.0 - Pandas version: 1.5.3
2023-12-18T07:32:32Z
5,574
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-24T07:57:32Z
https://api.github.com/repos/huggingface/datasets/issues/5574/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5574/timeline
c4 dataset streaming fails with `FileNotFoundError`
https://api.github.com/repos/huggingface/datasets/issues/5574/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/202907?v=4", "events_url": "https://api.github.com/users/krasserm/events{/privacy}", "followers_url": "https://api.github.com/users/krasserm/followers", "following_url": "https://api.github.com/users/krasserm/following{/other_user}", "gists_url": "https://api.github.com/users/krasserm/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/krasserm", "id": 202907, "login": "krasserm", "node_id": "MDQ6VXNlcjIwMjkwNw==", "organizations_url": "https://api.github.com/users/krasserm/orgs", "received_events_url": "https://api.github.com/users/krasserm/received_events", "repos_url": "https://api.github.com/users/krasserm/repos", "site_admin": false, "starred_url": "https://api.github.com/users/krasserm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/krasserm/subscriptions", "type": "User", "url": "https://api.github.com/users/krasserm" }
[]
null
completed
NONE
2023-02-27T04:03:38Z
null
I_kwDODunzps5fQSRz
[ "Also encountering this issue for every dataset I try to stream! Installed datasets from main:\r\n```\r\n- `datasets` version: 2.10.1.dev0\r\n- Platform: macOS-13.1-arm64-arm-64bit\r\n- Python version: 3.9.13\r\n- PyArrow version: 10.0.1\r\n- Pandas version: 1.5.2\r\n```\r\n\r\nRepro:\r\n```python\r\nfrom datasets ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5574/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5574
https://github.com/huggingface/datasets/issues/5574
false
1,597,400,836
https://api.github.com/repos/huggingface/datasets/issues/5573/labels{/name}
I've removed `torchaudio` completely and switched to use `soundfile` for everything. With the new version of `soundfile` package this should work smoothly because the `libsndfile` C library is bundled, in Linux wheels too. Let me know if you think it's too harsh and we should continue to support `torchaudio` decoding. I decided that we can drop it completely because: 1. it's always something wrong with `torchaudio` (for example recently https://github.com/huggingface/datasets/issues/5488 ) 2. the results of mp3 decoding are different depending on `torchaudio` version 3. `soundfile` is slightly faster then the latest `torchaudio` 4. anyway users can pass any custom decoding function with any library they want if needed (worth putting a snippet in the docs). cc @sanchit-gandhi @vaibhavad
2023-02-28T20:25:14Z
5,573
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-23T19:19:44Z
https://api.github.com/repos/huggingface/datasets/issues/5573/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5573/timeline
Use soundfile for mp3 decoding instead of torchaudio
https://api.github.com/repos/huggingface/datasets/issues/5573/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
null
null
CONTRIBUTOR
2023-02-28T20:16:02Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5573.diff", "html_url": "https://github.com/huggingface/datasets/pull/5573", "merged_at": "2023-02-28T20:16:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/5573.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5573" }
PR_kwDODunzps5Kop7n
[ "_The documentation is not available anymore as the PR was closed or merged._", "@mariosasko thank you for the review! do you have any idea why `test_hash_torch_tensor` fails on \"ubuntu-latest deps-minimum\"? I removed the `torchaudio<0.12.0` test dependency so it uses the latest `torch` now, might it be connect...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 3, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/5573/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5573
https://github.com/huggingface/datasets/pull/5573
true
1,597,257,624
https://api.github.com/repos/huggingface/datasets/issues/5572/labels{/name}
### Describe the bug download_mode="reuse_dataset_if_exists" will always consider that a dataset doesn't exist. Specifically, upon losing an internet connection trying to load a dataset for a second time in ten seconds, a connection error results, showing a breakpoint of: ``` File ~/jupyterlab/.direnv/python-3.9.6/lib/python3.9/site-packages/datasets/load.py:1174, in dataset_module_factory(path, revision, download_config, download_mode, dynamic_modules_path, data_dir, data_files, **download_kwargs) 1165 except Exception as e: # noqa: catch any exception of hf_hub and consider that the dataset doesn't exist 1166 if isinstance( 1167 e, 1168 ( (...) 1172 ), 1173 ): -> 1174 raise ConnectionError(f"Couldn't reach '{path}' on the Hub ({type(e).__name__})") 1175 elif "404" in str(e): 1176 msg = f"Dataset '{path}' doesn't exist on the Hub" ConnectionError: Couldn't reach 'lsb/tenk' on the Hub (ConnectionError) ``` This has been around since at least v2.0. ### Steps to reproduce the bug ``` from datasets import load_dataset import numpy as np tenk = load_dataset("lsb/tenk") # ten thousand integers print(np.average(tenk['train']['a'])) # prints 4999.5 ### now disconnect your internet tenk_too = load_dataset("lsb/tenk", download_mode="reuse_dataset_if_exists") # Raises ConnectionError: Couldn't reach 'lsb/tenk' on the Hub (ConnectionError) ``` ### Expected behavior I expected that I would be able to reuse the dataset I just downloaded. ### Environment info - `datasets` version: 2.10.0 - Platform: macOS-13.1-arm64-arm-64bit - Python version: 3.9.6 - PyArrow version: 7.0.0 - Pandas version: 1.5.2
2023-02-23T18:03:55Z
5,572
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-23T17:28:11Z
https://api.github.com/repos/huggingface/datasets/issues/5572/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5572/timeline
Datasets 2.10.0 does not reuse the dataset cache
https://api.github.com/repos/huggingface/datasets/issues/5572/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/45281?v=4", "events_url": "https://api.github.com/users/lsb/events{/privacy}", "followers_url": "https://api.github.com/users/lsb/followers", "following_url": "https://api.github.com/users/lsb/following{/other_user}", "gists_url": "https://api.github.com/users/lsb/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lsb", "id": 45281, "login": "lsb", "node_id": "MDQ6VXNlcjQ1Mjgx", "organizations_url": "https://api.github.com/users/lsb/orgs", "received_events_url": "https://api.github.com/users/lsb/received_events", "repos_url": "https://api.github.com/users/lsb/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lsb/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lsb/subscriptions", "type": "User", "url": "https://api.github.com/users/lsb" }
[]
null
completed
NONE
2023-02-23T18:03:55Z
null
I_kwDODunzps5fNDeY
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5572/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5572
https://github.com/huggingface/datasets/issues/5572
false
1,597,198,953
https://api.github.com/repos/huggingface/datasets/issues/5571/labels{/name}
### Describe the bug Steps: 1. Created a dataset in a Linux VM and created a small sample using dataset.to_json() method. 2. Downloaded the JSON file to my local Windows machine for working and saved in say - r"C:\Users\name\file.json" 3. I am reading the file in my local PyCharm - the location of python file is different than the location of the JSON. 4. When I read using load_dataset("json",args.input_json), it throws and error from builder.py. raise InvalidConfigName( f"Bad characters from black list '{invalid_windows_characters}' found in '{self.name}'. " f"They could create issues when creating a directory for this config on Windows filesystem." 6. When I bring the data to the current directory, it works fine. ### Steps to reproduce the bug Steps: 1. Created a dataset in a Linux VM and created a small sample using dataset.to_json() method. 2. Downloaded the JSON file to my local Windows machine for working and saved in say - r"C:\Users\name\file.json" 3. I am reading the file in my local PyCharm - the location of python file is different than the location of the JSON. 4. When I read using load_dataset("json",args.input_json), it throws and error from builder.py. raise InvalidConfigName( f"Bad characters from black list '{invalid_windows_characters}' found in '{self.name}'. " f"They could create issues when creating a directory for this config on Windows filesystem." 6. When I bring the data to the current directory, it works fine. ### Expected behavior Should be able to read from a path different than current directory in Windows machine. ### Environment info datasets version: 2.3.1 python version: 3.8 Windows OS
2023-02-24T13:21:47Z
5,571
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-23T16:50:11Z
https://api.github.com/repos/huggingface/datasets/issues/5571/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5571/timeline
load_dataset fails for JSON in windows
https://api.github.com/repos/huggingface/datasets/issues/5571/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/11876897?v=4", "events_url": "https://api.github.com/users/abinashsahu/events{/privacy}", "followers_url": "https://api.github.com/users/abinashsahu/followers", "following_url": "https://api.github.com/users/abinashsahu/following{/other_user}", "gists_url": "https://api.github.com/users/abinashsahu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/abinashsahu", "id": 11876897, "login": "abinashsahu", "node_id": "MDQ6VXNlcjExODc2ODk3", "organizations_url": "https://api.github.com/users/abinashsahu/orgs", "received_events_url": "https://api.github.com/users/abinashsahu/received_events", "repos_url": "https://api.github.com/users/abinashsahu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/abinashsahu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/abinashsahu/subscriptions", "type": "User", "url": "https://api.github.com/users/abinashsahu" }
[]
null
completed
NONE
2023-02-24T13:21:47Z
null
I_kwDODunzps5fM1Jp
[ "Hi! \r\n\r\nYou need to pass an input json file explicitly as `data_files` to `load_dataset` to avoid this error:\r\n```python\r\n ds = load_dataset(\"json\", data_files=args.input_json)\r\n```\r\n\r\n", "Thanks it worked!" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5571/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5571
https://github.com/huggingface/datasets/issues/5571
false
1,597,190,926
https://api.github.com/repos/huggingface/datasets/issues/5570/labels{/name}
### Describe the bug When calling ```load_dataset('imagenet-1k')``` FileNotFoundError is raised, if not logged in and if logged in with huggingface-cli but not having accepted the licence on the hub. There is no error once accepting. ### Steps to reproduce the bug ``` from datasets import load_dataset imagenet = load_dataset("imagenet-1k", split="train", streaming=True) FileNotFoundError: Couldn't find a dataset script at /content/imagenet-1k/imagenet-1k.py or any data file in the same directory. Couldn't find 'imagenet-1k' on the Hugging Face Hub either: FileNotFoundError: Dataset 'imagenet-1k' doesn't exist on the Hub ``` tested on a colab notebook. ### Expected behavior I would expect a specific error indicating that I have to login then accept the dataset licence. I find this bug very relevant as this code is on a guide on the [Huggingface documentation for Datasets](https://huggingface.co/docs/datasets/about_mapstyle_vs_iterable) ### Environment info google colab cpu-only instance
2023-07-24T15:18:50Z
5,570
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-23T16:44:32Z
https://api.github.com/repos/huggingface/datasets/issues/5570/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5570/timeline
load_dataset gives FileNotFoundError on imagenet-1k if license is not accepted on the hub
https://api.github.com/repos/huggingface/datasets/issues/5570/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/38630200?v=4", "events_url": "https://api.github.com/users/buoi/events{/privacy}", "followers_url": "https://api.github.com/users/buoi/followers", "following_url": "https://api.github.com/users/buoi/following{/other_user}", "gists_url": "https://api.github.com/users/buoi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/buoi", "id": 38630200, "login": "buoi", "node_id": "MDQ6VXNlcjM4NjMwMjAw", "organizations_url": "https://api.github.com/users/buoi/orgs", "received_events_url": "https://api.github.com/users/buoi/received_events", "repos_url": "https://api.github.com/users/buoi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/buoi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/buoi/subscriptions", "type": "User", "url": "https://api.github.com/users/buoi" }
[]
null
completed
NONE
2023-07-24T15:18:50Z
null
I_kwDODunzps5fMzMO
[ "Hi, thanks for the feedback! Would it help to add a tip or note saying the dataset is gated and you need to accept the license before downloading it?", "The error is now more informative:\r\n```\r\nFileNotFoundError: Couldn't find a dataset script at /content/imagenet-1k/imagenet-1k.py or any data file in the sa...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5570/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5570
https://github.com/huggingface/datasets/issues/5570
false
1,597,132,383
https://api.github.com/repos/huggingface/datasets/issues/5569/labels{/name}
[5558](https://github.com/huggingface/datasets/issues/5568)
2023-02-24T14:06:37Z
5,569
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-23T16:06:04Z
https://api.github.com/repos/huggingface/datasets/issues/5569/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5569/timeline
pass the dataset features to the IterableDataset.from_generator function
https://api.github.com/repos/huggingface/datasets/issues/5569/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4", "events_url": "https://api.github.com/users/Hubert-Bonisseur/events{/privacy}", "followers_url": "https://api.github.com/users/Hubert-Bonisseur/followers", "following_url": "https://api.github.com/users/Hubert-Bonisseur/following{/other_user}", "gists_url": "https://api.github.com/users/Hubert-Bonisseur/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Hubert-Bonisseur", "id": 48770768, "login": "Hubert-Bonisseur", "node_id": "MDQ6VXNlcjQ4NzcwNzY4", "organizations_url": "https://api.github.com/users/Hubert-Bonisseur/orgs", "received_events_url": "https://api.github.com/users/Hubert-Bonisseur/received_events", "repos_url": "https://api.github.com/users/Hubert-Bonisseur/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Hubert-Bonisseur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hubert-Bonisseur/subscriptions", "type": "User", "url": "https://api.github.com/users/Hubert-Bonisseur" }
[]
null
null
CONTRIBUTOR
2023-02-23T18:15:16Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5569.diff", "html_url": "https://github.com/huggingface/datasets/pull/5569", "merged_at": "2023-02-23T18:15:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/5569.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5569" }
PR_kwDODunzps5KnwHD
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5569/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5569
https://github.com/huggingface/datasets/pull/5569
true
1,596,900,532
https://api.github.com/repos/huggingface/datasets/issues/5568/labels{/name}
### Describe the bug Hello, I like the new `to_iterable_dataset` feature but I noticed something that seems to be missing. When using `to_iterable_dataset` to transform your map style dataset into iterable dataset, you lose valuable metadata like the features. These metadata are useful if you want to interleave iterable datasets, cast columns etc. ### Steps to reproduce the bug ```python dataset = load_dataset("lhoestq/demo1")["train"] print(dataset.features) # {'id': Value(dtype='string', id=None), 'package_name': Value(dtype='string', id=None), 'review': Value(dtype='string', id=None), 'date': Value(dtype='string', id=None), 'star': Value(dtype='int64', id=None), 'version_id': Value(dtype='int64', id=None)} dataset = dataset.to_iterable_dataset() print(dataset.features) # None ``` ### Expected behavior Keep the relevant information ### Environment info datasets==2.10.0
2023-02-24T13:22:36Z
5,568
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "7057ff", "default": true...
2023-02-23T13:45:33Z
https://api.github.com/repos/huggingface/datasets/issues/5568/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4", "events_url": "https://api.github.com/users/Hubert-Bonisseur/events{/privacy}", "followers_url": "https://api.github.com/users/Hubert-Bonisseur/followers", "following_url": "https://api.github.com/users/Hubert-Bonisseur/following{/other_user}", "gists_url": "https://api.github.com/users/Hubert-Bonisseur/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Hubert-Bonisseur", "id": 48770768, "login": "Hubert-Bonisseur", "node_id": "MDQ6VXNlcjQ4NzcwNzY4", "organizations_url": "https://api.github.com/users/Hubert-Bonisseur/orgs", "received_events_url": "https://api.github.com/users/Hubert-Bonisseur/received_events", "repos_url": "https://api.github.com/users/Hubert-Bonisseur/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Hubert-Bonisseur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hubert-Bonisseur/subscriptions", "type": "User", "url": "https://api.github.com/users/Hubert-Bonisseur" }
https://api.github.com/repos/huggingface/datasets/issues/5568/timeline
dataset.to_iterable_dataset() loses useful info like dataset features
https://api.github.com/repos/huggingface/datasets/issues/5568/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4", "events_url": "https://api.github.com/users/Hubert-Bonisseur/events{/privacy}", "followers_url": "https://api.github.com/users/Hubert-Bonisseur/followers", "following_url": "https://api.github.com/users/Hubert-Bonisseur/following{/other_user}", "gists_url": "https://api.github.com/users/Hubert-Bonisseur/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Hubert-Bonisseur", "id": 48770768, "login": "Hubert-Bonisseur", "node_id": "MDQ6VXNlcjQ4NzcwNzY4", "organizations_url": "https://api.github.com/users/Hubert-Bonisseur/orgs", "received_events_url": "https://api.github.com/users/Hubert-Bonisseur/received_events", "repos_url": "https://api.github.com/users/Hubert-Bonisseur/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Hubert-Bonisseur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hubert-Bonisseur/subscriptions", "type": "User", "url": "https://api.github.com/users/Hubert-Bonisseur" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/48770768?v=4", "events_url": "https://api.github.com/users/Hubert-Bonisseur/events{/privacy}", "followers_url": "https://api.github.com/users/Hubert-Bonisseur/followers", "following_url": "https://api.github.com/users/Hubert-Bonisseur/followin...
null
completed
CONTRIBUTOR
2023-02-24T13:22:36Z
null
I_kwDODunzps5fLsS0
[ "Hi ! Oh good catch. I think the features should be passed to `IterableDataset.from_generator()` in `to_iterable_dataset()` indeed.\r\n\r\nSetting this as a good first issue if someone would like to contribute, otherwise we can take care of it :)", "#self-assign", "seems like the feature parameter is missing fr...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5568/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5568
https://github.com/huggingface/datasets/issues/5568
false
1,595,916,674
https://api.github.com/repos/huggingface/datasets/issues/5566/labels{/name}
### Feature request Right now, we have to read the get the parquet file to the local storage. So having ability to read given the bucket directly address would be benificial ### Motivation In a production set up, this feature can help us a lot. So we do not need move training datafiles in between storage. ### Your contribution I am willing to help if there's anyway.
2023-02-23T11:03:29Z
5,566
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists", "id": 1935892865, "name": "duplicate", "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate" }, { "color": "a2eeef", ...
2023-02-22T22:13:40Z
https://api.github.com/repos/huggingface/datasets/issues/5566/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5566/timeline
Directly reading parquet files in a s3 bucket from the load_dataset method
https://api.github.com/repos/huggingface/datasets/issues/5566/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16892570?v=4", "events_url": "https://api.github.com/users/shamanez/events{/privacy}", "followers_url": "https://api.github.com/users/shamanez/followers", "following_url": "https://api.github.com/users/shamanez/following{/other_user}", "gists_url": "https://api.github.com/users/shamanez/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/shamanez", "id": 16892570, "login": "shamanez", "node_id": "MDQ6VXNlcjE2ODkyNTcw", "organizations_url": "https://api.github.com/users/shamanez/orgs", "received_events_url": "https://api.github.com/users/shamanez/received_events", "repos_url": "https://api.github.com/users/shamanez/repos", "site_admin": false, "starred_url": "https://api.github.com/users/shamanez/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shamanez/subscriptions", "type": "User", "url": "https://api.github.com/users/shamanez" }
[]
null
null
NONE
null
null
I_kwDODunzps5fH8GC
[ "Hi ! I think is in the scope of this other issue: to https://github.com/huggingface/datasets/issues/5281 " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5566/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5566
https://github.com/huggingface/datasets/issues/5566
false
1,595,281,752
https://api.github.com/repos/huggingface/datasets/issues/5565/labels{/name}
This way we can control the size of the record_batches/row_groups of arrow/parquet files. This can be useful for `datasets-server` to keep control of the row groups size which can affect random access performance for audio/image/video datasets Right now having 1,000 examples per row group cause some image datasets to be pretty slow for random access (e.g. 4seconds for `beans` to get 20 rows)
2023-03-10T13:53:03Z
5,565
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-22T15:09:30Z
https://api.github.com/repos/huggingface/datasets/issues/5565/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5565/timeline
Add writer_batch_size for ArrowBasedBuilder
https://api.github.com/repos/huggingface/datasets/issues/5565/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-03-10T13:45:43Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5565.diff", "html_url": "https://github.com/huggingface/datasets/pull/5565", "merged_at": "2023-03-10T13:45:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/5565.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5565" }
PR_kwDODunzps5KhfTH
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5565/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5565
https://github.com/huggingface/datasets/pull/5565
true
1,595,064,698
https://api.github.com/repos/huggingface/datasets/issues/5564/labels{/name}
null
2023-02-22T13:09:26Z
5,564
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-22T13:00:09Z
https://api.github.com/repos/huggingface/datasets/issues/5564/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5564/timeline
Set dev version
https://api.github.com/repos/huggingface/datasets/issues/5564/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-22T13:00:25Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5564.diff", "html_url": "https://github.com/huggingface/datasets/pull/5564", "merged_at": "2023-02-22T13:00:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/5564.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5564" }
PR_kwDODunzps5KgwzU
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5564). All of your documentation changes will be reflected on that endpoint.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5564/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5564
https://github.com/huggingface/datasets/pull/5564
true
1,595,049,025
https://api.github.com/repos/huggingface/datasets/issues/5563/labels{/name}
null
2023-02-22T13:05:55Z
5,563
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-22T12:48:52Z
https://api.github.com/repos/huggingface/datasets/issues/5563/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5563/timeline
Release: 2.10.0
https://api.github.com/repos/huggingface/datasets/issues/5563/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-22T12:56:48Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5563.diff", "html_url": "https://github.com/huggingface/datasets/pull/5563", "merged_at": "2023-02-22T12:56:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/5563.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5563" }
PR_kwDODunzps5KgtbL
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5563/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5563
https://github.com/huggingface/datasets/pull/5563
true
1,594,625,539
https://api.github.com/repos/huggingface/datasets/issues/5562/labels{/name}
Removed mangle_dup_cols=True from BuilderConfig. It triggered following deprecation warning: /usr/local/lib/python3.8/dist-packages/datasets/download/streaming_download_manager.py:776: FutureWarning: the 'mangle_dupe_cols' keyword is deprecated and will be removed in a future version. Please take steps to stop the use of 'mangle_dupe_cols' return pd.read_csv(xopen(filepath_or_buffer, "rb", use_auth_token=use_auth_token), **kwargs) Further documentation of pandas: https://pandas.pydata.org/docs/whatsnew/v1.4.0.html#mangle-dupe-cols-in-read-csv-no-longer-renames-unique-columns-conflicting-with-target-names At first sight it seems like this flag is resolved internally, it might need some more research.
2023-02-23T11:07:49Z
5,562
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-22T07:56:10Z
https://api.github.com/repos/huggingface/datasets/issues/5562/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5562/timeline
Update csv.py
https://api.github.com/repos/huggingface/datasets/issues/5562/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/54279069?v=4", "events_url": "https://api.github.com/users/XDoubleU/events{/privacy}", "followers_url": "https://api.github.com/users/XDoubleU/followers", "following_url": "https://api.github.com/users/XDoubleU/following{/other_user}", "gists_url": "https://api.github.com/users/XDoubleU/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/XDoubleU", "id": 54279069, "login": "XDoubleU", "node_id": "MDQ6VXNlcjU0Mjc5MDY5", "organizations_url": "https://api.github.com/users/XDoubleU/orgs", "received_events_url": "https://api.github.com/users/XDoubleU/received_events", "repos_url": "https://api.github.com/users/XDoubleU/repos", "site_admin": false, "starred_url": "https://api.github.com/users/XDoubleU/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/XDoubleU/subscriptions", "type": "User", "url": "https://api.github.com/users/XDoubleU" }
[]
null
null
CONTRIBUTOR
2023-02-23T11:00:58Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5562.diff", "html_url": "https://github.com/huggingface/datasets/pull/5562", "merged_at": "2023-02-23T11:00:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/5562.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5562" }
PR_kwDODunzps5KfTUT
[ "_The documentation is not available anymore as the PR was closed or merged._", "Removed it :)", "Changed it :)", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_format...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5562/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5562
https://github.com/huggingface/datasets/pull/5562
true
1,593,862,388
https://api.github.com/repos/huggingface/datasets/issues/5561/labels{/name}
@huggingface/datasets do you think it would be useful? Motivation - sometimes PRs are like 30% "fix: style" commits :) If so - I need to double check the config but for me locally it works as expected.
2023-02-28T15:37:22Z
5,561
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-21T17:35:07Z
https://api.github.com/repos/huggingface/datasets/issues/5561/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5561/timeline
Add pre-commit config yaml file to enable automatic code formatting
https://api.github.com/repos/huggingface/datasets/issues/5561/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
null
null
CONTRIBUTOR
2023-02-23T18:23:29Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5561.diff", "html_url": "https://github.com/huggingface/datasets/pull/5561", "merged_at": "2023-02-23T18:23:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/5561.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5561" }
PR_kwDODunzps5Kcxw_
[ "_The documentation is not available anymore as the PR was closed or merged._", "Better yet have someone enable pre-commit CI https://pre-commit.ci/ and it will apply the pre-commit fixes to the PR automatically as an additional commit.", "@Skylion007 hi! I agree with @nateraw here, I'd better not force to use ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5561/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5561
https://github.com/huggingface/datasets/pull/5561
true
1,593,809,978
https://api.github.com/repos/huggingface/datasets/issues/5560/labels{/name}
This PR modifies `map` to: * ensure the TQDM bar gets the last progress update * when a map function fails, avoid throwing a chained exception in the single-proc mode
2023-02-21T18:26:23Z
5,560
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-21T16:56:17Z
https://api.github.com/repos/huggingface/datasets/issues/5560/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5560/timeline
Ensure last tqdm update in `map`
https://api.github.com/repos/huggingface/datasets/issues/5560/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2023-02-21T18:19:09Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5560.diff", "html_url": "https://github.com/huggingface/datasets/pull/5560", "merged_at": "2023-02-21T18:19:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/5560.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5560" }
PR_kwDODunzps5Kcml6
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5560/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5560
https://github.com/huggingface/datasets/pull/5560
true
1,593,676,489
https://api.github.com/repos/huggingface/datasets/issues/5559/labels{/name}
#5455 introduced a small bug that lead `map` to ignore the `suffix_template` argument and not put suffixes to cached files in multiprocessing. I fixed this and also improved a few things: - regarding logging: "Loading cached processed dataset" is now logged only once even in multiprocessing (it used to be logged `num_proc` times) - regarding new_fingerprint: I made sure that the returned dataset satisfies `ds._fingerprint==new_fingerprint` if `new_fingerprint` is passed to `map`
2023-02-21T17:21:37Z
5,559
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-21T15:26:26Z
https://api.github.com/repos/huggingface/datasets/issues/5559/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5559/timeline
Fix map suffix_template
https://api.github.com/repos/huggingface/datasets/issues/5559/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-21T17:14:29Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5559.diff", "html_url": "https://github.com/huggingface/datasets/pull/5559", "merged_at": "2023-02-21T17:14:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/5559.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5559" }
PR_kwDODunzps5KcKSb
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/5559/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5559
https://github.com/huggingface/datasets/pull/5559
true
1,593,655,815
https://api.github.com/repos/huggingface/datasets/issues/5558/labels{/name}
Colab now has Ubuntu 20.04 which already has `ffmpeg` of required (>4) version.
2023-03-01T13:46:04Z
5,558
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-21T15:13:36Z
https://api.github.com/repos/huggingface/datasets/issues/5558/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5558/timeline
Remove instructions for `ffmpeg` system package installation on Colab
https://api.github.com/repos/huggingface/datasets/issues/5558/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
null
null
CONTRIBUTOR
2023-02-23T13:50:27Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5558.diff", "html_url": "https://github.com/huggingface/datasets/pull/5558", "merged_at": "2023-02-23T13:50:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/5558.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5558" }
PR_kwDODunzps5KcF5E
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5558/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5558
https://github.com/huggingface/datasets/pull/5558
true
1,593,545,324
https://api.github.com/repos/huggingface/datasets/issues/5557/labels{/name}
Otherwise it would show a `Map` progress bar, since it uses `map` under the hood
2023-02-21T14:19:54Z
5,557
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-21T14:04:42Z
https://api.github.com/repos/huggingface/datasets/issues/5557/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5557/timeline
Add filter desc
https://api.github.com/repos/huggingface/datasets/issues/5557/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-21T14:12:39Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5557.diff", "html_url": "https://github.com/huggingface/datasets/pull/5557", "merged_at": "2023-02-21T14:12:39Z", "patch_url": "https://github.com/huggingface/datasets/pull/5557.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5557" }
PR_kwDODunzps5Kbube
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5557/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5557
https://github.com/huggingface/datasets/pull/5557
true
1,593,246,936
https://api.github.com/repos/huggingface/datasets/issues/5556/labels{/name}
...instead of relying on the optional librosa dependency `resampy`. It was only used for `_decode_non_mp3_file_like` anyway and not for the other ones - removing it fixes consistency between decoding methods (except torchaudio decoding) Therefore I think it is a better solution than adding `resampy` as a dependency in https://github.com/huggingface/datasets/pull/5554 cc @polinaeterna
2023-02-21T12:49:50Z
5,556
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-21T10:45:50Z
https://api.github.com/repos/huggingface/datasets/issues/5556/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5556/timeline
Use default audio resampling type
https://api.github.com/repos/huggingface/datasets/issues/5556/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-21T12:42:52Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5556.diff", "html_url": "https://github.com/huggingface/datasets/pull/5556", "merged_at": "2023-02-21T12:42:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/5556.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5556" }
PR_kwDODunzps5KauVL
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5556/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5556
https://github.com/huggingface/datasets/pull/5556
true
1,592,469,938
https://api.github.com/repos/huggingface/datasets/issues/5555/labels{/name}
### Describe the bug ``` --------------------------------------------------------------------------- ValueError Traceback (most recent call last) Cell In [16], line 1 ----> 1 train_dataset = train_dataset.shuffle() File /opt/conda/envs/pytorch/lib/python3.9/site-packages/datasets/arrow_dataset.py:551, in transmit_format.<locals>.wrapper(*args, **kwargs) 544 self_format = { 545 "type": self._format_type, 546 "format_kwargs": self._format_kwargs, 547 "columns": self._format_columns, 548 "output_all_columns": self._output_all_columns, 549 } 550 # apply actual function --> 551 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 552 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 553 # re-apply format to the output File /opt/conda/envs/pytorch/lib/python3.9/site-packages/datasets/fingerprint.py:480, in fingerprint_transform.<locals>._fingerprint.<locals>.wrapper(*args, **kwargs) 476 validate_fingerprint(kwargs[fingerprint_name]) 478 # Call actual function --> 480 out = func(self, *args, **kwargs) 482 # Update fingerprint of in-place transforms + update in-place history of transforms 484 if inplace: # update after calling func so that the fingerprint doesn't change if the function fails File /opt/conda/envs/pytorch/lib/python3.9/site-packages/datasets/arrow_dataset.py:3616, in Dataset.shuffle(self, seed, generator, keep_in_memory, load_from_cache_file, indices_cache_file_name, writer_batch_size, new_fingerprint) 3610 return self._new_dataset_with_indices( 3611 fingerprint=new_fingerprint, indices_cache_file_name=indices_cache_file_name 3612 ) 3614 permutation = generator.permutation(len(self)) -> 3616 return self.select( 3617 indices=permutation, 3618 keep_in_memory=keep_in_memory, 3619 indices_cache_file_name=indices_cache_file_name if not keep_in_memory else None, 3620 writer_batch_size=writer_batch_size, 3621 new_fingerprint=new_fingerprint, 3622 ) File /opt/conda/envs/pytorch/lib/python3.9/site-packages/datasets/arrow_dataset.py:551, in transmit_format.<locals>.wrapper(*args, **kwargs) 544 self_format = { 545 "type": self._format_type, 546 "format_kwargs": self._format_kwargs, 547 "columns": self._format_columns, 548 "output_all_columns": self._output_all_columns, 549 } 550 # apply actual function --> 551 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 552 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 553 # re-apply format to the output File /opt/conda/envs/pytorch/lib/python3.9/site-packages/datasets/fingerprint.py:480, in fingerprint_transform.<locals>._fingerprint.<locals>.wrapper(*args, **kwargs) 476 validate_fingerprint(kwargs[fingerprint_name]) 478 # Call actual function --> 480 out = func(self, *args, **kwargs) 482 # Update fingerprint of in-place transforms + update in-place history of transforms 484 if inplace: # update after calling func so that the fingerprint doesn't change if the function fails File /opt/conda/envs/pytorch/lib/python3.9/site-packages/datasets/arrow_dataset.py:3266, in Dataset.select(self, indices, keep_in_memory, indices_cache_file_name, writer_batch_size, new_fingerprint) 3263 return self._select_contiguous(start, length, new_fingerprint=new_fingerprint) 3265 # If not contiguous, we need to create a new indices mapping -> 3266 return self._select_with_indices_mapping( 3267 indices, 3268 keep_in_memory=keep_in_memory, 3269 indices_cache_file_name=indices_cache_file_name, 3270 writer_batch_size=writer_batch_size, 3271 new_fingerprint=new_fingerprint, 3272 ) File /opt/conda/envs/pytorch/lib/python3.9/site-packages/datasets/arrow_dataset.py:551, in transmit_format.<locals>.wrapper(*args, **kwargs) 544 self_format = { 545 "type": self._format_type, 546 "format_kwargs": self._format_kwargs, 547 "columns": self._format_columns, 548 "output_all_columns": self._output_all_columns, 549 } 550 # apply actual function --> 551 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 552 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 553 # re-apply format to the output File /opt/conda/envs/pytorch/lib/python3.9/site-packages/datasets/fingerprint.py:480, in fingerprint_transform.<locals>._fingerprint.<locals>.wrapper(*args, **kwargs) 476 validate_fingerprint(kwargs[fingerprint_name]) 478 # Call actual function --> 480 out = func(self, *args, **kwargs) 482 # Update fingerprint of in-place transforms + update in-place history of transforms 484 if inplace: # update after calling func so that the fingerprint doesn't change if the function fails File /opt/conda/envs/pytorch/lib/python3.9/site-packages/datasets/arrow_dataset.py:3389, in Dataset._select_with_indices_mapping(self, indices, keep_in_memory, indices_cache_file_name, writer_batch_size, new_fingerprint) 3387 logger.info(f"Caching indices mapping at {indices_cache_file_name}") 3388 tmp_file = tempfile.NamedTemporaryFile("wb", dir=os.path.dirname(indices_cache_file_name), delete=False) -> 3389 writer = ArrowWriter( 3390 path=tmp_file.name, writer_batch_size=writer_batch_size, fingerprint=new_fingerprint, unit="indices" 3391 ) 3393 indices = indices if isinstance(indices, list) else list(indices) 3395 size = len(self) File /opt/conda/envs/pytorch/lib/python3.9/site-packages/datasets/arrow_writer.py:315, in ArrowWriter.__init__(self, schema, features, path, stream, fingerprint, writer_batch_size, hash_salt, check_duplicates, disable_nullable, update_features, with_metadata, unit, embed_local_files, storage_options) 312 self._disable_nullable = disable_nullable 314 if stream is None: --> 315 fs_token_paths = fsspec.get_fs_token_paths(path, storage_options=storage_options) 316 self._fs: fsspec.AbstractFileSystem = fs_token_paths[0] 317 self._path = ( 318 fs_token_paths[2][0] 319 if not is_remote_filesystem(self._fs) 320 else self._fs.unstrip_protocol(fs_token_paths[2][0]) 321 ) File /opt/conda/envs/pytorch/lib/python3.9/site-packages/fsspec/core.py:593, in get_fs_token_paths(urlpath, mode, num, name_function, storage_options, protocol, expand) 591 else: 592 urlpath = stringify_path(urlpath) --> 593 chain = _un_chain(urlpath, storage_options or {}) 594 if len(chain) > 1: 595 inkwargs = {} File /opt/conda/envs/pytorch/lib/python3.9/site-packages/fsspec/core.py:330, in _un_chain(path, kwargs) 328 for bit in reversed(bits): 329 protocol = split_protocol(bit)[0] or "file" --> 330 cls = get_filesystem_class(protocol) 331 extra_kwargs = cls._get_kwargs_from_urls(bit) 332 kws = kwargs.get(protocol, {}) File /opt/conda/envs/pytorch/lib/python3.9/site-packages/fsspec/registry.py:240, in get_filesystem_class(protocol) 238 if protocol not in registry: 239 if protocol not in known_implementations: --> 240 raise ValueError("Protocol not known: %s" % protocol) 241 bit = known_implementations[protocol] 242 try: ValueError: Protocol not known: parent ``` This is what the `train_dataset` object looks like ``` Dataset({ features: ['label', 'input_ids', 'attention_mask'], num_rows: 364166 }) ``` ### Steps to reproduce the bug The `train_dataset` obj is created by concatenating two datasets And then shuffle is called, but it throws the mentioned error. ### Expected behavior Should shuffle the dataset properly. ### Environment info - `datasets` version: 2.6.1 - Platform: Linux-5.15.0-1022-aws-x86_64-with-glibc2.31 - Python version: 3.9.13 - PyArrow version: 10.0.0 - Pandas version: 1.4.4
2023-02-27T09:23:34Z
5,555
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-20T21:33:45Z
https://api.github.com/repos/huggingface/datasets/issues/5555/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5555/timeline
`.shuffle` throwing error `ValueError: Protocol not known: parent`
https://api.github.com/repos/huggingface/datasets/issues/5555/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/10768588?v=4", "events_url": "https://api.github.com/users/prabhakar267/events{/privacy}", "followers_url": "https://api.github.com/users/prabhakar267/followers", "following_url": "https://api.github.com/users/prabhakar267/following{/other_user}", "gists_url": "https://api.github.com/users/prabhakar267/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/prabhakar267", "id": 10768588, "login": "prabhakar267", "node_id": "MDQ6VXNlcjEwNzY4NTg4", "organizations_url": "https://api.github.com/users/prabhakar267/orgs", "received_events_url": "https://api.github.com/users/prabhakar267/received_events", "repos_url": "https://api.github.com/users/prabhakar267/repos", "site_admin": false, "starred_url": "https://api.github.com/users/prabhakar267/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/prabhakar267/subscriptions", "type": "User", "url": "https://api.github.com/users/prabhakar267" }
[]
null
null
NONE
null
null
I_kwDODunzps5e6ymy
[ "Hi ! The indices mapping is written in the same cachedirectory as your dataset.\r\n\r\nCan you run this to show your current cache directory ?\r\n```python\r\nprint(train_dataset.cache_files)\r\n```", "```\r\n[{'filename': '.../train/dataset.arrow'}, {'filename': '.../train/dataset.arrow'}]\r\n```\r\n\r\nThese a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5555/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5555
https://github.com/huggingface/datasets/issues/5555
false
1,592,285,062
https://api.github.com/repos/huggingface/datasets/issues/5554/labels{/name}
In librosa 0.10 they removed the `resmpy` dependency and set it to optional. However it is necessary for resampling. I added it to the "audio" extra dependencies.
2023-09-24T10:07:29Z
5,554
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-20T18:15:43Z
https://api.github.com/repos/huggingface/datasets/issues/5554/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5554/timeline
Add resampy dep
https://api.github.com/repos/huggingface/datasets/issues/5554/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-21T12:43:38Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5554.diff", "html_url": "https://github.com/huggingface/datasets/pull/5554", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5554.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5554" }
PR_kwDODunzps5KXhZh
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5554/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5554
https://github.com/huggingface/datasets/pull/5554
true
1,592,236,998
https://api.github.com/repos/huggingface/datasets/issues/5553/labels{/name}
Solves #5539
2023-02-21T13:08:25Z
5,553
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-20T17:29:14Z
https://api.github.com/repos/huggingface/datasets/issues/5553/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5553/timeline
improved message error row formatting
https://api.github.com/repos/huggingface/datasets/issues/5553/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/26489385?v=4", "events_url": "https://api.github.com/users/Plutone11011/events{/privacy}", "followers_url": "https://api.github.com/users/Plutone11011/followers", "following_url": "https://api.github.com/users/Plutone11011/following{/other_user}", "gists_url": "https://api.github.com/users/Plutone11011/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Plutone11011", "id": 26489385, "login": "Plutone11011", "node_id": "MDQ6VXNlcjI2NDg5Mzg1", "organizations_url": "https://api.github.com/users/Plutone11011/orgs", "received_events_url": "https://api.github.com/users/Plutone11011/received_events", "repos_url": "https://api.github.com/users/Plutone11011/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Plutone11011/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Plutone11011/subscriptions", "type": "User", "url": "https://api.github.com/users/Plutone11011" }
[]
null
null
CONTRIBUTOR
2023-02-21T12:58:12Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5553.diff", "html_url": "https://github.com/huggingface/datasets/pull/5553", "merged_at": "2023-02-21T12:58:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/5553.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5553" }
PR_kwDODunzps5KXXUq
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5553/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5553
https://github.com/huggingface/datasets/pull/5553
true
1,592,186,703
https://api.github.com/repos/huggingface/datasets/issues/5552/labels{/name}
Fix for https://discord.com/channels/879548962464493619/1075729627546406912/1075729627546406912
2023-02-21T13:20:42Z
5,552
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-20T16:50:09Z
https://api.github.com/repos/huggingface/datasets/issues/5552/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5552/timeline
Make tiktoken tokenizers hashable
https://api.github.com/repos/huggingface/datasets/issues/5552/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2023-02-21T13:13:05Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5552.diff", "html_url": "https://github.com/huggingface/datasets/pull/5552", "merged_at": "2023-02-21T13:13:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/5552.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5552" }
PR_kwDODunzps5KXMjA
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5552/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5552
https://github.com/huggingface/datasets/pull/5552
true
1,592,140,836
https://api.github.com/repos/huggingface/datasets/issues/5551/labels{/name}
This is kinda unimportant fix but, the suggested `pip install sklearn` does not work. The current error message if sklearn is not installed: ``` ImportError: To be able to use [dataset name], you need to install the following dependency: sklearn. Please install it using 'pip install sklearn' for instance. ```
2023-02-21T13:27:57Z
5,551
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-20T16:16:57Z
https://api.github.com/repos/huggingface/datasets/issues/5551/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5551/timeline
Suggest scikit-learn instead of sklearn
https://api.github.com/repos/huggingface/datasets/issues/5551/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/74963545?v=4", "events_url": "https://api.github.com/users/osbm/events{/privacy}", "followers_url": "https://api.github.com/users/osbm/followers", "following_url": "https://api.github.com/users/osbm/following{/other_user}", "gists_url": "https://api.github.com/users/osbm/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/osbm", "id": 74963545, "login": "osbm", "node_id": "MDQ6VXNlcjc0OTYzNTQ1", "organizations_url": "https://api.github.com/users/osbm/orgs", "received_events_url": "https://api.github.com/users/osbm/received_events", "repos_url": "https://api.github.com/users/osbm/repos", "site_admin": false, "starred_url": "https://api.github.com/users/osbm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/osbm/subscriptions", "type": "User", "url": "https://api.github.com/users/osbm" }
[]
null
null
CONTRIBUTOR
2023-02-21T13:21:07Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5551.diff", "html_url": "https://github.com/huggingface/datasets/pull/5551", "merged_at": "2023-02-21T13:21:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/5551.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5551" }
PR_kwDODunzps5KXCof
[ "good catch!", "_The documentation is not available anymore as the PR was closed or merged._", "The test fail is unrelated to this PR and fixed on `main` - merging :)", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: ...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5551/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5551
https://github.com/huggingface/datasets/pull/5551
true
1,591,409,475
https://api.github.com/repos/huggingface/datasets/issues/5550/labels{/name}
Hello! ## Pull Request overview * Resolve 4 broken references in the docs ## The problems Two broken references [here](https://huggingface.co/docs/datasets/package_reference/main_classes#datasets.Dataset.class_encode_column): ![image](https://user-images.githubusercontent.com/37621491/220056232-366b64dc-33c9-461b-8f82-1ac4aa570280.png) --- One broken reference [here](https://huggingface.co/docs/datasets/package_reference/main_classes#datasets.Dataset.unique): ![image](https://user-images.githubusercontent.com/37621491/220057135-2f249d60-c01d-48b5-82bb-5085a7635198.png) --- One missing reference [here](https://huggingface.co/docs/datasets/v2.9.0/en/package_reference/main_classes#datasets.DatasetDict.class_encode_column): ![image](https://user-images.githubusercontent.com/37621491/220057025-4a8e5556-5041-4ec7-b8d8-ed4fdc266495.png) - Tom Aarsen
2023-02-20T15:16:13Z
5,550
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-20T08:52:11Z
https://api.github.com/repos/huggingface/datasets/issues/5550/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5550/timeline
Resolve four broken refs in the docs
https://api.github.com/repos/huggingface/datasets/issues/5550/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/37621491?v=4", "events_url": "https://api.github.com/users/tomaarsen/events{/privacy}", "followers_url": "https://api.github.com/users/tomaarsen/followers", "following_url": "https://api.github.com/users/tomaarsen/following{/other_user}", "gists_url": "https://api.github.com/users/tomaarsen/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/tomaarsen", "id": 37621491, "login": "tomaarsen", "node_id": "MDQ6VXNlcjM3NjIxNDkx", "organizations_url": "https://api.github.com/users/tomaarsen/orgs", "received_events_url": "https://api.github.com/users/tomaarsen/received_events", "repos_url": "https://api.github.com/users/tomaarsen/repos", "site_admin": false, "starred_url": "https://api.github.com/users/tomaarsen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tomaarsen/subscriptions", "type": "User", "url": "https://api.github.com/users/tomaarsen" }
[]
null
null
MEMBER
2023-02-20T15:09:13Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5550.diff", "html_url": "https://github.com/huggingface/datasets/pull/5550", "merged_at": "2023-02-20T15:09:13Z", "patch_url": "https://github.com/huggingface/datasets/pull/5550.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5550" }
PR_kwDODunzps5KUl5i
[ "_The documentation is not available anymore as the PR was closed or merged._", "See the resolved changes [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5550/en/package_reference/main_classes#datasets.Dataset.class_encode_column), [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5550/en/pa...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5550/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5550
https://github.com/huggingface/datasets/pull/5550
true
1,590,836,848
https://api.github.com/repos/huggingface/datasets/issues/5549/labels{/name}
Fix #5548 Apply ruff's flake8-comprehension checks for better performance, and more readable code.
2023-02-23T14:06:39Z
5,549
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-19T20:09:28Z
https://api.github.com/repos/huggingface/datasets/issues/5549/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5549/timeline
Apply ruff flake8-comprehension checks
https://api.github.com/repos/huggingface/datasets/issues/5549/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/2053727?v=4", "events_url": "https://api.github.com/users/Skylion007/events{/privacy}", "followers_url": "https://api.github.com/users/Skylion007/followers", "following_url": "https://api.github.com/users/Skylion007/following{/other_user}", "gists_url": "https://api.github.com/users/Skylion007/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Skylion007", "id": 2053727, "login": "Skylion007", "node_id": "MDQ6VXNlcjIwNTM3Mjc=", "organizations_url": "https://api.github.com/users/Skylion007/orgs", "received_events_url": "https://api.github.com/users/Skylion007/received_events", "repos_url": "https://api.github.com/users/Skylion007/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Skylion007/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Skylion007/subscriptions", "type": "User", "url": "https://api.github.com/users/Skylion007" }
[]
null
null
CONTRIBUTOR
2023-02-23T13:59:39Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5549.diff", "html_url": "https://github.com/huggingface/datasets/pull/5549", "merged_at": "2023-02-23T13:59:39Z", "patch_url": "https://github.com/huggingface/datasets/pull/5549.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5549" }
PR_kwDODunzps5KSsi3
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5549/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5549
https://github.com/huggingface/datasets/pull/5549
true
1,590,835,479
https://api.github.com/repos/huggingface/datasets/issues/5548/labels{/name}
### Feature request Apply ruff flake8 comprehension checks to codebase. ### Motivation This should strictly improve the performance / readability of the codebase by removing unnecessary iteration, function calls, etc. This should generate better Python bytecode which should strictly improve performance. I already applied this fixes to PyTorch and Sympy with little issue and have opened PRs to diffusers and transformers todo this as well. ### Your contribution Making a PR.
2023-02-23T13:59:41Z
5,548
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-02-19T20:05:38Z
https://api.github.com/repos/huggingface/datasets/issues/5548/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5548/timeline
Apply flake8-comprehensions to codebase
https://api.github.com/repos/huggingface/datasets/issues/5548/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/2053727?v=4", "events_url": "https://api.github.com/users/Skylion007/events{/privacy}", "followers_url": "https://api.github.com/users/Skylion007/followers", "following_url": "https://api.github.com/users/Skylion007/following{/other_user}", "gists_url": "https://api.github.com/users/Skylion007/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Skylion007", "id": 2053727, "login": "Skylion007", "node_id": "MDQ6VXNlcjIwNTM3Mjc=", "organizations_url": "https://api.github.com/users/Skylion007/orgs", "received_events_url": "https://api.github.com/users/Skylion007/received_events", "repos_url": "https://api.github.com/users/Skylion007/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Skylion007/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Skylion007/subscriptions", "type": "User", "url": "https://api.github.com/users/Skylion007" }
[]
null
completed
CONTRIBUTOR
2023-02-23T13:59:41Z
null
I_kwDODunzps5e0jkX
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5548/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5548
https://github.com/huggingface/datasets/issues/5548
false
1,590,468,200
https://api.github.com/repos/huggingface/datasets/issues/5547/labels{/name}
## What's in this PR? After exploring for a while the JAX integration in 🤗`datasets`, I found out that, even though JAX prioritizes the TPU and GPU as the default device when available, the `JaxFormatter` doesn't let you specify the device where you want to place the `jax.Array`s in case you don't want to rely on JAX's default array placement. So on, I've included the `device` param in `JaxFormatter` but there are some things to take into consideration: * A formatted `Dataset` is copied with `copy.deepcopy` which means that if one adds the param `device` in `JaxFormatter` as a `jaxlib.xla_extension.Device`, it "fails" because that object cannot be serialized (instead of serializing the param adds a random hash instead). That's the reason why I added a function `_map_devices_to_str` to basically create a mapping of strings to `jaxlib.xla_extension.Device`s so that `self.device` is a string and not a `jaxlib.xla_extension.Device`. * To create a `jax.Array` in a device you need to either create it in the default device and then move it to the desired device with `jax.device_put` or directly create it in the device you want with `jax.default_device()` context manager. * JAX will create an array by default in `jax.devices()[0]` More information on JAX device management is available at https://jax.readthedocs.io/en/latest/faq.html#controlling-data-and-computation-placement-on-devices ## What's missing in this PR? I've tested it both locally in CPU (Mac M2 and Mac M1, as no GPU support for Mac yet), and in GPU and TPU in Google Colab, let me know if you want me to provide you the Notebook for the latter. But I did not implement any integration test as I wanted to get your feedback first.
2023-02-21T16:10:55Z
5,547
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-18T20:57:40Z
https://api.github.com/repos/huggingface/datasets/issues/5547/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5547/timeline
Add JAX device selection when formatting
https://api.github.com/repos/huggingface/datasets/issues/5547/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
[]
null
null
CONTRIBUTOR
2023-02-21T16:04:03Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5547.diff", "html_url": "https://github.com/huggingface/datasets/pull/5547", "merged_at": "2023-02-21T16:04:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/5547.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5547" }
PR_kwDODunzps5KRmcf
[ "The code below was throwing a warning:\r\n\r\n```python\r\nclass JaxFormatter(Formatter[Mapping, \"jax.Array\", Mapping]):\r\n def __init__(self, features=None, device=None, **jnp_array_kwargs):\r\n super().__init__(features=features)\r\n import jax\r\n from jaxlib.xla_extension import Devi...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5547/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5547
https://github.com/huggingface/datasets/pull/5547
true
1,590,346,349
https://api.github.com/repos/huggingface/datasets/issues/5546/labels{/name}
### Describe the bug In the huggingface course (https://huggingface.co/course/chapter3/2?fw=pt) it said that if we set HF_HOME, downloaded datasets would be cached at specified address but it does not. downloaded models from checkpoint names are downloaded and cached at HF_HOME but this is not the case for datasets, they are still cached at ~/.cache/huggingface/datasets. ### Steps to reproduce the bug Run the following code ``` from datasets import load_dataset raw_datasets = load_dataset("glue", "mrpc") raw_datasets ``` it downloads and store dataset at ~/.cache/huggingface/datasets ### Expected behavior to cache dataset at HF_HOME. ### Environment info python 3.10.6 Kubuntu 22.04 HF_HOME located on a separate partition
2023-07-24T14:22:43Z
5,546
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-18T13:30:35Z
https://api.github.com/repos/huggingface/datasets/issues/5546/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5546/timeline
Downloaded datasets do not cache at $HF_HOME
https://api.github.com/repos/huggingface/datasets/issues/5546/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/79091831?v=4", "events_url": "https://api.github.com/users/ErfanMoosaviMonazzah/events{/privacy}", "followers_url": "https://api.github.com/users/ErfanMoosaviMonazzah/followers", "following_url": "https://api.github.com/users/ErfanMoosaviMonazzah/following{/other_user}", "gists_url": "https://api.github.com/users/ErfanMoosaviMonazzah/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ErfanMoosaviMonazzah", "id": 79091831, "login": "ErfanMoosaviMonazzah", "node_id": "MDQ6VXNlcjc5MDkxODMx", "organizations_url": "https://api.github.com/users/ErfanMoosaviMonazzah/orgs", "received_events_url": "https://api.github.com/users/ErfanMoosaviMonazzah/received_events", "repos_url": "https://api.github.com/users/ErfanMoosaviMonazzah/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ErfanMoosaviMonazzah/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ErfanMoosaviMonazzah/subscriptions", "type": "User", "url": "https://api.github.com/users/ErfanMoosaviMonazzah" }
[]
null
completed
NONE
2023-07-24T14:22:43Z
null
I_kwDODunzps5eysJt
[ "Hi ! Can you make sure you set `HF_HOME` before importing `datasets` ?\r\n\r\nThen you can print\r\n```python\r\nprint(datasets.config.HF_CACHE_HOME)\r\nprint(datasets.config.HF_DATASETS_CACHE)\r\n```" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5546/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5546
https://github.com/huggingface/datasets/issues/5546
false
1,590,315,972
https://api.github.com/repos/huggingface/datasets/issues/5545/labels{/name}
Hi, I was missing the ability to easily open the pushed dataset and it seemed like a quick fix. Maybe we also want to log this info somewhere, but let me know if I need to add that too. Cheers, David
2023-12-18T16:57:56Z
5,545
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-18T11:26:25Z
https://api.github.com/repos/huggingface/datasets/issues/5545/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5545/timeline
Added return methods for URL-references to the pushed dataset
https://api.github.com/repos/huggingface/datasets/issues/5545/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/25269220?v=4", "events_url": "https://api.github.com/users/davidberenstein1957/events{/privacy}", "followers_url": "https://api.github.com/users/davidberenstein1957/followers", "following_url": "https://api.github.com/users/davidberenstein1957/following{/other_user}", "gists_url": "https://api.github.com/users/davidberenstein1957/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/davidberenstein1957", "id": 25269220, "login": "davidberenstein1957", "node_id": "MDQ6VXNlcjI1MjY5MjIw", "organizations_url": "https://api.github.com/users/davidberenstein1957/orgs", "received_events_url": "https://api.github.com/users/davidberenstein1957/received_events", "repos_url": "https://api.github.com/users/davidberenstein1957/repos", "site_admin": false, "starred_url": "https://api.github.com/users/davidberenstein1957/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/davidberenstein1957/subscriptions", "type": "User", "url": "https://api.github.com/users/davidberenstein1957" }
[]
null
null
NONE
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/5545.diff", "html_url": "https://github.com/huggingface/datasets/pull/5545", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5545.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5545" }
PR_kwDODunzps5KRKct
[ "Hi ! Maybe we'd need to align with `transformers` and other libraries that implement `push_to_hub` to agree on what it should return.\r\n\r\ne.g. in `transformers` the typing says it returns a string, but in practice it returns a `CommitInfo`.\r\n\r\nTherefore I'd not add an output to `push_to_hub` here unless we ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5545/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5545
https://github.com/huggingface/datasets/pull/5545
true
1,588,951,379
https://api.github.com/repos/huggingface/datasets/issues/5543/labels{/name}
### Describe the bug in #3627, the host url of the pile dataset became `https://mystic.the-eye.eu`. Now the new url is broken, but `https://the-eye.eu` seems to work again. ### Steps to reproduce the bug ```python3 from datasets import load_dataset dataset = load_dataset("bookcorpusopen") ``` shows ```python3 ConnectionError: Couldn't reach https://mystic.the-eye.eu/public/AI/pile_preliminary_components/books1.tar.gz (ProxyError(MaxRetryError("HTTPSConnectionPool(host='mystic.the-eye.eu', port=443): Max retries exceeded with url: /public/AI/pile_pr eliminary_components/books1.tar.gz (Caused by ProxyError('Cannot connect to proxy.', OSError('Tunnel connection failed: 504 Gateway Timeout')))"))) ``` ### Expected behavior Downloading as normal. ### Environment info - `datasets` version: 2.9.0 - Platform: Linux-5.4.143.bsk.7-amd64-x86_64-with-glibc2.31 - Python version: 3.9.2 - PyArrow version: 6.0.1 - Pandas version: 1.5.3
2023-02-21T06:37:00Z
5,543
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-17T08:40:11Z
https://api.github.com/repos/huggingface/datasets/issues/5543/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5543/timeline
the pile datasets url seems to change back
https://api.github.com/repos/huggingface/datasets/issues/5543/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/5126316?v=4", "events_url": "https://api.github.com/users/wjfwzzc/events{/privacy}", "followers_url": "https://api.github.com/users/wjfwzzc/followers", "following_url": "https://api.github.com/users/wjfwzzc/following{/other_user}", "gists_url": "https://api.github.com/users/wjfwzzc/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/wjfwzzc", "id": 5126316, "login": "wjfwzzc", "node_id": "MDQ6VXNlcjUxMjYzMTY=", "organizations_url": "https://api.github.com/users/wjfwzzc/orgs", "received_events_url": "https://api.github.com/users/wjfwzzc/received_events", "repos_url": "https://api.github.com/users/wjfwzzc/repos", "site_admin": false, "starred_url": "https://api.github.com/users/wjfwzzc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wjfwzzc/subscriptions", "type": "User", "url": "https://api.github.com/users/wjfwzzc" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
NONE
2023-02-20T08:41:33Z
null
I_kwDODunzps5etXlT
[ "Thanks for reporting, @wjfwzzc.\r\n\r\nI am transferring this issue to the corresponding dataset on the Hub: https://huggingface.co/datasets/bookcorpusopen/discussions/1", "Thank you. All fixes are done:\r\n- [x] https://huggingface.co/datasets/bookcorpusopen/discussions/2\r\n- [x] https://huggingface.co/dataset...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5543/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5543
https://github.com/huggingface/datasets/issues/5543
false
1,588,633,724
https://api.github.com/repos/huggingface/datasets/issues/5542/labels{/name}
Fixes https://github.com/huggingface/datasets/issues/5541
2023-02-17T19:20:49Z
5,542
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-17T01:52:38Z
https://api.github.com/repos/huggingface/datasets/issues/5542/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5542/timeline
Avoid saving sparse ChunkedArrays in pyarrow tables
https://api.github.com/repos/huggingface/datasets/issues/5542/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6591505?v=4", "events_url": "https://api.github.com/users/marioga/events{/privacy}", "followers_url": "https://api.github.com/users/marioga/followers", "following_url": "https://api.github.com/users/marioga/following{/other_user}", "gists_url": "https://api.github.com/users/marioga/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/marioga", "id": 6591505, "login": "marioga", "node_id": "MDQ6VXNlcjY1OTE1MDU=", "organizations_url": "https://api.github.com/users/marioga/orgs", "received_events_url": "https://api.github.com/users/marioga/received_events", "repos_url": "https://api.github.com/users/marioga/repos", "site_admin": false, "starred_url": "https://api.github.com/users/marioga/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/marioga/subscriptions", "type": "User", "url": "https://api.github.com/users/marioga" }
[]
null
null
CONTRIBUTOR
2023-02-17T11:12:32Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5542.diff", "html_url": "https://github.com/huggingface/datasets/pull/5542", "merged_at": "2023-02-17T11:12:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/5542.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5542" }
PR_kwDODunzps5KLjMl
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5542/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5542
https://github.com/huggingface/datasets/pull/5542
true
1,588,633,555
https://api.github.com/repos/huggingface/datasets/issues/5541/labels{/name}
### Describe the bug If we perform a `select` (or `shuffle`, `train_test_split`, etc.) operation on a dataset , we end up with a dataset with an `indices_table`. Currently, flattening such dataset consumes a lot of memory and the resulting flat dataset contains ChunkedArrays with as many chunks as there are rows. This is extremely inefficient and slows down the operations on the flat dataset, e.g., saving/loading the dataset to disk becomes really slow. Perhaps more importantly, loading the dataset back from disk basically loads the whole table into RAM, as it cannot take advantage of memory mapping. ### Steps to reproduce the bug The following script reproduces the issue: ```python import gc import os import psutil import tempfile import time from datasets import Dataset DATASET_SIZE = 5000000 def profile(func): def wrapper(*args, **kwargs): mem_before = psutil.Process(os.getpid()).memory_info().rss / (1024 * 1024) start = time.time() # Run function here out = func(*args, **kwargs) end = time.time() mem_after = psutil.Process(os.getpid()).memory_info().rss / (1024 * 1024) print(f"{func.__name__} -- RAM memory used: {mem_after - mem_before} MB -- Total time: {end - start:.6f} s") return out return wrapper def main(): ds = Dataset.from_list([{'col': i} for i in range(DATASET_SIZE)]) print(f"Num chunks for original ds: {ds.data['col'].num_chunks}") with tempfile.TemporaryDirectory() as tmpdir: path1 = os.path.join(tmpdir, 'ds1') print("Original ds save/load") profile(ds.save_to_disk)(path1) ds_loaded = profile(Dataset.load_from_disk)(path1) print(f"Num chunks for original ds after reloading: {ds_loaded.data['col'].num_chunks}") print("") ds_select = ds.select(reversed(range(len(ds)))) print(f"Num chunks for selected ds: {ds_select.data['col'].num_chunks}") del ds del ds_loaded gc.collect() # This would happen anyway when we call save_to_disk ds_select = profile(ds_select.flatten_indices)() print(f"Num chunks for selected ds after flattening: {ds_select.data['col'].num_chunks}") print("") path2 = os.path.join(tmpdir, 'ds2') print("Selected ds save/load") profile(ds_select.save_to_disk)(path2) del ds_select gc.collect() ds_select_loaded = profile(Dataset.load_from_disk)(path2) print(f"Num chunks for selected ds after reloading: {ds_select_loaded.data['col'].num_chunks}") if __name__ == '__main__': main() ``` Sample result: ``` Num chunks for original ds: 1 Original ds save/load save_to_disk -- RAM memory used: 0.515625 MB -- Total time: 0.253888 s load_from_disk -- RAM memory used: 42.765625 MB -- Total time: 0.015176 s Num chunks for original ds after reloading: 5000 Num chunks for selected ds: 1 flatten_indices -- RAM memory used: 4852.609375 MB -- Total time: 46.116774 s Num chunks for selected ds after flattening: 5000000 Selected ds save/load save_to_disk -- RAM memory used: 1326.65625 MB -- Total time: 42.309825 s load_from_disk -- RAM memory used: 2085.953125 MB -- Total time: 11.659137 s Num chunks for selected ds after reloading: 5000000 ``` ### Expected behavior Saving/loading the dataset should be much faster and consume almost no extra memory thanks to pyarrow memory mapping. ### Environment info - `datasets` version: 2.9.1.dev0 - Platform: macOS-13.1-arm64-arm-64bit - Python version: 3.10.8 - PyArrow version: 11.0.0 - Pandas version: 1.5.3
2023-02-22T13:15:20Z
5,541
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-17T01:52:24Z
https://api.github.com/repos/huggingface/datasets/issues/5541/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5541/timeline
Flattening indices in selected datasets is extremely inefficient
https://api.github.com/repos/huggingface/datasets/issues/5541/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6591505?v=4", "events_url": "https://api.github.com/users/marioga/events{/privacy}", "followers_url": "https://api.github.com/users/marioga/followers", "following_url": "https://api.github.com/users/marioga/following{/other_user}", "gists_url": "https://api.github.com/users/marioga/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/marioga", "id": 6591505, "login": "marioga", "node_id": "MDQ6VXNlcjY1OTE1MDU=", "organizations_url": "https://api.github.com/users/marioga/orgs", "received_events_url": "https://api.github.com/users/marioga/received_events", "repos_url": "https://api.github.com/users/marioga/repos", "site_admin": false, "starred_url": "https://api.github.com/users/marioga/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/marioga/subscriptions", "type": "User", "url": "https://api.github.com/users/marioga" }
[]
null
completed
CONTRIBUTOR
2023-02-17T11:12:33Z
null
I_kwDODunzps5esJ_T
[ "Running the script above on the branch https://github.com/huggingface/datasets/pull/5542 results in the expected behaviour:\r\n```\r\nNum chunks for original ds: 1\r\nOriginal ds save/load\r\nsave_to_disk -- RAM memory used: 0.671875 MB -- Total time: 0.255265 s\r\nload_from_disk -- RAM memory used: 42.796875 MB -...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5541/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5541
https://github.com/huggingface/datasets/issues/5541
false
1,588,438,344
https://api.github.com/repos/huggingface/datasets/issues/5540/labels{/name}
A tutorial for creating datasets based on the folder-based builders and `from_dict` and `from_generator` methods. I've also mentioned loading scripts as a next step, but I think we should keep the tutorial focused on the low-code methods. Let me know what you think! 🙂
2023-02-17T18:50:46Z
5,540
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-16T22:09:35Z
https://api.github.com/repos/huggingface/datasets/issues/5540/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5540/timeline
Tutorial for creating a dataset
https://api.github.com/repos/huggingface/datasets/issues/5540/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
null
null
MEMBER
2023-02-17T18:41:28Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5540.diff", "html_url": "https://github.com/huggingface/datasets/pull/5540", "merged_at": "2023-02-17T18:41:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/5540.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5540" }
PR_kwDODunzps5KK5qz
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5540/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5540
https://github.com/huggingface/datasets/pull/5540
true
1,587,970,083
https://api.github.com/repos/huggingface/datasets/issues/5539/labels{/name}
### Describe the bug When dataset contains a 0-dim tensor, formatting.py raises a following error and fails. ```bash Traceback (most recent call last): File "<path>/lib/python3.8/site-packages/datasets/formatting/formatting.py", line 501, in format_row return _unnest(formatted_batch) File "<path>/lib/python3.8/site-packages/datasets/formatting/formatting.py", line 137, in _unnest return {key: array[0] for key, array in py_dict.items()} File "<path>/lib/python3.8/site-packages/datasets/formatting/formatting.py", line 137, in <dictcomp> return {key: array[0] for key, array in py_dict.items()} IndexError: invalid index of a 0-dim tensor. Use `tensor.item()` in Python or `tensor.item<T>()` in C++ to convert a 0-dim tensor to a number ``` ### Steps to reproduce the bug Load whichever dataset and add transform method to add 0-dim tensor. Or create/find a dataset containing 0-dim tensor. E.g. ```python from datasets import load_dataset import torch dataset = load_dataset("lambdalabs/pokemon-blip-captions", split='train') def t(batch): return {"test": torch.tensor(1)} dataset.set_transform(t) d_0 = dataset[0] ``` ### Expected behavior Extractor will correctly get a row from the dataset, even if it contains 0-dim tensor. ### Environment info `datasets==2.8.0`, but it looks like it is also applicable to main branch version (as of 16th February)
2023-02-22T10:30:30Z
5,539
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "7057ff", "default": true, "description": "Good for newcomers", "id": 1935892877, "name": "good first issue", "node_id": "MDU6TGFiZWwxOTM1ODkyODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20first%20issue" } ]
2023-02-16T16:08:51Z
https://api.github.com/repos/huggingface/datasets/issues/5539/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5539/timeline
IndexError: invalid index of a 0-dim tensor. Use `tensor.item()` in Python or `tensor.item<T>()` in C++ to convert a 0-dim tensor to a number
https://api.github.com/repos/huggingface/datasets/issues/5539/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/41912135?v=4", "events_url": "https://api.github.com/users/aalbersk/events{/privacy}", "followers_url": "https://api.github.com/users/aalbersk/followers", "following_url": "https://api.github.com/users/aalbersk/following{/other_user}", "gists_url": "https://api.github.com/users/aalbersk/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/aalbersk", "id": 41912135, "login": "aalbersk", "node_id": "MDQ6VXNlcjQxOTEyMTM1", "organizations_url": "https://api.github.com/users/aalbersk/orgs", "received_events_url": "https://api.github.com/users/aalbersk/received_events", "repos_url": "https://api.github.com/users/aalbersk/repos", "site_admin": false, "starred_url": "https://api.github.com/users/aalbersk/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/aalbersk/subscriptions", "type": "User", "url": "https://api.github.com/users/aalbersk" }
[]
null
completed
NONE
2023-02-21T13:03:57Z
null
I_kwDODunzps5epoAj
[ "Hi! The `set_transform` does not apply a custom formatting transform on a single example but the entire batch, so the fixed version of your transform would look as follows:\r\n```python\r\nfrom datasets import load_dataset\r\nimport torch\r\n\r\ndataset = load_dataset(\"lambdalabs/pokemon-blip-captions\", split='t...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5539/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5539
https://github.com/huggingface/datasets/issues/5539
false
1,587,732,596
https://api.github.com/repos/huggingface/datasets/issues/5538/labels{/name}
TimeoutError Traceback (most recent call last) ~\anaconda3\lib\urllib\request.py in do_open(self, http_class, req, **http_conn_args) 1345 try: -> 1346 h.request(req.get_method(), req.selector, req.data, headers, 1347 encode_chunked=req.has_header('Transfer-encoding')) ~\anaconda3\lib\http\client.py in request(self, method, url, body, headers, encode_chunked) 1278 """Send a complete request to the server.""" -> 1279 self._send_request(method, url, body, headers, encode_chunked) 1280 ~\anaconda3\lib\http\client.py in _send_request(self, method, url, body, headers, encode_chunked) 1324 body = _encode(body, 'body') -> 1325 self.endheaders(body, encode_chunked=encode_chunked) 1326 ~\anaconda3\lib\http\client.py in endheaders(self, message_body, encode_chunked) 1273 raise CannotSendHeader() -> 1274 self._send_output(message_body, encode_chunked=encode_chunked) 1275 ~\anaconda3\lib\http\client.py in _send_output(self, message_body, encode_chunked) 1033 del self._buffer[:] -> 1034 self.send(msg) 1035 ~\anaconda3\lib\http\client.py in send(self, data) 973 if self.auto_open: --> 974 self.connect() 975 else: ~\anaconda3\lib\http\client.py in connect(self) 1440 -> 1441 super().connect() 1442 ~\anaconda3\lib\http\client.py in connect(self) 944 """Connect to the host and port specified in __init__.""" --> 945 self.sock = self._create_connection( 946 (self.host,self.port), self.timeout, self.source_address) ~\anaconda3\lib\socket.py in create_connection(address, timeout, source_address) 843 try: --> 844 raise err 845 finally: ~\anaconda3\lib\socket.py in create_connection(address, timeout, source_address) 831 sock.bind(source_address) --> 832 sock.connect(sa) 833 # Break explicitly a reference cycle TimeoutError: [WinError 10060] A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond During handling of the above exception, another exception occurred: URLError Traceback (most recent call last) ~\AppData\Local\Temp/ipykernel_12220/2927704185.py in <module> 1 import seaborn as sn ----> 2 iris = sn.load_dataset('iris') ~\anaconda3\lib\site-packages\seaborn\utils.py in load_dataset(name, cache, data_home, **kws) 594 if name not in get_dataset_names(): 595 raise ValueError(f"'{name}' is not one of the example datasets.") --> 596 urlretrieve(url, cache_path) 597 full_path = cache_path 598 else: ~\anaconda3\lib\urllib\request.py in urlretrieve(url, filename, reporthook, data) 237 url_type, path = _splittype(url) 238 --> 239 with contextlib.closing(urlopen(url, data)) as fp: 240 headers = fp.info() 241 ~\anaconda3\lib\urllib\request.py in urlopen(url, data, timeout, cafile, capath, cadefault, context) 212 else: 213 opener = _opener --> 214 return opener.open(url, data, timeout) 215 216 def install_opener(opener): ~\anaconda3\lib\urllib\request.py in open(self, fullurl, data, timeout) 515 516 sys.audit('urllib.Request', req.full_url, req.data, req.headers, req.get_method()) --> 517 response = self._open(req, data) 518 519 # post-process response ~\anaconda3\lib\urllib\request.py in _open(self, req, data) 532 533 protocol = req.type --> 534 result = self._call_chain(self.handle_open, protocol, protocol + 535 '_open', req) 536 if result: ~\anaconda3\lib\urllib\request.py in _call_chain(self, chain, kind, meth_name, *args) 492 for handler in handlers: 493 func = getattr(handler, meth_name) --> 494 result = func(*args) 495 if result is not None: 496 return result ~\anaconda3\lib\urllib\request.py in https_open(self, req) 1387 1388 def https_open(self, req): -> 1389 return self.do_open(http.client.HTTPSConnection, req, 1390 context=self._context, check_hostname=self._check_hostname) 1391 ~\anaconda3\lib\urllib\request.py in do_open(self, http_class, req, **http_conn_args) 1347 encode_chunked=req.has_header('Transfer-encoding')) 1348 except OSError as err: # timeout error -> 1349 raise URLError(err) 1350 r = h.getresponse() 1351 except: URLError: <urlopen error [WinError 10060] A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond>
2023-02-16T14:44:36Z
5,538
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-16T14:01:58Z
https://api.github.com/repos/huggingface/datasets/issues/5538/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5538/timeline
load_dataset in seaborn is not working for me. getting this error.
https://api.github.com/repos/huggingface/datasets/issues/5538/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/125575109?v=4", "events_url": "https://api.github.com/users/reemaranibarik/events{/privacy}", "followers_url": "https://api.github.com/users/reemaranibarik/followers", "following_url": "https://api.github.com/users/reemaranibarik/following{/other_user}", "gists_url": "https://api.github.com/users/reemaranibarik/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/reemaranibarik", "id": 125575109, "login": "reemaranibarik", "node_id": "U_kgDOB3wfxQ", "organizations_url": "https://api.github.com/users/reemaranibarik/orgs", "received_events_url": "https://api.github.com/users/reemaranibarik/received_events", "repos_url": "https://api.github.com/users/reemaranibarik/repos", "site_admin": false, "starred_url": "https://api.github.com/users/reemaranibarik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/reemaranibarik/subscriptions", "type": "User", "url": "https://api.github.com/users/reemaranibarik" }
[]
null
completed
NONE
2023-02-16T14:44:36Z
null
I_kwDODunzps5eouB0
[ "Hi! `seaborn`'s `load_dataset` pulls datasets from [here](https://github.com/mwaskom/seaborn-data) and not from our Hub, so this issue is not related to our library in any way and should be reported in their repo instead." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5538/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5538
https://github.com/huggingface/datasets/issues/5538
false
1,587,567,464
https://api.github.com/repos/huggingface/datasets/issues/5537/labels{/name}
Certain datasets like `bigcode/the-stack-dedup` have so many files that loading them takes forever right from the data files resolution step. `datasets` uses file patterns to check the structure of the repository but it takes too much time to iterate over and over again on all the data files. This comes from `resolve_patterns_in_dataset_repository` which calls `_resolve_single_pattern_in_dataset_repository`, which iterates on all the files at ```python glob_iter = [PurePath(filepath) for filepath in fs.glob(PurePath(pattern).as_posix()) if fs.isfile(filepath)] ``` but calling `glob` on such a dataset is too expensive. Indeed it calls `ls()` in `hffilesystem.py` too many times. Maybe `glob` can be more optimized in `hffilesystem.py`, or the data files resolution can directly be implemented in the filesystem by checking its `dir_cache` ?
2023-12-15T13:12:31Z
5,537
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "BDE59C", "default": fals...
2023-02-16T12:11:45Z
https://api.github.com/repos/huggingface/datasets/issues/5537/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/35013374?v=4", "events_url": "https://api.github.com/users/semajyllek/events{/privacy}", "followers_url": "https://api.github.com/users/semajyllek/followers", "following_url": "https://api.github.com/users/semajyllek/following{/other_user}", "gists_url": "https://api.github.com/users/semajyllek/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/semajyllek", "id": 35013374, "login": "semajyllek", "node_id": "MDQ6VXNlcjM1MDEzMzc0", "organizations_url": "https://api.github.com/users/semajyllek/orgs", "received_events_url": "https://api.github.com/users/semajyllek/received_events", "repos_url": "https://api.github.com/users/semajyllek/repos", "site_admin": false, "starred_url": "https://api.github.com/users/semajyllek/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/semajyllek/subscriptions", "type": "User", "url": "https://api.github.com/users/semajyllek" }
https://api.github.com/repos/huggingface/datasets/issues/5537/timeline
Increase speed of data files resolution
https://api.github.com/repos/huggingface/datasets/issues/5537/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/35013374?v=4", "events_url": "https://api.github.com/users/semajyllek/events{/privacy}", "followers_url": "https://api.github.com/users/semajyllek/followers", "following_url": "https://api.github.com/users/semajyllek/following{/other_user}", ...
null
completed
MEMBER
2023-12-15T13:12:31Z
null
I_kwDODunzps5eoFto
[ "#self-assign", "You were right, if `self.dir_cache` is not None in glob, it is exactly the same as what is returned by find, at least for all the tests we have, and some extended evaluation I did across a random sample of about 1000 datasets. \r\n\r\nThanks for the nice hints, and let me know if this is not exac...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5537/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5537
https://github.com/huggingface/datasets/issues/5537
false
1,586,930,643
https://api.github.com/repos/huggingface/datasets/issues/5536/labels{/name}
### Describe the bug _Parameter 'function'=<function process at 0x7f1ec4388af0> of the transform datasets.arrow_dataset.Dataset.\_map_single couldn't be hashed properly, a random hash was used instead. Make sure your transforms and parameters are serializable with pickle or dill for the dataset fingerprinting and caching to work. If you reuse this transform, the caching mechanism will consider it to be different from the previous calls and recompute everything. This warning is only showed once. Subsequent hashing failures won't be showed._ This issue with `.map()` happens for me consistently, as also described in closed issue #4506 Dataset indices can be individually serialized using dill and pickle without any errors. I'm using tiktoken to encode in the function passed to map(). Similarly, indices can be individually encoded without error. ### Steps to reproduce the bug ```py from datasets import load_dataset import tiktoken dataset = load_dataset("stas/openwebtext-10k") enc = tiktoken.get_encoding("gpt2") tokenized = dataset.map( process, remove_columns=['text'], desc="tokenizing the OWT splits", ) def process(example): ids = enc.encode(example['text']) ids.append(enc.eot_token) out = {'ids': ids, 'len': len(ids)} return out ``` ### Expected behavior Should encode simple text objects. ### Environment info Python versions tried: both 3.8 and 3.10.10 `PYTHONUTF8=1` as env variable Datasets tried: - stas/openwebtext-10k - rotten_tomatoes - local text file OS: Ubuntu Linux 20.04 Package versions: - torch 1.13.1 - dill 0.3.4 (if using 0.3.6 - same issue) - datasets 2.9.0 - tiktoken 0.2.0
2023-09-08T21:06:01Z
5,536
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-16T03:12:07Z
https://api.github.com/repos/huggingface/datasets/issues/5536/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5536/timeline
Failure to hash function when using .map()
https://api.github.com/repos/huggingface/datasets/issues/5536/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6916056?v=4", "events_url": "https://api.github.com/users/venzen/events{/privacy}", "followers_url": "https://api.github.com/users/venzen/followers", "following_url": "https://api.github.com/users/venzen/following{/other_user}", "gists_url": "https://api.github.com/users/venzen/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/venzen", "id": 6916056, "login": "venzen", "node_id": "MDQ6VXNlcjY5MTYwNTY=", "organizations_url": "https://api.github.com/users/venzen/orgs", "received_events_url": "https://api.github.com/users/venzen/received_events", "repos_url": "https://api.github.com/users/venzen/repos", "site_admin": false, "starred_url": "https://api.github.com/users/venzen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/venzen/subscriptions", "type": "User", "url": "https://api.github.com/users/venzen" }
[]
null
completed
NONE
2023-02-16T14:56:41Z
null
I_kwDODunzps5elqPT
[ "Hi ! `enc` is not hashable:\r\n```python\r\nimport tiktoken\r\nfrom datasets.fingerprint import Hasher\r\n\r\nenc = tiktoken.get_encoding(\"gpt2\")\r\nHasher.hash(enc)\r\n# raises TypeError: cannot pickle 'builtins.CoreBPE' object\r\n```\r\nIt happens because it's not picklable, and because of that it's not possib...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5536/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5536
https://github.com/huggingface/datasets/issues/5536
false
1,586,520,369
https://api.github.com/repos/huggingface/datasets/issues/5535/labels{/name}
## What's in this PR? As a follow-up of #5522, I've created this entry in the documentation to explain how to use `.with_format("jax")` and why is it useful. @lhoestq Feel free to drop any feedback and/or suggestion, as probably more useful features can be included there!
2023-02-20T10:39:42Z
5,535
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-15T20:35:11Z
https://api.github.com/repos/huggingface/datasets/issues/5535/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5535/timeline
Add JAX-formatting documentation
https://api.github.com/repos/huggingface/datasets/issues/5535/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
[]
null
null
CONTRIBUTOR
2023-02-20T10:32:39Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5535.diff", "html_url": "https://github.com/huggingface/datasets/pull/5535", "merged_at": "2023-02-20T10:32:39Z", "patch_url": "https://github.com/huggingface/datasets/pull/5535.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5535" }
PR_kwDODunzps5KEb5L
[ "_The documentation is not available anymore as the PR was closed or merged._", "> Awesome thank you !\r\n> \r\n> Could you also explain how to use certain types like ClassLabel, Image or Audio with jax ? You can get a lot of inspiration from the \"Other feature types\" section in the [PyTorch page](https://huggi...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5535/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5535
https://github.com/huggingface/datasets/pull/5535
true
1,586,177,862
https://api.github.com/repos/huggingface/datasets/issues/5534/labels{/name}
### Describe the bug `map()` magically breaks when using a `Array3D` feature and mapping it. I created a very simple dummy dataset (see below). When filtering it down to 95 elements I can apply map, but it breaks when filtering it down to just 96 entries with the following exception: ``` Traceback (most recent call last): File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3255, in _map_single writer.finalize() # close_stream=bool(buf_writer is None)) # We only close if we are writing in a file File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/arrow_writer.py", line 581, in finalize self.write_examples_on_file() File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/arrow_writer.py", line 440, in write_examples_on_file batch_examples[col] = array_concat(arrays) File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/table.py", line 1931, in array_concat return _concat_arrays(arrays) File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/table.py", line 1901, in _concat_arrays return array_type.wrap_array(_concat_arrays([array.storage for array in arrays])) File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/table.py", line 1922, in _concat_arrays _concat_arrays([array.values for array in arrays]), File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/table.py", line 1922, in _concat_arrays _concat_arrays([array.values for array in arrays]), File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/table.py", line 1920, in _concat_arrays return pa.ListArray.from_arrays( File "pyarrow/array.pxi", line 1997, in pyarrow.lib.ListArray.from_arrays File "pyarrow/array.pxi", line 1527, in pyarrow.lib.Array.validate File "pyarrow/error.pxi", line 100, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: Negative offsets in list array During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 2815, in map return self._map_single( File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 546, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 513, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/fingerprint.py", line 480, in wrapper out = func(self, *args, **kwargs) File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 3259, in _map_single writer.finalize() File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/arrow_writer.py", line 581, in finalize self.write_examples_on_file() File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/arrow_writer.py", line 440, in write_examples_on_file batch_examples[col] = array_concat(arrays) File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/table.py", line 1931, in array_concat return _concat_arrays(arrays) File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/table.py", line 1901, in _concat_arrays return array_type.wrap_array(_concat_arrays([array.storage for array in arrays])) File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/table.py", line 1922, in _concat_arrays _concat_arrays([array.values for array in arrays]), File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/table.py", line 1922, in _concat_arrays _concat_arrays([array.values for array in arrays]), File "/home/arbi01/miniconda3/envs/tmp9/lib/python3.9/site-packages/datasets/table.py", line 1920, in _concat_arrays return pa.ListArray.from_arrays( File "pyarrow/array.pxi", line 1997, in pyarrow.lib.ListArray.from_arrays File "pyarrow/array.pxi", line 1527, in pyarrow.lib.Array.validate File "pyarrow/error.pxi", line 100, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: Negative offsets in list array ``` ### Steps to reproduce the bug 1. put following dataset loading script into: debug/debug.py ```python import datasets import numpy as np class DEBUG(datasets.GeneratorBasedBuilder): """DEBUG dataset.""" def _info(self): return datasets.DatasetInfo( features=datasets.Features( { "id": datasets.Value("uint8"), "img_data": datasets.Array3D(shape=(3, 224, 224), dtype="uint8"), }, ), supervised_keys=None, ) def _split_generators(self, dl_manager): return [datasets.SplitGenerator(name=datasets.Split.TRAIN)] def _generate_examples(self): for i in range(149): image_np = np.zeros(shape=(3, 224, 224), dtype=np.int8).tolist() yield f"id_{i}", {"id": i, "img_data": image_np} ``` 2. try the following code: ```python import datasets def add_dummy_col(ex): ex["dummy"] = "test" return ex ds = datasets.load_dataset(path="debug", split="train") # works ds_filtered_works = ds.filter(lambda example: example["id"] < 95) print(f"filtered result size: {len(ds_filtered_works)}") # output: # filtered result size: 95 ds_mapped_works = ds_filtered_works.map(add_dummy_col) # fails ds_filtered_error = ds.filter(lambda example: example["id"] < 96) print(f"filtered result size: {len(ds_filtered_error)}") # output: # filtered result size: 96 ds_mapped_error = ds_filtered_error.map(add_dummy_col) ``` ### Expected behavior The example code does not fail. ### Environment info Python 3.9.16 (main, Jan 11 2023, 16:05:54); [GCC 11.2.0] :: Anaconda, Inc. on linux datasets 2.9.0
2023-03-03T16:31:33Z
5,534
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-15T16:34:25Z
https://api.github.com/repos/huggingface/datasets/issues/5534/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5534/timeline
map() breaks at certain dataset size when using Array3D
https://api.github.com/repos/huggingface/datasets/issues/5534/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3375489?v=4", "events_url": "https://api.github.com/users/ArneBinder/events{/privacy}", "followers_url": "https://api.github.com/users/ArneBinder/followers", "following_url": "https://api.github.com/users/ArneBinder/following{/other_user}", "gists_url": "https://api.github.com/users/ArneBinder/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ArneBinder", "id": 3375489, "login": "ArneBinder", "node_id": "MDQ6VXNlcjMzNzU0ODk=", "organizations_url": "https://api.github.com/users/ArneBinder/orgs", "received_events_url": "https://api.github.com/users/ArneBinder/received_events", "repos_url": "https://api.github.com/users/ArneBinder/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ArneBinder/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ArneBinder/subscriptions", "type": "User", "url": "https://api.github.com/users/ArneBinder" }
[]
null
null
NONE
null
null
I_kwDODunzps5eiydG
[ "Hi! This code works for me locally or in Colab. What's the output of `python -c \"import pyarrow as pa; print(pa.__version__)\"` when you run it inside your environment?", "Thanks for looking into this!\r\nThe output of `python -c \"import pyarrow as pa; print(pa.__version__)\"` is:\r\n```\r\n11.0.0\r\n```\r\n\...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5534/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5534
https://github.com/huggingface/datasets/issues/5534
false
1,585,885,871
https://api.github.com/repos/huggingface/datasets/issues/5533/labels{/name}
This PR closes #5496 . I tried to imitate the `reduce`-method from `functools`, i.e. the function input must be a binary operation. I assume that the input type has an empty element, i.e. `input_type()` is defined, as the acumulant is instantiated as this object - im not sure that is this a reasonable assumption? If `batched= True` the reduction of each shard is _not_ returned, but the reduction of the entire dataset. I was unsure wether this was an intuitive API, or it would make more sense to return the reduction of each shard?
2023-02-28T14:46:13Z
5,533
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-15T13:44:01Z
https://api.github.com/repos/huggingface/datasets/issues/5533/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5533/timeline
Add reduce function
https://api.github.com/repos/huggingface/datasets/issues/5533/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/38854604?v=4", "events_url": "https://api.github.com/users/AJDERS/events{/privacy}", "followers_url": "https://api.github.com/users/AJDERS/followers", "following_url": "https://api.github.com/users/AJDERS/following{/other_user}", "gists_url": "https://api.github.com/users/AJDERS/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/AJDERS", "id": 38854604, "login": "AJDERS", "node_id": "MDQ6VXNlcjM4ODU0NjA0", "organizations_url": "https://api.github.com/users/AJDERS/orgs", "received_events_url": "https://api.github.com/users/AJDERS/received_events", "repos_url": "https://api.github.com/users/AJDERS/repos", "site_admin": false, "starred_url": "https://api.github.com/users/AJDERS/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AJDERS/subscriptions", "type": "User", "url": "https://api.github.com/users/AJDERS" }
[]
null
null
NONE
2023-02-28T14:46:12Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5533.diff", "html_url": "https://github.com/huggingface/datasets/pull/5533", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5533.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5533" }
PR_kwDODunzps5KCR5I
[ "I agree that it would be a good idea to introduce a `combiner` argument in another PR.\r\n\r\nI did take quite a lot of inspiration from the implementation of `map`, but it did not seem obvious how to resuse `map` for the implementation. Do you have any suggestions, i could give a try?\r\n\r\nThose were exactly m...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5533/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5533
https://github.com/huggingface/datasets/pull/5533
true
1,584,505,128
https://api.github.com/repos/huggingface/datasets/issues/5532/labels{/name}
### Describe the bug When I have a dataset with very few (e.g. 1) examples per class and I call the train_test_split function on it, sometimes the single class will be in the test set. thus will never be considered for training. ### Steps to reproduce the bug ``` import numpy as np from datasets import Dataset data = [ {'label': 0, 'text': "example1"}, {'label': 1, 'text': "example2"}, {'label': 1, 'text': "example3"}, {'label': 1, 'text': "example4"}, {'label': 0, 'text': "example5"}, {'label': 1, 'text': "example6"}, {'label': 2, 'text': "example7"}, {'label': 2, 'text': "example8"} ] for _ in range(10): data_set = Dataset.from_list(data) data_set = data_set.train_test_split(test_size=0.5) data_set["train"] unique_labels_train = np.unique(data_set["train"][:]["label"]) unique_labels_test = np.unique(data_set["test"][:]["label"]) assert len(unique_labels_train) >= len(unique_labels_test) ``` ### Expected behavior I expect to have every available class at least once in my training set. ### Environment info - `datasets` version: 2.9.0 - Platform: Linux-5.15.65+-x86_64-with-debian-bullseye-sid - Python version: 3.7.12 - PyArrow version: 11.0.0 - Pandas version: 1.3.5
2023-02-15T16:09:19Z
5,532
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-14T16:52:29Z
https://api.github.com/repos/huggingface/datasets/issues/5532/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5532/timeline
train_test_split in arrow_dataset does not ensure to keep single classes in test set
https://api.github.com/repos/huggingface/datasets/issues/5532/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/37191008?v=4", "events_url": "https://api.github.com/users/Ulipenitz/events{/privacy}", "followers_url": "https://api.github.com/users/Ulipenitz/followers", "following_url": "https://api.github.com/users/Ulipenitz/following{/other_user}", "gists_url": "https://api.github.com/users/Ulipenitz/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Ulipenitz", "id": 37191008, "login": "Ulipenitz", "node_id": "MDQ6VXNlcjM3MTkxMDA4", "organizations_url": "https://api.github.com/users/Ulipenitz/orgs", "received_events_url": "https://api.github.com/users/Ulipenitz/received_events", "repos_url": "https://api.github.com/users/Ulipenitz/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Ulipenitz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Ulipenitz/subscriptions", "type": "User", "url": "https://api.github.com/users/Ulipenitz" }
[]
null
completed
NONE
2023-02-15T16:09:19Z
null
I_kwDODunzps5ecaEo
[ "Hi! You can get this behavior by specifying `stratify_by_column=\"label\"` in `train_test_split`.\r\n\r\nThis is the full example:\r\n```python\r\nimport numpy as np\r\nfrom datasets import Dataset, ClassLabel\r\n\r\ndata = [\r\n {'label': 0, 'text': \"example1\"},\r\n {'label': 1, 'text': \"example2\"},\r\n...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5532/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5532
https://github.com/huggingface/datasets/issues/5532
false
1,584,387,276
https://api.github.com/repos/huggingface/datasets/issues/5531/labels{/name}
This code fails: ```python from datasets import Dataset ds = Dataset.from_json(path_to_file) ds.data.validate() ``` raises ```python ArrowInvalid: Column 2: In chunk 1: Invalid: Struct child array #3 invalid: Invalid: Length spanned by list offsets (4064) larger than values array (length 4063) ``` This causes many issues for @TevenLeScao: - `map` fails because it fails to rewrite invalid arrow arrays ```python ~/Desktop/hf/datasets/src/datasets/arrow_writer.py in write_examples_on_file(self) 438 if all(isinstance(row[0][col], (pa.Array, pa.ChunkedArray)) for row in self.current_examples): 439 arrays = [row[0][col] for row in self.current_examples] --> 440 batch_examples[col] = array_concat(arrays) 441 else: 442 batch_examples[col] = [ ~/Desktop/hf/datasets/src/datasets/table.py in array_concat(arrays) 1885 1886 if not _is_extension_type(array_type): -> 1887 return pa.concat_arrays(arrays) 1888 1889 def _offsets_concat(offsets): ~/.virtualenvs/hf-datasets/lib/python3.7/site-packages/pyarrow/array.pxi in pyarrow.lib.concat_arrays() ~/.virtualenvs/hf-datasets/lib/python3.7/site-packages/pyarrow/error.pxi in pyarrow.lib.pyarrow_internal_check_status() ~/.virtualenvs/hf-datasets/lib/python3.7/site-packages/pyarrow/error.pxi in pyarrow.lib.check_status() ArrowIndexError: array slice would exceed array length ``` - `to_dict()` **segfaults** ⚠️ ```python /Users/runner/work/crossbow/crossbow/arrow/cpp/src/arrow/array/data.cc:99: Check failed: (off) <= (length) Slice offset greater than array length ``` To reproduce: unzip the archive and run the above code using `sanity_oscar_en.jsonl` [sanity_oscar_en.jsonl.zip](https://github.com/huggingface/datasets/files/10734124/sanity_oscar_en.jsonl.zip) PS: reading using pandas and converting to Arrow works though (note that the dataset lives in RAM in this case): ```python ds = Dataset.from_pandas(pd.read_json(path_to_file, lines=True)) ds.data.validate() ```
2023-02-14T15:46:09Z
5,531
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2023-02-14T15:39:49Z
https://api.github.com/repos/huggingface/datasets/issues/5531/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5531/timeline
Invalid Arrow data from JSONL
https://api.github.com/repos/huggingface/datasets/issues/5531/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
null
null
I_kwDODunzps5eb9TM
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5531/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5531
https://github.com/huggingface/datasets/issues/5531
false
1,582,938,241
https://api.github.com/repos/huggingface/datasets/issues/5530/labels{/name}
## What's in this PR? As discussed with @lhoestq in https://github.com/huggingface/datasets/pull/5522, the license for `NumpyFormatter` at `datasets/formatting/np_formatter.py` was missing, but present on the rest of the `formatting/*.py` files. So this PR is basically to include it there.
2023-02-14T14:40:41Z
5,530
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-13T19:33:23Z
https://api.github.com/repos/huggingface/datasets/issues/5530/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5530/timeline
Add missing license in `NumpyFormatter`
https://api.github.com/repos/huggingface/datasets/issues/5530/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
[]
null
null
CONTRIBUTOR
2023-02-14T12:23:58Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5530.diff", "html_url": "https://github.com/huggingface/datasets/pull/5530", "merged_at": "2023-02-14T12:23:58Z", "patch_url": "https://github.com/huggingface/datasets/pull/5530.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5530" }
PR_kwDODunzps5J4W_4
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5530/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5530
https://github.com/huggingface/datasets/pull/5530
true
1,582,501,233
https://api.github.com/repos/huggingface/datasets/issues/5529/labels{/name}
## What's in this PR? After playing around a little bit with 🤗`datasets` in Google Cloud Storage (GCS), I found out some things that should be fixed IMO in the code: * `datasets.load_from_disk` is not checking whether `state.json` is there too when trying to load a `Dataset`, just `dataset_info.json` is checked * `DatasetDict.load_from_disk` is not checking whether `state.json` is there too when redirecting the user to load it as `datasets.load_from_disk`, just `dataset_info.json` is checked, which is misleading, as it won't be loadable that way either * `Dataset.load_from_disk` is missing the `extract_path_from_uri` call before checking in the `fs` whether `dataset_info.json` and `dataset_dict.json` exist, which when using `gcsfs` leads to 400 error code (not blocking) due to `gcsfs.retry.HttpError: Invalid bucket name: 'gs:', 400` * And, finally, the exception messages are a little bit misleading / incomplete IMO so I've tried to include all the relevant information in the messages to avoid issues when interpreting the exceptions
2023-02-23T18:14:32Z
5,529
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-13T14:54:55Z
https://api.github.com/repos/huggingface/datasets/issues/5529/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5529/timeline
Fix `datasets.load_from_disk`, `DatasetDict.load_from_disk` and `Dataset.load_from_disk`
https://api.github.com/repos/huggingface/datasets/issues/5529/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
[]
null
null
CONTRIBUTOR
2023-02-23T18:05:26Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5529.diff", "html_url": "https://github.com/huggingface/datasets/pull/5529", "merged_at": "2023-02-23T18:05:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/5529.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5529" }
PR_kwDODunzps5J26Fq
[ "_The documentation is not available anymore as the PR was closed or merged._", "Hmm, should this also be updated in `Dataset.load_from_disk` and `DatasetDict.load_from_disk`? https://github.com/huggingface/datasets/pull/5466 As there the paths are joined using `Path(..., ...)` and it won't work on Windows OS acc...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5529/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5529
https://github.com/huggingface/datasets/pull/5529
true
1,582,195,085
https://api.github.com/repos/huggingface/datasets/issues/5528/labels{/name}
Fixes #5492. Introduce new kwarg `create_pr` in `push_to_hub`, which is passed to `HFapi.upload_file`.
2023-10-06T21:58:02Z
5,528
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-13T11:43:47Z
https://api.github.com/repos/huggingface/datasets/issues/5528/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5528/timeline
Push to hub in a pull request
https://api.github.com/repos/huggingface/datasets/issues/5528/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/38854604?v=4", "events_url": "https://api.github.com/users/AJDERS/events{/privacy}", "followers_url": "https://api.github.com/users/AJDERS/followers", "following_url": "https://api.github.com/users/AJDERS/following{/other_user}", "gists_url": "https://api.github.com/users/AJDERS/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/AJDERS", "id": 38854604, "login": "AJDERS", "node_id": "MDQ6VXNlcjM4ODU0NjA0", "organizations_url": "https://api.github.com/users/AJDERS/orgs", "received_events_url": "https://api.github.com/users/AJDERS/received_events", "repos_url": "https://api.github.com/users/AJDERS/repos", "site_admin": false, "starred_url": "https://api.github.com/users/AJDERS/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AJDERS/subscriptions", "type": "User", "url": "https://api.github.com/users/AJDERS" }
[]
null
null
NONE
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/5528.diff", "html_url": "https://github.com/huggingface/datasets/pull/5528", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5528.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5528" }
PR_kwDODunzps5J13wC
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5528). All of your documentation changes will be reflected on that endpoint.", "It seems that the parameter `create_pr` is available for [`0.8.0`](https://huggingface.co/docs/huggingface_hub/v0.8.1/en/package_reference/hf_api#h...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5528/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5528
https://github.com/huggingface/datasets/pull/5528
true
1,581,228,531
https://api.github.com/repos/huggingface/datasets/issues/5527/labels{/name}
fix https://github.com/huggingface/datasets/actions/runs/4156059127/jobs/7189576331
2023-02-13T10:29:03Z
5,527
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-12T11:51:25Z
https://api.github.com/repos/huggingface/datasets/issues/5527/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5527/timeline
Fix benchmarks CI - pin protobuf
https://api.github.com/repos/huggingface/datasets/issues/5527/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-13T09:24:16Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5527.diff", "html_url": "https://github.com/huggingface/datasets/pull/5527", "merged_at": "2023-02-13T09:24:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/5527.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5527" }
PR_kwDODunzps5JysSM
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5527/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5527
https://github.com/huggingface/datasets/pull/5527
true
1,580,488,133
https://api.github.com/repos/huggingface/datasets/issues/5526/labels{/name}
Fixes #5428 Allow loading/saving of FAISS index using fsspec: 1. Simply use BufferedIOWriter/Reader to Read/Write indices on fsspec stream. 2. Needed `mockfs` in the test, so I took it out of the `TestCase`. Let me know if that makes sense. I can work on the documentation once the code changes are approved.
2023-03-27T15:26:46Z
5,526
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-10T23:37:14Z
https://api.github.com/repos/huggingface/datasets/issues/5526/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5526/timeline
Allow loading/saving of FAISS index using fsspec
https://api.github.com/repos/huggingface/datasets/issues/5526/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8976546?v=4", "events_url": "https://api.github.com/users/Dref360/events{/privacy}", "followers_url": "https://api.github.com/users/Dref360/followers", "following_url": "https://api.github.com/users/Dref360/following{/other_user}", "gists_url": "https://api.github.com/users/Dref360/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Dref360", "id": 8976546, "login": "Dref360", "node_id": "MDQ6VXNlcjg5NzY1NDY=", "organizations_url": "https://api.github.com/users/Dref360/orgs", "received_events_url": "https://api.github.com/users/Dref360/received_events", "repos_url": "https://api.github.com/users/Dref360/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Dref360/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Dref360/subscriptions", "type": "User", "url": "https://api.github.com/users/Dref360" }
[]
null
null
CONTRIBUTOR
2023-03-27T15:18:20Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5526.diff", "html_url": "https://github.com/huggingface/datasets/pull/5526", "merged_at": "2023-03-27T15:18:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/5526.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5526" }
PR_kwDODunzps5JwVol
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks for the quick review! I updated the code with your suggestion", "Thanks for the quick review @albertvillanova! I updated the code with your suggestions", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5526/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5526
https://github.com/huggingface/datasets/pull/5526
true
1,580,342,729
https://api.github.com/repos/huggingface/datasets/issues/5525/labels{/name}
### Describe the bug Processing a dataset I alredy uploaded to the Hub (https://huggingface.co/datasets/tj-solergibert/Europarl-ST) I found that for some splits and some languages (test split, source_lang = "nl") after applying a map function I get the mentioned error. I alredy tried reseting the shorter strings (reset_cortas function). It only happends with NL, PL, RO and PT. It does not make sense since when processing the other languages I also use the corpus of those that fail and it does not cause any errors. I suspect that the error may be in this direction: We use cast_array_to_feature to support casting to custom types like Audio and Image # Also, when trying type "string", we don't want to convert integers or floats to "string". # We only do it if trying_type is False - since this is what the user asks for. ### Steps to reproduce the bug Here I link a colab notebook to reproduce the error: https://colab.research.google.com/drive/1JCrS7FlGfu_kFqChMrwKZ_bpabnIMqbP?authuser=1#scrollTo=FBAvlhMxIzpA ### Expected behavior Data processing does not fail. A correct example can be seen here: https://huggingface.co/datasets/tj-solergibert/Europarl-ST-processed-mt-en ### Environment info - `datasets` version: 2.9.0 - Platform: Linux-5.10.147+-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 9.0.0 - Pandas version: 1.3.5
2023-02-14T17:41:08Z
5,525
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-10T21:12:36Z
https://api.github.com/repos/huggingface/datasets/issues/5525/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5525/timeline
TypeError: Couldn't cast array of type string to null
https://api.github.com/repos/huggingface/datasets/issues/5525/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/74564958?v=4", "events_url": "https://api.github.com/users/TJ-Solergibert/events{/privacy}", "followers_url": "https://api.github.com/users/TJ-Solergibert/followers", "following_url": "https://api.github.com/users/TJ-Solergibert/following{/other_user}", "gists_url": "https://api.github.com/users/TJ-Solergibert/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/TJ-Solergibert", "id": 74564958, "login": "TJ-Solergibert", "node_id": "MDQ6VXNlcjc0NTY0OTU4", "organizations_url": "https://api.github.com/users/TJ-Solergibert/orgs", "received_events_url": "https://api.github.com/users/TJ-Solergibert/received_events", "repos_url": "https://api.github.com/users/TJ-Solergibert/repos", "site_admin": false, "starred_url": "https://api.github.com/users/TJ-Solergibert/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TJ-Solergibert/subscriptions", "type": "User", "url": "https://api.github.com/users/TJ-Solergibert" }
[]
null
completed
NONE
2023-02-14T09:35:49Z
null
I_kwDODunzps5eMh3J
[ "Thanks for reporting, @TJ-Solergibert.\r\n\r\nWe cannot access your Colab notebook: `There was an error loading this notebook. Ensure that the file is accessible and try again.`\r\nCould you please make it publicly accessible?\r\n", "I swear it's public, I've checked the settings and I've been able to open it in...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5525/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5525
https://github.com/huggingface/datasets/issues/5525
false
1,580,219,454
https://api.github.com/repos/huggingface/datasets/issues/5524/labels{/name}
Hi to whoever is reading this! 🤗 ## What's in this PR? ~~Basically, I've removed the 🤗`datasets` installation as `python -m pip install ".[quality]" in the `check_code_quality` job in `.github/workflows/ci.yaml`, as we don't need to install the whole package to run the CI, unless that's done on purpose e.g. to check that the Python package installation succeeds before running the tests over the matrix of os?~~ ~~So I just wanted to check whether the time was reduced doing this (which I assume it will), plus whether this is something that can be improved, or just discarded in case you're also using that step to make sure that the package can be installed.~~ ## What's missing? ~~I was just wondering whether you consider replacing `isort` and `flake8` with `ruff` (if possible), since it's way faster, more information at [`ruff`](https://github.com/charliermarsh/ruff). Before creating this PR the average time of the `check_code_quality` job was around 40s.~~ ## Edit Sorry for the inconvenience this may have caused, didn't realise that the config is defined in `setup.cfg` and `pyproject.toml`, so running those without installing the Python package leads to failure, my bad 😞
2023-02-10T19:51:45Z
5,524
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-10T19:35:50Z
https://api.github.com/repos/huggingface/datasets/issues/5524/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5524/timeline
[INVALID PR]
https://api.github.com/repos/huggingface/datasets/issues/5524/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
[]
null
null
CONTRIBUTOR
2023-02-10T19:49:12Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5524.diff", "html_url": "https://github.com/huggingface/datasets/pull/5524", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5524.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5524" }
PR_kwDODunzps5JvbMw
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5524/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5524
https://github.com/huggingface/datasets/pull/5524
true
1,580,193,015
https://api.github.com/repos/huggingface/datasets/issues/5523/labels{/name}
### Describe the bug Verification of split names (=indexing data by split) happens after downloading the data. So when the split name is incorrect, users learn about that only after the data is fully downloaded, for large datasets it might take a lot of time. ### Steps to reproduce the bug Load any dataset with random split name, for example: ```python from datasets import load_dataset load_dataset("mozilla-foundation/common_voice_11_0", "en", split="blabla") ``` and the download will start smoothly, despite there is no split named "blabla". ### Expected behavior Raise error when split name is incorrect. ### Environment info `datasets==2.9.1.dev0`
2023-02-10T19:14:50Z
5,523
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2023-02-10T19:13:03Z
https://api.github.com/repos/huggingface/datasets/issues/5523/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
https://api.github.com/repos/huggingface/datasets/issues/5523/timeline
Checking that split name is correct happens only after the data is downloaded
https://api.github.com/repos/huggingface/datasets/issues/5523/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_use...
null
null
CONTRIBUTOR
null
null
I_kwDODunzps5eL9T3
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5523/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5523
https://github.com/huggingface/datasets/issues/5523
false
1,580,183,124
https://api.github.com/repos/huggingface/datasets/issues/5522/labels{/name}
Hi to whoever is reading this! 🤗 ## What's in this PR? I was exploring the code regarding the `JaxFormatter` implemented in 🤗`datasets`, and found some things that IMO could be changed. Those are mainly regarding the docstrings and the type-hints based on `jax`'s 0.4.1 release where `jax.Array` was introduced as the default type for JAX-arrays (instead of `jnp.DeviceArray`, `jnp.SharedDeviceArray`, and `jnp.GlobalDeviceArray`). Even though `isinstance(..., jax.Array)` also works with lower versions such as e.g. `0.3.25`. More information about the latter at [`jax` v0.4.1 - Release Notes](https://github.com/google/jax/releases/tag/jax-v0.4.1) and [jax.Array migration - JAX documentation](https://jax.readthedocs.io/en/latest/jax_array_migration.html). ## What's missing? * Do you want me to write an entry in the documentation on how to use 🤗`datasets` with JAX as https://huggingface.co/docs/datasets/use_with_pytorch with PyTorch? * Do we need to actually include `pyarrow` under the `TYPE_CHECKING` when needed? I just did it for JAX, but if we are OK with that, I can do that with the rest of the formatters, just LMK. * Should the License header be included in `datasets.formatting.np_formatter`? If so, do I include the one from 2020 e.g. https://github.com/huggingface/datasets/blob/b065547654efa0ec633cf373ac1512884c68b2e1/src/datasets/formatting/tf_formatter.py#L1-L13 * Is there any reason why `jnp.array` is being used instead of `jnp.asarray`? There's no difference between both, just that `jnp.asarray` has `copy=False` as default, even though `numpy` to `jax.numpy` conversion is not zero-copy, but just asking :)
2023-02-15T14:48:27Z
5,522
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-10T19:05:00Z
https://api.github.com/repos/huggingface/datasets/issues/5522/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5522/timeline
Minor changes in JAX-formatting docstrings & type-hints
https://api.github.com/repos/huggingface/datasets/issues/5522/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
[]
null
null
CONTRIBUTOR
2023-02-15T13:19:06Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5522.diff", "html_url": "https://github.com/huggingface/datasets/pull/5522", "merged_at": "2023-02-15T13:19:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/5522.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5522" }
PR_kwDODunzps5JvTVp
[ "P.S. For more context, I'm currently exploring the integration of 🤗`datasets` with JAX, so in case you need any help or want me to try something specific just let me know! (`jnp.asarray`/`jnp.array(..., copy=False)` still no zero-copy 😭)", "_The documentation is not available anymore as the PR was closed or me...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5522/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5522
https://github.com/huggingface/datasets/pull/5522
true
1,578,418,289
https://api.github.com/repos/huggingface/datasets/issues/5521/labels{/name}
Fix https://github.com/huggingface/datasets/issues/5520.
2023-02-13T20:40:48Z
5,521
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-09T18:47:59Z
https://api.github.com/repos/huggingface/datasets/issues/5521/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5521/timeline
Fix bug when casting empty array to class labels
https://api.github.com/repos/huggingface/datasets/issues/5521/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6591505?v=4", "events_url": "https://api.github.com/users/marioga/events{/privacy}", "followers_url": "https://api.github.com/users/marioga/followers", "following_url": "https://api.github.com/users/marioga/following{/other_user}", "gists_url": "https://api.github.com/users/marioga/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/marioga", "id": 6591505, "login": "marioga", "node_id": "MDQ6VXNlcjY1OTE1MDU=", "organizations_url": "https://api.github.com/users/marioga/orgs", "received_events_url": "https://api.github.com/users/marioga/received_events", "repos_url": "https://api.github.com/users/marioga/repos", "site_admin": false, "starred_url": "https://api.github.com/users/marioga/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/marioga/subscriptions", "type": "User", "url": "https://api.github.com/users/marioga" }
[]
null
null
CONTRIBUTOR
2023-02-12T11:17:17Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5521.diff", "html_url": "https://github.com/huggingface/datasets/pull/5521", "merged_at": "2023-02-12T11:17:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/5521.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5521" }
PR_kwDODunzps5JpWnp
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5521/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5521
https://github.com/huggingface/datasets/pull/5521
true
1,578,417,074
https://api.github.com/repos/huggingface/datasets/issues/5520/labels{/name}
### Describe the bug `ClassLabel.cast_storage` raises `TypeError` when called on an empty `IntegerArray`. ### Steps to reproduce the bug Minimal steps: ```python import pyarrow as pa from datasets import ClassLabel ClassLabel(names=['foo', 'bar']).cast_storage(pa.array([], pa.int64())) ``` In practice, this bug arises in situations like the one below: ```python from datasets import ClassLabel, Dataset, Features, Sequence dataset = Dataset.from_dict({'labels': [[], []]}, features=Features({'labels': Sequence(ClassLabel(names=['foo', 'bar']))})) # this raises TypeError dataset.map(batched=True, batch_size=1) ``` ### Expected behavior `ClassLabel.cast_storage` should return an empty Int64Array. ### Environment info - `datasets` version: 2.9.1.dev0 - Platform: Linux-4.15.0-1032-aws-x86_64-with-glibc2.27 - Python version: 3.10.6 - PyArrow version: 11.0.0 - Pandas version: 1.5.3
2023-02-12T11:17:18Z
5,520
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-09T18:46:52Z
https://api.github.com/repos/huggingface/datasets/issues/5520/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5520/timeline
ClassLabel.cast_storage raises TypeError when called on an empty IntegerArray
https://api.github.com/repos/huggingface/datasets/issues/5520/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6591505?v=4", "events_url": "https://api.github.com/users/marioga/events{/privacy}", "followers_url": "https://api.github.com/users/marioga/followers", "following_url": "https://api.github.com/users/marioga/following{/other_user}", "gists_url": "https://api.github.com/users/marioga/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/marioga", "id": 6591505, "login": "marioga", "node_id": "MDQ6VXNlcjY1OTE1MDU=", "organizations_url": "https://api.github.com/users/marioga/orgs", "received_events_url": "https://api.github.com/users/marioga/received_events", "repos_url": "https://api.github.com/users/marioga/repos", "site_admin": false, "starred_url": "https://api.github.com/users/marioga/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/marioga/subscriptions", "type": "User", "url": "https://api.github.com/users/marioga" }
[]
null
completed
CONTRIBUTOR
2023-02-12T11:17:18Z
null
I_kwDODunzps5eFLuy
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5520/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5520
https://github.com/huggingface/datasets/issues/5520
false
1,578,341,785
https://api.github.com/repos/huggingface/datasets/issues/5519/labels{/name}
Use `ruff` for formatting instead of `isort` and `black` to be consistent with [`transformers`](https://github.com/huggingface/transformers/pull/21480) and [`hfh`](https://github.com/huggingface/huggingface_hub/pull/1323). TODO: - [x] ~Merge the community contributors' PR to avoid having to run `make style` on their PR branches~ (we have some new PRs, but fixing those shouldn't be too big of a problem)
2023-02-14T16:28:27Z
5,519
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-09T17:50:21Z
https://api.github.com/repos/huggingface/datasets/issues/5519/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5519/timeline
Format code with `ruff`
https://api.github.com/repos/huggingface/datasets/issues/5519/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2023-02-14T16:18:38Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5519.diff", "html_url": "https://github.com/huggingface/datasets/pull/5519", "merged_at": "2023-02-14T16:18:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/5519.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5519" }
PR_kwDODunzps5JpGPl
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5519/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5519
https://github.com/huggingface/datasets/pull/5519
true
1,578,203,962
https://api.github.com/repos/huggingface/datasets/issues/5518/labels{/name}
Fix https://github.com/huggingface/datasets/issues/3841
2023-02-13T13:55:49Z
5,518
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-09T16:22:29Z
https://api.github.com/repos/huggingface/datasets/issues/5518/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5518/timeline
Remove py.typed
https://api.github.com/repos/huggingface/datasets/issues/5518/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2023-02-13T13:48:40Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5518.diff", "html_url": "https://github.com/huggingface/datasets/pull/5518", "merged_at": "2023-02-13T13:48:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/5518.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5518" }
PR_kwDODunzps5Joom3
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5518/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5518
https://github.com/huggingface/datasets/pull/5518
true
1,577,976,608
https://api.github.com/repos/huggingface/datasets/issues/5517/labels{/name}
### Describe the bug When I create a dataset with a `float64` feature, then apply numpy formatting the returned numpy arrays are silently downcasted to `float32`. ### Steps to reproduce the bug ```python import datasets dataset = datasets.Dataset.from_dict({'a': [1.0, 2.0, 3.0]}).with_format("numpy") print("feature dtype:", dataset.features['a'].dtype) print("array dtype:", dataset['a'].dtype) ``` output: ``` feature dtype: float64 array dtype: float32 ``` ### Expected behavior ``` feature dtype: float64 array dtype: float64 ``` ### Environment info - `datasets` version: 2.8.0 - Platform: Linux-5.4.0-135-generic-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 10.0.1 - Pandas version: 1.4.4 ### Suggested Fix Changing [the `_tensorize` function of the numpy formatter](https://github.com/huggingface/datasets/blob/b065547654efa0ec633cf373ac1512884c68b2e1/src/datasets/formatting/np_formatter.py#L32) to ```python def _tensorize(self, value): if isinstance(value, (str, bytes, type(None))): return value elif isinstance(value, (np.character, np.ndarray)) and np.issubdtype(value.dtype, np.character): return value elif isinstance(value, np.number): return value return np.asarray(value, **self.np_array_kwargs) ``` fixes this particular issue for me. Not sure if this would break other tests. This should also avoid unnecessary copying of the array.
2024-01-18T08:42:17Z
5,517
{ "closed_at": null, "closed_issues": 1, "created_at": "2023-02-13T16:22:42Z", "creator": { "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }, "description": "Next major release", "due_on": null, "html_url": "https://github.com/huggingface/datasets/milestone/10", "id": 9038583, "labels_url": "https://api.github.com/repos/huggingface/datasets/milestones/10/labels", "node_id": "MI_kwDODunzps4Aier3", "number": 10, "open_issues": 3, "state": "open", "title": "3.0", "updated_at": "2024-03-19T15:29:43Z", "url": "https://api.github.com/repos/huggingface/datasets/milestones/10" }
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-09T14:18:00Z
https://api.github.com/repos/huggingface/datasets/issues/5517/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5517/timeline
`with_format("numpy")` silently downcasts float64 to float32 features
https://api.github.com/repos/huggingface/datasets/issues/5517/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1250234?v=4", "events_url": "https://api.github.com/users/ernestum/events{/privacy}", "followers_url": "https://api.github.com/users/ernestum/followers", "following_url": "https://api.github.com/users/ernestum/following{/other_user}", "gists_url": "https://api.github.com/users/ernestum/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ernestum", "id": 1250234, "login": "ernestum", "node_id": "MDQ6VXNlcjEyNTAyMzQ=", "organizations_url": "https://api.github.com/users/ernestum/orgs", "received_events_url": "https://api.github.com/users/ernestum/received_events", "repos_url": "https://api.github.com/users/ernestum/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ernestum/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ernestum/subscriptions", "type": "User", "url": "https://api.github.com/users/ernestum" }
[]
null
null
NONE
null
null
I_kwDODunzps5eDgMg
[ "Hi! This behavior stems from these lines:\r\n\r\nhttps://github.com/huggingface/datasets/blob/b065547654efa0ec633cf373ac1512884c68b2e1/src/datasets/formatting/np_formatter.py#L45-L46\r\n\r\nI agree we should preserve the original type whenever possible and downcast explicitly with a warning.\r\n\r\n@lhoestq Do you...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5517/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5517
https://github.com/huggingface/datasets/issues/5517
false
1,577,661,640
https://api.github.com/repos/huggingface/datasets/issues/5516/labels{/name}
Resolves #5482. Attaches feature metadata to parquet files serialised using `Dataset.to_parquet`. This allows retrieving data with "rich" feature types (e.g., `datasets.features.image.Image` or `datasets.features.audio.Audio`) from parquet files without cumbersome casting (for an example, see #5482). @lhoestq It seems that it is sufficient to attach metadata to the schema prior to serialising and features are loaded back with correct types afterwards automatically. I used the following script to test the implementation: ```python from pathlib import Path import datasets dataset_name = "Maysee/tiny-imagenet" ds = datasets.load_dataset(dataset_name, split=datasets.Split.TRAIN) output_directory_path = Path(__file__).parent.joinpath("example_test_outputs", dataset_name.replace("/", "_")) output_directory_path.mkdir(exist_ok=True, parents=True) output_filepath = output_directory_path.joinpath("ds.parquet") ds.to_parquet(str(output_filepath)) reloaded_ds = datasets.load_dataset(str(output_directory_path), split=datasets.Split.TRAIN) assert ds.features == reloaded_ds.features ``` Prior to the change in this PR this script raises an `AssertionError` and the `Image` features lose their type after serialisation. After the change in this PR, the assertion does not raise an error and manual inspection of the features shows type `Image` for the respective columns of `reloaded_ds `. Some open questions: * How/where can I best add new unit tests for this implementation? * What dataset would I best use in the tests? I chose `Maysee/tiny-imagenet` mainly because it is small and contains an ?Image` feature that can be used to test, but I'd be happy for suggestions on a suitable data source to use. * Currently I'm calling `datasets.arrow_writer.ArrowWriter._build_metadata` as I need the same logic. However, I'm not happy with the coupling between `datasets.io.parquet` and `datasets.arrow_writer` it leaves me with. Suggest to factor this common logic out into a helper function and reuse it from both of these. Do you agree and if yes, could you please guide me where I would best place this function? Many thanks in advance and kind regards, MFreidank
2023-02-12T16:00:00Z
5,516
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-09T10:52:15Z
https://api.github.com/repos/huggingface/datasets/issues/5516/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5516/timeline
Reload features from Parquet metadata
https://api.github.com/repos/huggingface/datasets/issues/5516/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6368040?v=4", "events_url": "https://api.github.com/users/MFreidank/events{/privacy}", "followers_url": "https://api.github.com/users/MFreidank/followers", "following_url": "https://api.github.com/users/MFreidank/following{/other_user}", "gists_url": "https://api.github.com/users/MFreidank/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/MFreidank", "id": 6368040, "login": "MFreidank", "node_id": "MDQ6VXNlcjYzNjgwNDA=", "organizations_url": "https://api.github.com/users/MFreidank/orgs", "received_events_url": "https://api.github.com/users/MFreidank/received_events", "repos_url": "https://api.github.com/users/MFreidank/repos", "site_admin": false, "starred_url": "https://api.github.com/users/MFreidank/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MFreidank/subscriptions", "type": "User", "url": "https://api.github.com/users/MFreidank" }
[]
null
null
CONTRIBUTOR
2023-02-12T15:57:01Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5516.diff", "html_url": "https://github.com/huggingface/datasets/pull/5516", "merged_at": "2023-02-12T15:57:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/5516.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5516" }
PR_kwDODunzps5JmzPQ
[ "Thanks a lot for your help @lhoestq. I've simplified what turned out to be a simple fix and added the unit test.\r\n\r\nDoes this look ready to be merged or is there anything I'm still missing?", "Cool ! I think you just need to remove the unused import in `io/parquet.py`\r\n```\r\nsrc/datasets/io/parquet.py:4:1...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5516/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5516
https://github.com/huggingface/datasets/pull/5516
true
1,577,590,611
https://api.github.com/repos/huggingface/datasets/issues/5515/labels{/name}
* Updating type annotations for #`load_from_cache_file` * Added logic for cache checking if needed * Updated documentation following the wording of `Dataset.map`
2023-02-14T15:38:13Z
5,515
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-09T10:04:46Z
https://api.github.com/repos/huggingface/datasets/issues/5515/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5515/timeline
Unify `load_from_cache_file` type and logic
https://api.github.com/repos/huggingface/datasets/issues/5515/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/22773355?v=4", "events_url": "https://api.github.com/users/HallerPatrick/events{/privacy}", "followers_url": "https://api.github.com/users/HallerPatrick/followers", "following_url": "https://api.github.com/users/HallerPatrick/following{/other_user}", "gists_url": "https://api.github.com/users/HallerPatrick/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/HallerPatrick", "id": 22773355, "login": "HallerPatrick", "node_id": "MDQ6VXNlcjIyNzczMzU1", "organizations_url": "https://api.github.com/users/HallerPatrick/orgs", "received_events_url": "https://api.github.com/users/HallerPatrick/received_events", "repos_url": "https://api.github.com/users/HallerPatrick/repos", "site_admin": false, "starred_url": "https://api.github.com/users/HallerPatrick/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/HallerPatrick/subscriptions", "type": "User", "url": "https://api.github.com/users/HallerPatrick" }
[]
null
null
CONTRIBUTOR
2023-02-14T14:26:42Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5515.diff", "html_url": "https://github.com/huggingface/datasets/pull/5515", "merged_at": "2023-02-14T14:26:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/5515.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5515" }
PR_kwDODunzps5Jmj5X
[ "_The documentation is not available anymore as the PR was closed or merged._", "The commit also includes the changes to the `DatasetDict` methods or am I missing something?", "Oh, indeed. Feel free to mark the PR as \"Ready for review\" then.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5515/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5515
https://github.com/huggingface/datasets/pull/5515
true
1,576,453,837
https://api.github.com/repos/huggingface/datasets/issues/5514/labels{/name}
### Feature request 1. Replace the `load_from_cache_file` default value to `True`. 2. Remove or alter checks from `is_caching_enabled` logic. ### Motivation I stumbled over an inconsistency in the `Dataset.map` interface. The documentation (and source) states for the parameter `load_from_cache_file`: ``` load_from_cache_file (`bool`, defaults to `True` if caching is enabled): If a cache file storing the current computation from `function` can be identified, use it instead of recomputing. ``` 1. `load_from_cache_file` default value is `None`, while being annotated as `bool` 2. It is inconsistent with other method signatures like `filter`, that have the default value `True` 3. The logic is inconsistent, as the `map` method checks if caching is enabled through `is_caching_enabled`. This logic is not used for other similar methods. ### Your contribution I am not fully aware of the logic behind caching checks. If this is just a inconsistency that historically grew, I would suggest to remove the `is_caching_enabled` logic as the "default" logic. Maybe someone can give insights, if environment variables have a higher priority than local variables or vice versa. If this is clarified, I could adjust the source according to the "Feature request" section of this issue.
2023-02-14T14:26:44Z
5,514
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-02-08T16:40:44Z
https://api.github.com/repos/huggingface/datasets/issues/5514/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5514/timeline
Improve inconsistency of `Dataset.map` interface for `load_from_cache_file`
https://api.github.com/repos/huggingface/datasets/issues/5514/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/22773355?v=4", "events_url": "https://api.github.com/users/HallerPatrick/events{/privacy}", "followers_url": "https://api.github.com/users/HallerPatrick/followers", "following_url": "https://api.github.com/users/HallerPatrick/following{/other_user}", "gists_url": "https://api.github.com/users/HallerPatrick/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/HallerPatrick", "id": 22773355, "login": "HallerPatrick", "node_id": "MDQ6VXNlcjIyNzczMzU1", "organizations_url": "https://api.github.com/users/HallerPatrick/orgs", "received_events_url": "https://api.github.com/users/HallerPatrick/received_events", "repos_url": "https://api.github.com/users/HallerPatrick/repos", "site_admin": false, "starred_url": "https://api.github.com/users/HallerPatrick/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/HallerPatrick/subscriptions", "type": "User", "url": "https://api.github.com/users/HallerPatrick" }
[]
null
completed
CONTRIBUTOR
2023-02-14T14:26:44Z
null
I_kwDODunzps5d9sbN
[ "Hi, thanks for noticing this! We can't just remove the cache control as this allows us to control where the arrow files generated by the ops are written (cached on disk if enabled or a temporary directory if disabled). The right way to address this inconsistency would be by having `load_from_cache_file=None` by de...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5514/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5514
https://github.com/huggingface/datasets/issues/5514
false
1,576,300,803
https://api.github.com/repos/huggingface/datasets/issues/5513/labels{/name}
Hi @mariosasko, @lhoestq, or whoever reads this! :) After going through `ArrowDataset.set_format` I found out that the `type` param is actually named `type` which is a Python reserved name as you may already know, shouldn't that be renamed to `format_type` before the 3.0.0 is released? Just wanted to get your input, and if applicable, tackle this issue myself! Thanks 🤗
2023-07-24T16:02:18Z
5,513
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-08T15:13:46Z
https://api.github.com/repos/huggingface/datasets/issues/5513/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5513/timeline
Some functions use a param named `type` shouldn't that be avoided since it's a Python reserved name?
https://api.github.com/repos/huggingface/datasets/issues/5513/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36760800?v=4", "events_url": "https://api.github.com/users/alvarobartt/events{/privacy}", "followers_url": "https://api.github.com/users/alvarobartt/followers", "following_url": "https://api.github.com/users/alvarobartt/following{/other_user}", "gists_url": "https://api.github.com/users/alvarobartt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alvarobartt", "id": 36760800, "login": "alvarobartt", "node_id": "MDQ6VXNlcjM2NzYwODAw", "organizations_url": "https://api.github.com/users/alvarobartt/orgs", "received_events_url": "https://api.github.com/users/alvarobartt/received_events", "repos_url": "https://api.github.com/users/alvarobartt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alvarobartt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alvarobartt/subscriptions", "type": "User", "url": "https://api.github.com/users/alvarobartt" }
[]
null
completed
CONTRIBUTOR
2023-07-24T14:27:59Z
null
I_kwDODunzps5d9HED
[ "Hi! Let's not do this - renaming it would be a breaking change, and going through the deprecation cycle is only worth it if it improves user experience.", "Hi @mariosasko, ok it makes sense. Anyway, don't you think it's worth it at some point to start a deprecation cycle e.g. `fs` in `load_from_disk`? It doesn't...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5513/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5513
https://github.com/huggingface/datasets/issues/5513
false
1,576,142,432
https://api.github.com/repos/huggingface/datasets/issues/5512/labels{/name}
I implemented `__getitems__` to speed up batched data loading in PyTorch close https://github.com/huggingface/datasets/issues/5505
2023-02-19T18:35:09Z
5,512
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-08T13:38:59Z
https://api.github.com/repos/huggingface/datasets/issues/5512/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5512/timeline
Speed up batched PyTorch DataLoader
https://api.github.com/repos/huggingface/datasets/issues/5512/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-19T18:27:29Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5512.diff", "html_url": "https://github.com/huggingface/datasets/pull/5512", "merged_at": "2023-02-19T18:27:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/5512.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5512" }
PR_kwDODunzps5JhtQy
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5512/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5512
https://github.com/huggingface/datasets/pull/5512
true
1,575,851,768
https://api.github.com/repos/huggingface/datasets/issues/5511/labels{/name}
### Describe the bug I often want to create a dummy dataset from a bigger dataset for fast iteration when training. However, I'm having a hard time doing this especially when trying to upload the dataset to the Hub. ### Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("lambdalabs/pokemon-blip-captions") dataset["train"] = dataset["train"].select(range(20)) dataset.push_to_hub("patrickvonplaten/dummy_image_data") ``` gives: ``` ~/python_bin/datasets/arrow_dataset.py in _push_parquet_shards_to_hub(self, repo_id, split, private, token, branch, max_shard_size, embed_external_files) 4003 base_wait_time=2.0, 4004 max_retries=5, -> 4005 max_wait_time=20.0, 4006 ) 4007 return repo_id, split, uploaded_size, dataset_nbytes ~/python_bin/datasets/utils/file_utils.py in _retry(func, func_args, func_kwargs, exceptions, status_codes, max_retries, base_wait_time, max_wait_time) 328 while True: 329 try: --> 330 return func(*func_args, **func_kwargs) 331 except exceptions as err: 332 if retry >= max_retries or (status_codes and err.response.status_code not in status_codes): ~/hf/lib/python3.7/site-packages/huggingface_hub/utils/_validators.py in _inner_fn(*args, **kwargs) 122 ) 123 --> 124 return fn(*args, **kwargs) 125 126 return _inner_fn # type: ignore TypeError: upload_file() got an unexpected keyword argument 'identical_ok' In [2]: ``` ### Expected behavior I would have expected this to work. It's for me the most intuitive way of creating a dummy dataset. ### Environment info ``` - `datasets` version: 2.1.1.dev0 - Platform: Linux-4.19.0-22-cloud-amd64-x86_64-with-debian-10.13 - Python version: 3.7.3 - PyArrow version: 11.0.0 - Pandas version: 1.3.5 ```
2023-12-28T18:21:01Z
5,511
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-08T10:18:41Z
https://api.github.com/repos/huggingface/datasets/issues/5511/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5511/timeline
Creating a dummy dataset from a bigger one
https://api.github.com/repos/huggingface/datasets/issues/5511/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
[]
null
completed
CONTRIBUTOR
2023-02-08T10:35:48Z
null
I_kwDODunzps5d7Zb4
[ "Update `datasets` or downgrade `huggingface-hub` ;)\r\n\r\nThe `huggingface-hub` lib did a breaking change a few months ago, and you're using an old version of `datasets` that does't support it", "Awesome thanks a lot! Everything works just fine with `datasets==2.9.0` :-) ", "Getting same error with latest ver...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5511/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5511
https://github.com/huggingface/datasets/issues/5511
false
1,575,191,549
https://api.github.com/repos/huggingface/datasets/issues/5510/labels{/name}
Signed-off-by: Filip Haltmayer <filip.haltmayer@zilliz.com>
2023-02-24T16:45:09Z
5,510
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-07T23:30:26Z
https://api.github.com/repos/huggingface/datasets/issues/5510/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5510/timeline
Milvus integration for search
https://api.github.com/repos/huggingface/datasets/issues/5510/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/81822489?v=4", "events_url": "https://api.github.com/users/filip-halt/events{/privacy}", "followers_url": "https://api.github.com/users/filip-halt/followers", "following_url": "https://api.github.com/users/filip-halt/following{/other_user}", "gists_url": "https://api.github.com/users/filip-halt/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/filip-halt", "id": 81822489, "login": "filip-halt", "node_id": "MDQ6VXNlcjgxODIyNDg5", "organizations_url": "https://api.github.com/users/filip-halt/orgs", "received_events_url": "https://api.github.com/users/filip-halt/received_events", "repos_url": "https://api.github.com/users/filip-halt/repos", "site_admin": false, "starred_url": "https://api.github.com/users/filip-halt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/filip-halt/subscriptions", "type": "User", "url": "https://api.github.com/users/filip-halt" }
[]
null
null
NONE
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/5510.diff", "html_url": "https://github.com/huggingface/datasets/pull/5510", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5510.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5510" }
PR_kwDODunzps5JehbR
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5510). All of your documentation changes will be reflected on that endpoint.", "To the maintainer, sorry about the repeated run requests for formatting. Missed the `make style` outlined in contributing guidelines. ", "Anythin...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5510/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5510
https://github.com/huggingface/datasets/pull/5510
true
1,574,177,320
https://api.github.com/repos/huggingface/datasets/issues/5509/labels{/name}
This adds a static `__all__` field to `__init__.py`, allowing typecheckers to know which symbols are accessible from `datasets` at runtime. In particular [Pyright](https://github.com/microsoft/pylance-release/issues/2328#issuecomment-1029381258) seems to rely on this. At this point I have added all (modulo oversight) the symbols mentioned in the Reference part of [the docs](https://huggingface.co/docs/datasets), but that could be adjusted. As a side effect, only these symbols will be imported by `from datasets import *`, which may or may not be a good thing (and if it isn't, that's easy to fix). Another option would be to add a pyi stub, but I think `__all__` should be the most pythonic solution. This should fix #3841.
2023-02-08T17:48:24Z
5,509
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-07T11:42:40Z
https://api.github.com/repos/huggingface/datasets/issues/5509/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5509/timeline
Add a static `__all__` to `__init__.py` for typecheckers
https://api.github.com/repos/huggingface/datasets/issues/5509/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/14248012?v=4", "events_url": "https://api.github.com/users/LoicGrobol/events{/privacy}", "followers_url": "https://api.github.com/users/LoicGrobol/followers", "following_url": "https://api.github.com/users/LoicGrobol/following{/other_user}", "gists_url": "https://api.github.com/users/LoicGrobol/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/LoicGrobol", "id": 14248012, "login": "LoicGrobol", "node_id": "MDQ6VXNlcjE0MjQ4MDEy", "organizations_url": "https://api.github.com/users/LoicGrobol/orgs", "received_events_url": "https://api.github.com/users/LoicGrobol/received_events", "repos_url": "https://api.github.com/users/LoicGrobol/repos", "site_admin": false, "starred_url": "https://api.github.com/users/LoicGrobol/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LoicGrobol/subscriptions", "type": "User", "url": "https://api.github.com/users/LoicGrobol" }
[]
null
null
NONE
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/5509.diff", "html_url": "https://github.com/huggingface/datasets/pull/5509", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5509.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5509" }
PR_kwDODunzps5JbH-u
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5509). All of your documentation changes will be reflected on that endpoint.", "Hi! I've commented on the original issue to provide some context. Feel free to share your opinion there." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5509/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5509
https://github.com/huggingface/datasets/pull/5509
true
1,573,290,359
https://api.github.com/repos/huggingface/datasets/issues/5508/labels{/name}
### Describe the bug Saving a dataset after setting format to torch doesn't work, but only if filtering ### Steps to reproduce the bug ``` a = Dataset.from_dict({"b": [1, 2]}) a.set_format('torch') a.save_to_disk("test_save") # saves successfully a.filter(None).save_to_disk("test_save_filter") # does not >> [...] TypeError: Provided `function` which is applied to all elements of table returns a `dict` of types [<class 'torch.Tensor'>]. When using `batched=True`, make sure provided `function` returns a `dict` of types like `(<class 'list'>, <class 'numpy.ndarray'>)`. # note: skipping the format change to torch lets this work. ### Expected behavior Saving to work ### Environment info - `datasets` version: 2.4.0 - Platform: Linux-6.1.9-arch1-1-x86_64-with-glibc2.36 - Python version: 3.10.9 - PyArrow version: 9.0.0 - Pandas version: 1.4.4
2023-02-09T14:55:26Z
5,508
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-06T21:08:58Z
https://api.github.com/repos/huggingface/datasets/issues/5508/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5508/timeline
Saving a dataset after setting format to torch doesn't work, but only if filtering
https://api.github.com/repos/huggingface/datasets/issues/5508/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/13984157?v=4", "events_url": "https://api.github.com/users/joebhakim/events{/privacy}", "followers_url": "https://api.github.com/users/joebhakim/followers", "following_url": "https://api.github.com/users/joebhakim/following{/other_user}", "gists_url": "https://api.github.com/users/joebhakim/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/joebhakim", "id": 13984157, "login": "joebhakim", "node_id": "MDQ6VXNlcjEzOTg0MTU3", "organizations_url": "https://api.github.com/users/joebhakim/orgs", "received_events_url": "https://api.github.com/users/joebhakim/received_events", "repos_url": "https://api.github.com/users/joebhakim/repos", "site_admin": false, "starred_url": "https://api.github.com/users/joebhakim/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joebhakim/subscriptions", "type": "User", "url": "https://api.github.com/users/joebhakim" }
[]
null
completed
NONE
2023-02-09T14:55:26Z
null
I_kwDODunzps5dxoF3
[ "Hey, I'm a research engineer working on language modelling wanting to contribute to open source. I was wondering if I could give it a shot?", "Hi! This issue was fixed in https://github.com/huggingface/datasets/pull/4972, so please install `datasets>=2.5.0` to avoid it." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5508/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5508
https://github.com/huggingface/datasets/issues/5508
false
1,572,667,036
https://api.github.com/repos/huggingface/datasets/issues/5507/labels{/name}
_Originally [posted](https://huggingface.slack.com/archives/C02V51Q3800/p1675443873878489?thread_ts=1675418893.373479&cid=C02V51Q3800) on Slack_ Considering all this, perhaps for Datasets 3.0, we can do the following: * [ ] have `continuous=True` by default in `.shard` (requested in the survey and makes more sense for us since it doesn't create an indices mapping) * [x] allow calling `save_to_disk` on "unflattened" datasets * [ ] remove "hidden" expensive calls in `save_to_disk`, `unique`, `concatenate_datasets`, etc. For instance, instead of silently calling `flatten_indices` where it's needed, it's probably better to be explicit (considering how expensive these ops can be) and raise an error instead
2023-02-28T18:19:18Z
5,507
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-02-06T14:25:55Z
https://api.github.com/repos/huggingface/datasets/issues/5507/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
https://api.github.com/repos/huggingface/datasets/issues/5507/timeline
Optimise behaviour in respect to indices mapping
https://api.github.com/repos/huggingface/datasets/issues/5507/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", ...
null
null
CONTRIBUTOR
null
null
I_kwDODunzps5dvP6c
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5507/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5507
https://github.com/huggingface/datasets/issues/5507
false
1,571,838,641
https://api.github.com/repos/huggingface/datasets/issues/5506/labels{/name}
### Describe the bug I am training a Roberta model using 2 GPUs and the `Trainer` API with a batch size of 256. Initially I used a standard `Dataset`, but had issues with slow data loading. After reading [this issue](https://github.com/huggingface/datasets/issues/2252), I swapped to loading my dataset as contiguous shards and passing those to an `IterableDataset`. I observed an unexpected drop in GPU memory utilization, and found the batch size returned from the model had been cut in half. When using `Trainer` with 2 GPUs and a batch size of 256, `Dataset` returns a batch of size 512 (256 per GPU), while `IterableDataset` returns a batch size of 256 (256 total). My guess is `IterableDataset` isn't accounting for multiple cards. ### Steps to reproduce the bug ```python import datasets from datasets import IterableDataset from transformers import RobertaConfig from transformers import RobertaTokenizerFast from transformers import RobertaForMaskedLM from transformers import DataCollatorForLanguageModeling from transformers import Trainer, TrainingArguments use_iterable_dataset = True def gen_from_shards(shards): for shard in shards: for example in shard: yield example dataset = datasets.load_from_disk('my_dataset.hf') if use_iterable_dataset: n_shards = 100 shards = [dataset.shard(num_shards=n_shards, index=i) for i in range(n_shards)] dataset = IterableDataset.from_generator(gen_from_shards, gen_kwargs={"shards": shards}) tokenizer = RobertaTokenizerFast.from_pretrained("./my_tokenizer", max_len=160, use_fast=True) config = RobertaConfig( vocab_size=8248, max_position_embeddings=256, num_attention_heads=8, num_hidden_layers=6, type_vocab_size=1) model = RobertaForMaskedLM(config=config) data_collator = DataCollatorForLanguageModeling(tokenizer=tokenizer, mlm=True, mlm_probability=0.15) training_args = TrainingArguments( per_device_train_batch_size=256 # other args removed for brevity ) trainer = Trainer( model=model, args=training_args, data_collator=data_collator, train_dataset=dataset, ) trainer.train() ``` ### Expected behavior Expected `Dataset` and `IterableDataset` to have the same batch size behavior. If the current behavior is intentional, the batch size printout at the start of training should be updated. Currently, both dataset classes result in `Trainer` printing the same total batch size, even though the batch size sent to the GPUs are different. ### Environment info datasets 2.7.1 transformers 4.25.1
2023-02-08T18:30:08Z
5,506
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-06T03:26:03Z
https://api.github.com/repos/huggingface/datasets/issues/5506/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5506/timeline
IterableDataset and Dataset return different batch sizes when using Trainer with multiple GPUs
https://api.github.com/repos/huggingface/datasets/issues/5506/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/38166299?v=4", "events_url": "https://api.github.com/users/kheyer/events{/privacy}", "followers_url": "https://api.github.com/users/kheyer/followers", "following_url": "https://api.github.com/users/kheyer/following{/other_user}", "gists_url": "https://api.github.com/users/kheyer/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kheyer", "id": 38166299, "login": "kheyer", "node_id": "MDQ6VXNlcjM4MTY2Mjk5", "organizations_url": "https://api.github.com/users/kheyer/orgs", "received_events_url": "https://api.github.com/users/kheyer/received_events", "repos_url": "https://api.github.com/users/kheyer/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kheyer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kheyer/subscriptions", "type": "User", "url": "https://api.github.com/users/kheyer" }
[]
null
completed
NONE
2023-02-08T18:30:07Z
null
I_kwDODunzps5dsFqx
[ "Hi ! `datasets` doesn't do batching - the PyTorch DataLoader does and is created by the `Trainer`. Do you pass other arguments to training_args with respect to data loading ?\r\n\r\nAlso we recently released `.to_iterable_dataset` that does pretty much what you implemented, but using contiguous shards to get a bet...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5506/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5506
https://github.com/huggingface/datasets/issues/5506
false
1,571,720,814
https://api.github.com/repos/huggingface/datasets/issues/5505/labels{/name}
### Describe the bug In [the docs here](https://huggingface.co/docs/datasets/use_with_pytorch#use-a-batchsampler), it mentions the issue of the Dataset being read one-by-one, then states that using a BatchSampler resolves the issue. I'm not sure if this is a mistake in the docs or the code, but it seems that the only way for a Dataset to be passed a list of indexes by PyTorch (instead of one index at a time) is to define a `__getitems__` method (note the plural) on the Dataset object, and since the HF Dataset doesn't have this, PyTorch executes [this line of code](https://github.com/pytorch/pytorch/blob/master/torch/utils/data/_utils/fetch.py#L58), reverting to fetching one-by-one. ### Steps to reproduce the bug You can put a breakpoint in `Dataset.__getitem__()` or just print the args from there and see that it's called multiple times for a single `next(iter(dataloader))`, even when using the code from the docs: ```py from torch.utils.data.sampler import BatchSampler, RandomSampler batch_sampler = BatchSampler(RandomSampler(ds), batch_size=32, drop_last=False) dataloader = DataLoader(ds, batch_sampler=batch_sampler) ``` ### Expected behavior The expected behaviour would be for it to fetch batches from the dataset, rather than one-by-one. To demonstrate that there is room for improvement: once I have a HF dataset `ds`, if I just add this line: ```py ds.__getitems__ = ds.__getitem__ ``` ...then the time taken to loop over the dataset improves considerably (for wikitext-103, from one minute to 13 seconds with batch size 32). Probably not a big deal in the grand scheme of things, but seems like an easy win. ### Environment info - `datasets` version: 2.9.0 - Platform: Linux-5.10.102.1-microsoft-standard-WSL2-x86_64-with-glibc2.31 - Python version: 3.10.8 - PyArrow version: 10.0.1 - Pandas version: 1.5.3
2023-02-19T18:27:30Z
5,505
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-06T01:14:55Z
https://api.github.com/repos/huggingface/datasets/issues/5505/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5505/timeline
PyTorch BatchSampler still loads from Dataset one-by-one
https://api.github.com/repos/huggingface/datasets/issues/5505/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/4443482?v=4", "events_url": "https://api.github.com/users/davidgilbertson/events{/privacy}", "followers_url": "https://api.github.com/users/davidgilbertson/followers", "following_url": "https://api.github.com/users/davidgilbertson/following{/other_user}", "gists_url": "https://api.github.com/users/davidgilbertson/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/davidgilbertson", "id": 4443482, "login": "davidgilbertson", "node_id": "MDQ6VXNlcjQ0NDM0ODI=", "organizations_url": "https://api.github.com/users/davidgilbertson/orgs", "received_events_url": "https://api.github.com/users/davidgilbertson/received_events", "repos_url": "https://api.github.com/users/davidgilbertson/repos", "site_admin": false, "starred_url": "https://api.github.com/users/davidgilbertson/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/davidgilbertson/subscriptions", "type": "User", "url": "https://api.github.com/users/davidgilbertson" }
[]
null
completed
NONE
2023-02-19T18:27:30Z
null
I_kwDODunzps5dro5u
[ "This change seems to come from a few months ago in the PyTorch side. That's good news and it means we may not need to pass a batch_sampler as soon as we add `Dataset.__getitems__` to get the optimal speed :)\r\n\r\nThanks for reporting ! Would you like to open a PR to add `__getitems__` and remove this outdated do...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5505/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5505
https://github.com/huggingface/datasets/issues/5505
false
1,570,621,242
https://api.github.com/repos/huggingface/datasets/issues/5504/labels{/name}
Fixes https://github.com/huggingface/datasets/issues/5495 I'm not sure whether we prefer a test here or if timestamps are known to be unsupported (like booleans). The current test at least covers the bug
2023-02-08T17:28:50Z
5,504
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-03T23:39:04Z
https://api.github.com/repos/huggingface/datasets/issues/5504/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5504/timeline
don't zero copy timestamps
https://api.github.com/repos/huggingface/datasets/issues/5504/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/2512762?v=4", "events_url": "https://api.github.com/users/dwyatte/events{/privacy}", "followers_url": "https://api.github.com/users/dwyatte/followers", "following_url": "https://api.github.com/users/dwyatte/following{/other_user}", "gists_url": "https://api.github.com/users/dwyatte/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dwyatte", "id": 2512762, "login": "dwyatte", "node_id": "MDQ6VXNlcjI1MTI3NjI=", "organizations_url": "https://api.github.com/users/dwyatte/orgs", "received_events_url": "https://api.github.com/users/dwyatte/received_events", "repos_url": "https://api.github.com/users/dwyatte/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dwyatte/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dwyatte/subscriptions", "type": "User", "url": "https://api.github.com/users/dwyatte" }
[]
null
null
CONTRIBUTOR
2023-02-08T14:33:17Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5504.diff", "html_url": "https://github.com/huggingface/datasets/pull/5504", "merged_at": "2023-02-08T14:33:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/5504.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5504" }
PR_kwDODunzps5JPoWy
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5504/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5504
https://github.com/huggingface/datasets/pull/5504
true
1,570,091,225
https://api.github.com/repos/huggingface/datasets/issues/5502/labels{/name}
Added functionality implementation: sort datasets by multiple keys/columns as discussed in https://github.com/huggingface/datasets/issues/5425.
2023-02-21T14:46:49Z
5,502
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-03T16:17:00Z
https://api.github.com/repos/huggingface/datasets/issues/5502/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5502/timeline
Added functionality: sort datasets by multiple keys
https://api.github.com/repos/huggingface/datasets/issues/5502/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7805682?v=4", "events_url": "https://api.github.com/users/MichlF/events{/privacy}", "followers_url": "https://api.github.com/users/MichlF/followers", "following_url": "https://api.github.com/users/MichlF/following{/other_user}", "gists_url": "https://api.github.com/users/MichlF/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/MichlF", "id": 7805682, "login": "MichlF", "node_id": "MDQ6VXNlcjc4MDU2ODI=", "organizations_url": "https://api.github.com/users/MichlF/orgs", "received_events_url": "https://api.github.com/users/MichlF/received_events", "repos_url": "https://api.github.com/users/MichlF/repos", "site_admin": false, "starred_url": "https://api.github.com/users/MichlF/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MichlF/subscriptions", "type": "User", "url": "https://api.github.com/users/MichlF" }
[]
null
null
CONTRIBUTOR
2023-02-21T14:39:23Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5502.diff", "html_url": "https://github.com/huggingface/datasets/pull/5502", "merged_at": "2023-02-21T14:39:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/5502.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5502" }
PR_kwDODunzps5JN0aX
[ "_The documentation is not available anymore as the PR was closed or merged._", "> Thanks! I've left some comments.\r\n> \r\n> We should also add some tests, mainly to make sure `reverse` behaves as expected. Let me know if you need help with that.\r\n\r\nThanks for the offer! I couldn't find any guidelines on ho...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5502/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5502
https://github.com/huggingface/datasets/pull/5502
true
1,569,644,159
https://api.github.com/repos/huggingface/datasets/issues/5501/labels{/name}
Original fix: https://github.com/huggingface/huggingface_hub/pull/1267 Not sure this function is actually still called though. I haven't done benches on this. Is there a dataset where files are hosted on the hub through cloudfront so we can have the same setup as in `hf_hub` ?
2023-02-09T11:04:11Z
5,501
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-03T10:50:10Z
https://api.github.com/repos/huggingface/datasets/issues/5501/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5501/timeline
Increase chunk size for speeding up file downloads
https://api.github.com/repos/huggingface/datasets/issues/5501/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4", "events_url": "https://api.github.com/users/Narsil/events{/privacy}", "followers_url": "https://api.github.com/users/Narsil/followers", "following_url": "https://api.github.com/users/Narsil/following{/other_user}", "gists_url": "https://api.github.com/users/Narsil/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Narsil", "id": 204321, "login": "Narsil", "node_id": "MDQ6VXNlcjIwNDMyMQ==", "organizations_url": "https://api.github.com/users/Narsil/orgs", "received_events_url": "https://api.github.com/users/Narsil/received_events", "repos_url": "https://api.github.com/users/Narsil/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Narsil/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Narsil/subscriptions", "type": "User", "url": "https://api.github.com/users/Narsil" }
[]
null
null
CONTRIBUTOR
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/5501.diff", "html_url": "https://github.com/huggingface/datasets/pull/5501", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5501.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5501" }
PR_kwDODunzps5JMTn8
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5501). All of your documentation changes will be reflected on that endpoint.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5501/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5501
https://github.com/huggingface/datasets/pull/5501
true
1,569,257,240
https://api.github.com/repos/huggingface/datasets/issues/5500/labels{/name}
### Describe the bug I use the following scripts to download data from WMT19: ```python import datasets from datasets import inspect_dataset, load_dataset_builder from wmt19.wmt_utils import _TRAIN_SUBSETS,_DEV_SUBSETS ## this is a must due to: https://discuss.huggingface.co/t/load-dataset-hangs-with-local-files/28034/3 if __name__ == '__main__': dev_subsets,train_subsets = [],[] for subset in _TRAIN_SUBSETS: if subset.target=='en' and 'de' in subset.sources: train_subsets.append(subset.name) for subset in _DEV_SUBSETS: if subset.target=='en' and 'de' in subset.sources: dev_subsets.append(subset.name) inspect_dataset("wmt19", "./wmt19") builder = load_dataset_builder( "./wmt19/wmt_utils.py", language_pair=("de", "en"), subsets={ datasets.Split.TRAIN: train_subsets, datasets.Split.VALIDATION: dev_subsets, }, ) builder.download_and_prepare() ds = builder.as_dataset() ds.to_json("../data/wmt19/ende/data.json") ``` And I got the following error: ``` Traceback (most recent call last): | 0/2 [00:00<?, ?obj/s] File "draft.py", line 26, in <module> builder.download_and_prepare() | 0/1 [00:00<?, ?obj/s] File "/Users/hannibal046/anaconda3/lib/python3.8/site-packages/datasets/builder.py", line 605, in download_and_prepare self._download_and_prepare(%| | 0/1 [00:00<?, ?obj/s] File "/Users/hannibal046/anaconda3/lib/python3.8/site-packages/datasets/builder.py", line 1104, in _download_and_prepare super()._download_and_prepare(dl_manager, verify_infos, check_duplicate_keys=verify_infos) | 0/1 [00:00<?, ?obj/s] File "/Users/hannibal046/anaconda3/lib/python3.8/site-packages/datasets/builder.py", line 676, in _download_and_prepare verify_checksums(s #13: 0%| | 0/1 [00:00<?, ?obj/s] File "/Users/hannibal046/anaconda3/lib/python3.8/site-packages/datasets/utils/info_utils.py", line 35, in verify_checksums raise UnexpectedDownloadedFile(str(set(recorded_checksums) - set(expected_checksums))) | 0/1 [00:00<?, ?obj/s] datasets.utils.info_utils.UnexpectedDownloadedFile: {'https://s3.amazonaws.com/web-language-models/paracrawl/release1/paracrawl-release1.en-de.zipporah0-dedup-clean.tgz', 'https://huggingface.co/datasets/wmt/wmt13/resolve/main-zip/training-parallel-europarl-v7.zip', 'https://huggingface.co/datasets/wmt/wmt18/resolve/main-zip/translation-task/rapid2016.zip', 'https://huggingface.co/datasets/wmt/wmt18/resolve/main-zip/translation-task/training-parallel-nc-v13.zip', 'https://huggingface.co/datasets/wmt/wmt17/resolve/main-zip/translation-task/training-parallel-nc-v12.zip', 'https://huggingface.co/datasets/wmt/wmt14/resolve/main-zip/training-parallel-nc-v9.zip', 'https://huggingface.co/datasets/wmt/wmt15/resolve/main-zip/training-parallel-nc-v10.zip', 'https://huggingface.co/datasets/wmt/wmt16/resolve/main-zip/translation-task/training-parallel-nc-v11.zip'} ``` ### Steps to reproduce the bug see above ### Expected behavior download data successfully ### Environment info datasets==2.1.0 python==3.8
2023-02-03T05:52:56Z
5,500
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-03T05:45:37Z
https://api.github.com/repos/huggingface/datasets/issues/5500/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5500/timeline
WMT19 custom download checksum error
https://api.github.com/repos/huggingface/datasets/issues/5500/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/38466901?v=4", "events_url": "https://api.github.com/users/Hannibal046/events{/privacy}", "followers_url": "https://api.github.com/users/Hannibal046/followers", "following_url": "https://api.github.com/users/Hannibal046/following{/other_user}", "gists_url": "https://api.github.com/users/Hannibal046/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Hannibal046", "id": 38466901, "login": "Hannibal046", "node_id": "MDQ6VXNlcjM4NDY2OTAx", "organizations_url": "https://api.github.com/users/Hannibal046/orgs", "received_events_url": "https://api.github.com/users/Hannibal046/received_events", "repos_url": "https://api.github.com/users/Hannibal046/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Hannibal046/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Hannibal046/subscriptions", "type": "User", "url": "https://api.github.com/users/Hannibal046" }
[]
null
completed
NONE
2023-02-03T05:52:56Z
null
I_kwDODunzps5diPcY
[ "I update the `datatsets` version and it works." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5500/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5500
https://github.com/huggingface/datasets/issues/5500
false
1,568,937,026
https://api.github.com/repos/huggingface/datasets/issues/5499/labels{/name}
### Feature request When loading a dataset that has been cached locally, the `load_dataset` function takes a lot longer than it should take to fetch the dataset from disk (or memory). This is particularly noticeable for smaller datasets. For example, wikitext-2, comparing `load_data` (once cached) and `load_from_disk`, the `load_dataset` method takes 40 times longer. ⏱ 4.84s ⮜ load_dataset ⏱ 119ms ⮜ load_from_disk ### Motivation I assume this is doing something like checking for a newer version. If so, that's an age old problem: do you make the user wait _every single time they load from cache_ or do you do something like load from cache always, _then_ check for a newer version and alert if they have stale data. The decision usually revolves around what percentage of the time the data will have been updated, and how dangerous old data is. For most datasets it's extremely unlikely that there will be a newer version on any given run, so 99% of the time this is just wasted time. Maybe you don't want to make that decision for all users, but at least having the _option_ to not wait for checks would be an improvement. ### Your contribution .
2023-02-07T19:35:11Z
5,499
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-02-02T23:34:50Z
https://api.github.com/repos/huggingface/datasets/issues/5499/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5499/timeline
`load_dataset` has ~4 seconds of overhead for cached data
https://api.github.com/repos/huggingface/datasets/issues/5499/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/4443482?v=4", "events_url": "https://api.github.com/users/davidgilbertson/events{/privacy}", "followers_url": "https://api.github.com/users/davidgilbertson/followers", "following_url": "https://api.github.com/users/davidgilbertson/following{/other_user}", "gists_url": "https://api.github.com/users/davidgilbertson/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/davidgilbertson", "id": 4443482, "login": "davidgilbertson", "node_id": "MDQ6VXNlcjQ0NDM0ODI=", "organizations_url": "https://api.github.com/users/davidgilbertson/orgs", "received_events_url": "https://api.github.com/users/davidgilbertson/received_events", "repos_url": "https://api.github.com/users/davidgilbertson/repos", "site_admin": false, "starred_url": "https://api.github.com/users/davidgilbertson/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/davidgilbertson/subscriptions", "type": "User", "url": "https://api.github.com/users/davidgilbertson" }
[]
null
null
NONE
null
null
I_kwDODunzps5dhBRC
[ "Hi ! To skip the verification step that checks if newer data exist, you can enable offline mode with `HF_DATASETS_OFFLINE=1`.\r\n\r\nAlthough I agree this step should be much faster for datasets hosted on the HF Hub - we could just compare the commit hash from the local data and the remote git repository. We're no...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5499/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5499
https://github.com/huggingface/datasets/issues/5499
false
1,568,190,529
https://api.github.com/repos/huggingface/datasets/issues/5498/labels{/name}
### Describe the bug Hi, Thanks for the amazing work on the library! **Describe the bug** I think I might have noticed a small bug in the filter method. Having loaded a dataset using `load_dataset`, when I try to filter out empty entries with `batched=True`, I get a TypeError. ### Steps to reproduce the bug ``` train_dataset = train_dataset.filter( function=lambda example: example["image"] is not None, batched=True, batch_size=10) ``` Error message: ``` File .../lib/python3.9/site-packages/datasets/fingerprint.py:480, in fingerprint_transform.<locals>._fingerprint.<locals>.wrapper(*args, **kwargs) 476 validate_fingerprint(kwargs[fingerprint_name]) 478 # Call actual function --> 480 out = func(self, *args, **kwargs) ... -> 5666 indices_array = [i for i, to_keep in zip(indices, mask) if to_keep] 5667 if indices_mapping is not None: 5668 indices_array = pa.array(indices_array, type=pa.uint64()) TypeError: 'bool' object is not iterable ``` **Removing batched=True allows to bypass the issue.** ### Expected behavior According to the doc, "[batch_size corresponds to the] number of examples per batch provided to function if batched = True", so we shouldn't need to remove the batchd=True arg? source: https://huggingface.co/docs/datasets/v2.9.0/en/package_reference/main_classes#datasets.Dataset.filter ### Environment info - `datasets` version: 2.9.0 - Platform: Linux-5.4.0-122-generic-x86_64-with-glibc2.31 - Python version: 3.9.10 - PyArrow version: 10.0.1 - Pandas version: 1.5.3
2023-10-08T06:12:47Z
5,498
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-02-02T14:46:49Z
https://api.github.com/repos/huggingface/datasets/issues/5498/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5498/timeline
TypeError: 'bool' object is not iterable when filtering a datasets.arrow_dataset.Dataset
https://api.github.com/repos/huggingface/datasets/issues/5498/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/91255010?v=4", "events_url": "https://api.github.com/users/vmuel/events{/privacy}", "followers_url": "https://api.github.com/users/vmuel/followers", "following_url": "https://api.github.com/users/vmuel/following{/other_user}", "gists_url": "https://api.github.com/users/vmuel/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/vmuel", "id": 91255010, "login": "vmuel", "node_id": "MDQ6VXNlcjkxMjU1MDEw", "organizations_url": "https://api.github.com/users/vmuel/orgs", "received_events_url": "https://api.github.com/users/vmuel/received_events", "repos_url": "https://api.github.com/users/vmuel/repos", "site_admin": false, "starred_url": "https://api.github.com/users/vmuel/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vmuel/subscriptions", "type": "User", "url": "https://api.github.com/users/vmuel" }
[]
null
completed
NONE
2023-02-04T17:19:36Z
null
I_kwDODunzps5deLBB
[ "Hi! Instead of a single boolean, your filter function should return an iterable (of booleans) in the batched mode like so:\r\n```python\r\ntrain_dataset = train_dataset.filter(\r\n function=lambda batch: [image is not None for image in batch[\"image\"]], \r\n batched=True,\r\n batc...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5498/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5498
https://github.com/huggingface/datasets/issues/5498
false