id
int64
1.14B
2.23B
labels_url
stringlengths
75
75
body
stringlengths
2
33.9k
βŒ€
updated_at
stringlengths
20
20
number
int64
3.76k
6.79k
milestone
dict
repository_url
stringclasses
1 value
draft
bool
2 classes
labels
listlengths
0
4
created_at
stringlengths
20
20
comments_url
stringlengths
70
70
assignee
dict
timeline_url
stringlengths
70
70
title
stringlengths
1
290
events_url
stringlengths
68
68
active_lock_reason
null
user
dict
assignees
listlengths
0
3
performed_via_github_app
null
state_reason
stringclasses
3 values
author_association
stringclasses
3 values
closed_at
stringlengths
20
20
βŒ€
pull_request
dict
node_id
stringlengths
18
19
comments
listlengths
0
30
reactions
dict
state
stringclasses
2 values
locked
bool
1 class
url
stringlengths
61
61
html_url
stringlengths
49
51
is_pull_request
bool
2 classes
1,567,601,264
https://api.github.com/repos/huggingface/datasets/issues/5497/labels{/name}
Using `use_auth_token=True` is not needed anymore. If a user logged in, the token will be automatically retrieved. Also include a mention for gated repos See https://github.com/huggingface/huggingface_hub/pull/1064
2023-02-02T11:26:08Z
5,497
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-02T08:56:15Z
https://api.github.com/repos/huggingface/datasets/issues/5497/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5497/timeline
Improved error message for gated/private repos
https://api.github.com/repos/huggingface/datasets/issues/5497/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7246357?v=4", "events_url": "https://api.github.com/users/osanseviero/events{/privacy}", "followers_url": "https://api.github.com/users/osanseviero/followers", "following_url": "https://api.github.com/users/osanseviero/following{/other_user}", "gists_url": "https://api.github.com/users/osanseviero/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/osanseviero", "id": 7246357, "login": "osanseviero", "node_id": "MDQ6VXNlcjcyNDYzNTc=", "organizations_url": "https://api.github.com/users/osanseviero/orgs", "received_events_url": "https://api.github.com/users/osanseviero/received_events", "repos_url": "https://api.github.com/users/osanseviero/repos", "site_admin": false, "starred_url": "https://api.github.com/users/osanseviero/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/osanseviero/subscriptions", "type": "User", "url": "https://api.github.com/users/osanseviero" }
[]
null
null
MEMBER
2023-02-02T11:17:15Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5497.diff", "html_url": "https://github.com/huggingface/datasets/pull/5497", "merged_at": "2023-02-02T11:17:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/5497.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5497" }
PR_kwDODunzps5JFhvc
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5497/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5497
https://github.com/huggingface/datasets/pull/5497
true
1,567,301,765
https://api.github.com/repos/huggingface/datasets/issues/5496/labels{/name}
### Feature request Right now the `Dataset` class implements `map()` and `filter()`, but leaves out the third functional idiom popular among Python users: `reduce`. ### Motivation A `reduce` method is often useful when calculating dataset statistics, for example, the occurrence of a particular n-gram or the average line length of a code dataset. ### Your contribution I haven't contributed to `datasets` before, but I don't expect this will be too difficult, since the implementation will closely follow that of `map` and `filter`. I could have a crack over the weekend.
2023-07-21T14:24:32Z
5,496
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-02-02T04:30:22Z
https://api.github.com/repos/huggingface/datasets/issues/5496/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5496/timeline
Add a `reduce` method
https://api.github.com/repos/huggingface/datasets/issues/5496/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59542043?v=4", "events_url": "https://api.github.com/users/zhangir-azerbayev/events{/privacy}", "followers_url": "https://api.github.com/users/zhangir-azerbayev/followers", "following_url": "https://api.github.com/users/zhangir-azerbayev/following{/other_user}", "gists_url": "https://api.github.com/users/zhangir-azerbayev/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zhangir-azerbayev", "id": 59542043, "login": "zhangir-azerbayev", "node_id": "MDQ6VXNlcjU5NTQyMDQz", "organizations_url": "https://api.github.com/users/zhangir-azerbayev/orgs", "received_events_url": "https://api.github.com/users/zhangir-azerbayev/received_events", "repos_url": "https://api.github.com/users/zhangir-azerbayev/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zhangir-azerbayev/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zhangir-azerbayev/subscriptions", "type": "User", "url": "https://api.github.com/users/zhangir-azerbayev" }
[]
null
completed
NONE
2023-07-21T14:24:32Z
null
I_kwDODunzps5dayCF
[ "Hi! Sure, feel free to open a PR, so we can see the API you have in mind.", "I would like to give it a go! #self-assign", "Closing as `Dataset.map` can be used instead (see https://github.com/huggingface/datasets/pull/5533#issuecomment-1440571658 and https://github.com/huggingface/datasets/pull/5533#issuecomme...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5496/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5496
https://github.com/huggingface/datasets/issues/5496
false
1,566,803,452
https://api.github.com/repos/huggingface/datasets/issues/5495/labels{/name}
### Describe the bug There appears to be some eager behavior in `to_tf_dataset` that runs against every column in a dataset even if they aren't included in the columns argument. This is problematic with datetime UTC columns due to them not working with zero copy. If I don't have UTC information in my datetime column, then everything works as expected. ### Steps to reproduce the bug ```python import numpy as np import pandas as pd from datasets import Dataset df = pd.DataFrame(np.random.rand(2, 1), columns=["x"]) # df["dt"] = pd.to_datetime(["2023-01-01", "2023-01-01"]) # works fine df["dt"] = pd.to_datetime(["2023-01-01 00:00:00.00000+00:00", "2023-01-01 00:00:00.00000+00:00"]) df.to_parquet("test.pq") ds = Dataset.from_parquet("test.pq") tf_ds = ds.to_tf_dataset(columns=["x"], batch_size=2, shuffle=True) ``` ``` ArrowInvalid Traceback (most recent call last) Cell In[1], line 12 8 df.to_parquet("test.pq") 11 ds = Dataset.from_parquet("test.pq") ---> 12 tf_ds = ds.to_tf_dataset(columns=["r"], batch_size=2, shuffle=True) File ~/venv/lib/python3.8/site-packages/datasets/arrow_dataset.py:411, in TensorflowDatasetMixin.to_tf_dataset(self, batch_size, columns, shuffle, collate_fn, drop_remainder, collate_fn_args, label_cols, prefetch, num_workers) 407 dataset = self 409 # TODO(Matt, QL): deprecate the retention of label_ids and label --> 411 output_signature, columns_to_np_types = dataset._get_output_signature( 412 dataset, 413 collate_fn=collate_fn, 414 collate_fn_args=collate_fn_args, 415 cols_to_retain=cols_to_retain, 416 batch_size=batch_size if drop_remainder else None, 417 ) 419 if "labels" in output_signature: 420 if ("label_ids" in columns or "label" in columns) and "labels" not in columns: File ~/venv/lib/python3.8/site-packages/datasets/arrow_dataset.py:254, in TensorflowDatasetMixin._get_output_signature(dataset, collate_fn, collate_fn_args, cols_to_retain, batch_size, num_test_batches) 252 for _ in range(num_test_batches): 253 indices = sample(range(len(dataset)), test_batch_size) --> 254 test_batch = dataset[indices] 255 if cols_to_retain is not None: 256 test_batch = {key: value for key, value in test_batch.items() if key in cols_to_retain} File ~/venv/lib/python3.8/site-packages/datasets/arrow_dataset.py:2590, in Dataset.__getitem__(self, key) 2588 def __getitem__(self, key): # noqa: F811 2589 """Can be used to index columns (by string names) or rows (by integer index or iterable of indices or bools).""" -> 2590 return self._getitem( 2591 key, 2592 ) File ~/venv/lib/python3.8/site-packages/datasets/arrow_dataset.py:2575, in Dataset._getitem(self, key, **kwargs) 2573 formatter = get_formatter(format_type, features=self.features, **format_kwargs) 2574 pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None) -> 2575 formatted_output = format_table( 2576 pa_subtable, key, formatter=formatter, format_columns=format_columns, output_all_columns=output_all_columns 2577 ) 2578 return formatted_output File ~/venv/lib/python3.8/site-packages/datasets/formatting/formatting.py:634, in format_table(table, key, formatter, format_columns, output_all_columns) 632 python_formatter = PythonFormatter(features=None) 633 if format_columns is None: --> 634 return formatter(pa_table, query_type=query_type) 635 elif query_type == "column": 636 if key in format_columns: File ~/venv/lib/python3.8/site-packages/datasets/formatting/formatting.py:410, in Formatter.__call__(self, pa_table, query_type) 408 return self.format_column(pa_table) 409 elif query_type == "batch": --> 410 return self.format_batch(pa_table) File ~/venv/lib/python3.8/site-packages/datasets/formatting/np_formatter.py:78, in NumpyFormatter.format_batch(self, pa_table) 77 def format_batch(self, pa_table: pa.Table) -> Mapping: ---> 78 batch = self.numpy_arrow_extractor().extract_batch(pa_table) 79 batch = self.python_features_decoder.decode_batch(batch) 80 batch = self.recursive_tensorize(batch) File ~/venv/lib/python3.8/site-packages/datasets/formatting/formatting.py:164, in NumpyArrowExtractor.extract_batch(self, pa_table) 163 def extract_batch(self, pa_table: pa.Table) -> dict: --> 164 return {col: self._arrow_array_to_numpy(pa_table[col]) for col in pa_table.column_names} File ~/venv/lib/python3.8/site-packages/datasets/formatting/formatting.py:164, in <dictcomp>(.0) 163 def extract_batch(self, pa_table: pa.Table) -> dict: --> 164 return {col: self._arrow_array_to_numpy(pa_table[col]) for col in pa_table.column_names} File ~/venv/lib/python3.8/site-packages/datasets/formatting/formatting.py:185, in NumpyArrowExtractor._arrow_array_to_numpy(self, pa_array) 181 else: 182 zero_copy_only = _is_zero_copy_only(pa_array.type) and all( 183 not _is_array_with_nulls(chunk) for chunk in pa_array.chunks 184 ) --> 185 array: List = [ 186 row for chunk in pa_array.chunks for row in chunk.to_numpy(zero_copy_only=zero_copy_only) 187 ] 188 else: 189 if isinstance(pa_array.type, _ArrayXDExtensionType): 190 # don't call to_pylist() to preserve dtype of the fixed-size array File ~/venv/lib/python3.8/site-packages/datasets/formatting/formatting.py:186, in <listcomp>(.0) 181 else: 182 zero_copy_only = _is_zero_copy_only(pa_array.type) and all( 183 not _is_array_with_nulls(chunk) for chunk in pa_array.chunks 184 ) 185 array: List = [ --> 186 row for chunk in pa_array.chunks for row in chunk.to_numpy(zero_copy_only=zero_copy_only) 187 ] 188 else: 189 if isinstance(pa_array.type, _ArrayXDExtensionType): 190 # don't call to_pylist() to preserve dtype of the fixed-size array File ~/venv/lib/python3.8/site-packages/pyarrow/array.pxi:1475, in pyarrow.lib.Array.to_numpy() File ~/venv/lib/python3.8/site-packages/pyarrow/error.pxi:100, in pyarrow.lib.check_status() ArrowInvalid: Needed to copy 1 chunks with 0 nulls, but zero_copy_only was True ``` ### Expected behavior I think there are two potential issues/fixes 1. Proper handling of datetime UTC columns (perhaps there is something incorrect with zero copy handling here) 2. Not eagerly running against every column in a dataset when the columns argument of `to_tf_dataset` specifies a subset of columns (although I'm not sure if this is unavoidable) ### Environment info - `datasets` version: 2.9.0 - Platform: macOS-13.2-x86_64-i386-64bit - Python version: 3.8.12 - PyArrow version: 11.0.0 - Pandas version: 1.5.3
2023-02-08T14:33:19Z
5,495
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "7057ff", "default": true, "descript...
2023-02-01T20:47:33Z
https://api.github.com/repos/huggingface/datasets/issues/5495/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5495/timeline
to_tf_dataset fails with datetime UTC columns even if not included in columns argument
https://api.github.com/repos/huggingface/datasets/issues/5495/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/2512762?v=4", "events_url": "https://api.github.com/users/dwyatte/events{/privacy}", "followers_url": "https://api.github.com/users/dwyatte/followers", "following_url": "https://api.github.com/users/dwyatte/following{/other_user}", "gists_url": "https://api.github.com/users/dwyatte/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dwyatte", "id": 2512762, "login": "dwyatte", "node_id": "MDQ6VXNlcjI1MTI3NjI=", "organizations_url": "https://api.github.com/users/dwyatte/orgs", "received_events_url": "https://api.github.com/users/dwyatte/received_events", "repos_url": "https://api.github.com/users/dwyatte/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dwyatte/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dwyatte/subscriptions", "type": "User", "url": "https://api.github.com/users/dwyatte" }
[]
null
completed
CONTRIBUTOR
2023-02-08T14:33:19Z
null
I_kwDODunzps5dY4X8
[ "Hi! This is indeed a bug in our zero-copy logic.\r\n\r\nTo fix it, instead of the line:\r\nhttps://github.com/huggingface/datasets/blob/7cfac43b980ab9e4a69c2328f085770996323005/src/datasets/features/features.py#L702\r\n\r\nwe should have:\r\n```python\r\nreturn pa.types.is_primitive(pa_type) and not (pa.types.is_b...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5495/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5495
https://github.com/huggingface/datasets/issues/5495
false
1,566,655,348
https://api.github.com/repos/huggingface/datasets/issues/5494/labels{/name}
Our [installation documentation page](https://huggingface.co/docs/datasets/installation#audio) says that one can use Datasets for mp3 only with `torchaudio<0.12`. `torchaudio>0.12` is actually supported too but requires a specific version of ffmpeg which is not easily installed on all linux versions but there is a custom ubuntu repo for it, we have insctructions in the code: https://github.com/huggingface/datasets/blob/main/src/datasets/features/audio.py#L327 So we should update the doc page. But first investigate [this issue](5488).
2023-03-02T16:08:17Z
5,494
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "0075ca", "default": true, "description": "Improvements or additions to documentation", "id": 1935892861, "name": "documentation", "node_id": "MDU6TGFiZWwxOTM1ODkyODYx", "url": "https://api.github.com/repos/huggingface/datasets/labels/documentation" } ]
2023-02-01T19:07:50Z
https://api.github.com/repos/huggingface/datasets/issues/5494/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5494/timeline
Update audio installation doc page
https://api.github.com/repos/huggingface/datasets/issues/5494/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
null
completed
CONTRIBUTOR
2023-03-02T16:08:17Z
null
I_kwDODunzps5dYUN0
[ "Totally agree, the docs should be in sync with our code.\r\n\r\nIndeed to avoid confusing users, I think we should have updated the docs at the same time as this PR:\r\n- #5167", "@albertvillanova yeah sure I should have, but I forgot back then, sorry for that 😢", "No, @polinaeterna, nothing to be sorry about...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 1, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/5494/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5494
https://github.com/huggingface/datasets/issues/5494
false
1,566,637,806
https://api.github.com/repos/huggingface/datasets/issues/5493/labels{/name}
null
2023-02-08T15:10:46Z
5,493
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-01T18:57:48Z
https://api.github.com/repos/huggingface/datasets/issues/5493/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5493/timeline
Remove unused `load_from_cache_file` arg from `Dataset.shard()` docstring
https://api.github.com/repos/huggingface/datasets/issues/5493/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
null
null
CONTRIBUTOR
2023-02-08T15:03:50Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5493.diff", "html_url": "https://github.com/huggingface/datasets/pull/5493", "merged_at": "2023-02-08T15:03:50Z", "patch_url": "https://github.com/huggingface/datasets/pull/5493.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5493" }
PR_kwDODunzps5JCSAZ
[ "_The documentation is not available anymore as the PR was closed or merged._", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5493). All of your documentation changes will be reflected on that endpoint.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5493/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5493
https://github.com/huggingface/datasets/pull/5493
true
1,566,604,216
https://api.github.com/repos/huggingface/datasets/issues/5492/labels{/name}
Right now `ds.push_to_hub()` can push a dataset on `main` or on a new branch with `branch=`, but there is no way to open a pull request. Even passing `branch=refs/pr/x` doesn't seem to work: it tries to create a branch with that name cc @nateraw It should be possible to tweak the use of `huggingface_hub` in `push_to_hub` to make it open a PR or push to an existing PR
2023-10-16T13:30:48Z
5,492
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "7057ff", "default": true...
2023-02-01T18:32:14Z
https://api.github.com/repos/huggingface/datasets/issues/5492/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4", "events_url": "https://api.github.com/users/nateraw/events{/privacy}", "followers_url": "https://api.github.com/users/nateraw/followers", "following_url": "https://api.github.com/users/nateraw/following{/other_user}", "gists_url": "https://api.github.com/users/nateraw/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nateraw", "id": 32437151, "login": "nateraw", "node_id": "MDQ6VXNlcjMyNDM3MTUx", "organizations_url": "https://api.github.com/users/nateraw/orgs", "received_events_url": "https://api.github.com/users/nateraw/received_events", "repos_url": "https://api.github.com/users/nateraw/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nateraw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nateraw/subscriptions", "type": "User", "url": "https://api.github.com/users/nateraw" }
https://api.github.com/repos/huggingface/datasets/issues/5492/timeline
Push_to_hub in a pull request
https://api.github.com/repos/huggingface/datasets/issues/5492/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/32437151?v=4", "events_url": "https://api.github.com/users/nateraw/events{/privacy}", "followers_url": "https://api.github.com/users/nateraw/followers", "following_url": "https://api.github.com/users/nateraw/following{/other_user}", "gists...
null
completed
MEMBER
2023-10-16T13:30:48Z
null
I_kwDODunzps5dYHu4
[ "Assigned to myself and will get to it in the next week, but if someone finds this issue annoying and wants to submit a PR before I do, just ping me here and I'll reassign :). ", "I would like to be assigned to this issue, @nateraw . #self-assign" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5492/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5492
https://github.com/huggingface/datasets/issues/5492
false
1,566,235,012
https://api.github.com/repos/huggingface/datasets/issues/5491/labels{/name}
null
2023-02-02T07:42:28Z
5,491
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-01T14:39:39Z
https://api.github.com/repos/huggingface/datasets/issues/5491/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5491/timeline
[MINOR] Typo
https://api.github.com/repos/huggingface/datasets/issues/5491/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3664563?v=4", "events_url": "https://api.github.com/users/cakiki/events{/privacy}", "followers_url": "https://api.github.com/users/cakiki/followers", "following_url": "https://api.github.com/users/cakiki/following{/other_user}", "gists_url": "https://api.github.com/users/cakiki/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cakiki", "id": 3664563, "login": "cakiki", "node_id": "MDQ6VXNlcjM2NjQ1NjM=", "organizations_url": "https://api.github.com/users/cakiki/orgs", "received_events_url": "https://api.github.com/users/cakiki/received_events", "repos_url": "https://api.github.com/users/cakiki/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cakiki/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cakiki/subscriptions", "type": "User", "url": "https://api.github.com/users/cakiki" }
[]
null
null
CONTRIBUTOR
2023-02-02T07:35:14Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5491.diff", "html_url": "https://github.com/huggingface/datasets/pull/5491", "merged_at": "2023-02-02T07:35:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/5491.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5491" }
PR_kwDODunzps5JA9OD
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5491/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5491
https://github.com/huggingface/datasets/pull/5491
true
1,565,842,327
https://api.github.com/repos/huggingface/datasets/issues/5490/labels{/name}
As pointed out by @merveenoyan, default behavior of `Dataset.to_csv` adds the index as an additional column without name. This PR changes the default behavior, so that now the index column is not written. To add the index column, now you need to pass `index=True` and also `index_label=<name of the index colum>` to name that column. CC: @merveenoyan
2023-02-09T09:29:08Z
5,490
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-01T10:20:55Z
https://api.github.com/repos/huggingface/datasets/issues/5490/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5490/timeline
Do not add index column by default when exporting to CSV
https://api.github.com/repos/huggingface/datasets/issues/5490/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-02-09T09:22:23Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5490.diff", "html_url": "https://github.com/huggingface/datasets/pull/5490", "merged_at": "2023-02-09T09:22:23Z", "patch_url": "https://github.com/huggingface/datasets/pull/5490.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5490" }
PR_kwDODunzps5I_nz-
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5490/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5490
https://github.com/huggingface/datasets/pull/5490
true
1,565,761,705
https://api.github.com/repos/huggingface/datasets/issues/5489/labels{/name}
Pin `dill` lower version compatible with `datasets`. Related to: - #5487 - #288 Note that the required `dill._dill` module was introduced in dill-2.8.0, however we have heuristically tested that datasets can only be installed with dill>=3.0.0 (otherwise pip hangs indefinitely while preparing metadata for multiprocess-0.70.7)
2023-02-02T07:48:09Z
5,489
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-02-01T09:33:42Z
https://api.github.com/repos/huggingface/datasets/issues/5489/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5489/timeline
Pin dill lower version
https://api.github.com/repos/huggingface/datasets/issues/5489/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-02-02T07:40:43Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5489.diff", "html_url": "https://github.com/huggingface/datasets/pull/5489", "merged_at": "2023-02-02T07:40:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/5489.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5489" }
PR_kwDODunzps5I_WPH
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5489/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5489
https://github.com/huggingface/datasets/pull/5489
true
1,565,025,262
https://api.github.com/repos/huggingface/datasets/issues/5488/labels{/name}
### Describe the bug When loading a CommonVoice dataset with `datasets==2.9.0` and `torchaudio>=0.12.0`, I get an error reading the audio arrays: ```python --------------------------------------------------------------------------- LibsndfileError Traceback (most recent call last) ~/.local/lib/python3.8/site-packages/datasets/features/audio.py in _decode_mp3(self, path_or_file) 310 try: # try torchaudio anyway because sometimes it works (depending on the os and os packages installed) --> 311 array, sampling_rate = self._decode_mp3_torchaudio(path_or_file) 312 except RuntimeError: ~/.local/lib/python3.8/site-packages/datasets/features/audio.py in _decode_mp3_torchaudio(self, path_or_file) 351 --> 352 array, sampling_rate = torchaudio.load(path_or_file, format="mp3") 353 if self.sampling_rate and self.sampling_rate != sampling_rate: ~/.local/lib/python3.8/site-packages/torchaudio/backend/soundfile_backend.py in load(filepath, frame_offset, num_frames, normalize, channels_first, format) 204 """ --> 205 with soundfile.SoundFile(filepath, "r") as file_: 206 if file_.format != "WAV" or normalize: ~/.local/lib/python3.8/site-packages/soundfile.py in __init__(self, file, mode, samplerate, channels, subtype, endian, format, closefd) 654 format, subtype, endian) --> 655 self._file = self._open(file, mode_int, closefd) 656 if set(mode).issuperset('r+') and self.seekable(): ~/.local/lib/python3.8/site-packages/soundfile.py in _open(self, file, mode_int, closefd) 1212 err = _snd.sf_error(file_ptr) -> 1213 raise LibsndfileError(err, prefix="Error opening {0!r}: ".format(self.name)) 1214 if mode_int == _snd.SFM_WRITE: LibsndfileError: Error opening <_io.BytesIO object at 0x7fa539462090>: File contains data in an unknown format. ``` I assume this is because there's some issue with the mp3 decoding process. I've verified that I have `ffmpeg>=4` (on a Linux distro), which appears to be the fallback backend for `torchaudio,` (at least according to #4889). ### Steps to reproduce the bug ```python dataset = load_dataset("mozilla-foundation/common_voice_11_0", "be", split="train") dataset[0] ``` ### Expected behavior Similar behavior to `torchaudio<0.12.0`, which doesn't result in a `LibsndfileError` ### Environment info - `datasets` version: 2.9.0 - Platform: Linux-5.15.0-52-generic-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 10.0.1 - Pandas version: 1.5.1
2023-03-02T16:25:14Z
5,488
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-31T21:25:33Z
https://api.github.com/repos/huggingface/datasets/issues/5488/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5488/timeline
Error loading MP3 files from CommonVoice
https://api.github.com/repos/huggingface/datasets/issues/5488/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/110259722?v=4", "events_url": "https://api.github.com/users/kradonneoh/events{/privacy}", "followers_url": "https://api.github.com/users/kradonneoh/followers", "following_url": "https://api.github.com/users/kradonneoh/following{/other_user}", "gists_url": "https://api.github.com/users/kradonneoh/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kradonneoh", "id": 110259722, "login": "kradonneoh", "node_id": "U_kgDOBpJuCg", "organizations_url": "https://api.github.com/users/kradonneoh/orgs", "received_events_url": "https://api.github.com/users/kradonneoh/received_events", "repos_url": "https://api.github.com/users/kradonneoh/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kradonneoh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kradonneoh/subscriptions", "type": "User", "url": "https://api.github.com/users/kradonneoh" }
[]
null
completed
NONE
2023-03-02T16:25:13Z
null
I_kwDODunzps5dSGPu
[ "Hi @kradonneoh, thanks for reporting.\r\n\r\nPlease note that to work with audio datasets (and specifically with MP3 files) we have detailed installation instructions in our docs: https://huggingface.co/docs/datasets/installation#audio\r\n- one of the requirements is torchaudio<0.12.0\r\n\r\nLet us know if the pro...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5488/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5488
https://github.com/huggingface/datasets/issues/5488
false
1,564,480,121
https://api.github.com/repos/huggingface/datasets/issues/5487/labels{/name}
### Describe the bug I installed the `datasets` package and when I try to `import` it, I get the following error: ``` Traceback (most recent call last): File "/var/folders/jt/zw5g74ln6tqfdzsl8tx378j00000gn/T/ipykernel_3805/3458380017.py", line 1, in <module> import datasets File "/Users/avivbrokman/opt/anaconda3/lib/python3.9/site-packages/datasets/__init__.py", line 43, in <module> from .arrow_dataset import Dataset File "/Users/avivbrokman/opt/anaconda3/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 66, in <module> from .arrow_writer import ArrowWriter, OptimizedTypedSequence File "/Users/avivbrokman/opt/anaconda3/lib/python3.9/site-packages/datasets/arrow_writer.py", line 27, in <module> from .features import Features, Image, Value File "/Users/avivbrokman/opt/anaconda3/lib/python3.9/site-packages/datasets/features/__init__.py", line 17, in <module> from .audio import Audio File "/Users/avivbrokman/opt/anaconda3/lib/python3.9/site-packages/datasets/features/audio.py", line 12, in <module> from ..download.streaming_download_manager import xopen File "/Users/avivbrokman/opt/anaconda3/lib/python3.9/site-packages/datasets/download/__init__.py", line 9, in <module> from .download_manager import DownloadManager, DownloadMode File "/Users/avivbrokman/opt/anaconda3/lib/python3.9/site-packages/datasets/download/download_manager.py", line 36, in <module> from ..utils.py_utils import NestedDataStructure, map_nested, size_str File "/Users/avivbrokman/opt/anaconda3/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 602, in <module> class Pickler(dill.Pickler): File "/Users/avivbrokman/opt/anaconda3/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 605, in Pickler dispatch = dill._dill.MetaCatchingDict(dill.Pickler.dispatch.copy()) AttributeError: module 'dill' has no attribute '_dill' ``` Looking at the github source code for dill, it appears that `datasets` has a bug or is not compatible with the latest `dill`. Specifically, rather than `dill._dill.XXXX` it should be `dill.dill._dill.XXXX`. But given the popularity of `datasets` I feel confused about me being the first person to have this issue, so it makes me wonder if I'm misdiagnosing the issue. ### Steps to reproduce the bug Install `dill` and `datasets` packages and then `import datasets` ### Expected behavior I expect `datasets` to import. ### Environment info - `datasets` version: 2.9.0 - Platform: macOS-10.16-x86_64-i386-64bit - Python version: 3.9.13 - PyArrow version: 11.0.0 - Pandas version: 1.4.4
2023-02-24T16:18:36Z
5,487
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-31T15:01:08Z
https://api.github.com/repos/huggingface/datasets/issues/5487/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5487/timeline
Incorrect filepath for dill module
https://api.github.com/repos/huggingface/datasets/issues/5487/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/35349273?v=4", "events_url": "https://api.github.com/users/avivbrokman/events{/privacy}", "followers_url": "https://api.github.com/users/avivbrokman/followers", "following_url": "https://api.github.com/users/avivbrokman/following{/other_user}", "gists_url": "https://api.github.com/users/avivbrokman/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/avivbrokman", "id": 35349273, "login": "avivbrokman", "node_id": "MDQ6VXNlcjM1MzQ5Mjcz", "organizations_url": "https://api.github.com/users/avivbrokman/orgs", "received_events_url": "https://api.github.com/users/avivbrokman/received_events", "repos_url": "https://api.github.com/users/avivbrokman/repos", "site_admin": false, "starred_url": "https://api.github.com/users/avivbrokman/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/avivbrokman/subscriptions", "type": "User", "url": "https://api.github.com/users/avivbrokman" }
[]
null
completed
NONE
2023-02-24T16:18:36Z
null
I_kwDODunzps5dQBJ5
[ "Hi! The correct path is still `dill._dill.XXXX` in the latest release. What do you get when you run `python -c \"import dill; print(dill.__version__)\"` in your environment?", "`0.3.6` I feel like that's bad news, because it's probably not the issue.\r\n\r\nMy mistake, about the wrong path guess. I think I did...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5487/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5487
https://github.com/huggingface/datasets/issues/5487
false
1,564,059,749
https://api.github.com/repos/huggingface/datasets/issues/5486/labels{/name}
I have a local a `.txt` file that follows the `CONLL2003` format which I need to load using `load_script`. However, by using `sample_by='line'`, one can only split the dataset into lines without splitting each line into columns. Would it be reasonable to add a `sep` argument in combination with `sample_by='paragraph'` to parse a paragraph into an array for each column ? If so, I am happy to contribute! ## Environment * `python 3.8.10` * `datasets 2.9.0` ## Snippet of `train.txt` ```txt Distribution NN O O and NN O O dynamics NN O O of NN O O electron NN O B-RP complexes NN O I-RP in NN O O cyanobacterial NN O B-R membranes NN O I-R The NN O O occurrence NN O O of NN O O prostaglandin NN O B-R F2Ξ± NN O I-R in NN O O Pharbitis NN O B-R seedlings NN O I-R grown NN O O under NN O O short NN O B-P days NN O I-P or NN O I-P days NN O I-P ``` ## Current Behaviour ```python # defining 4 features ['tokens', 'pos_tags', 'chunk_tags', 'ner_tags'] here would fail with `ValueError: Length of names (4) does not match length of arrays (1)` dataset = datasets.load_dataset(path='text', features=features, data_files={'train': 'train.txt'}, sample_by='line') dataset['train']['tokens'][0] >>> 'Distribution\tNN\tO\tO' ``` ## Expected Behaviour / Suggestion ```python # suppose we defined 4 features ['tokens', 'pos_tags', 'chunk_tags', 'ner_tags'] dataset = datasets.load_dataset(path='text', features=features, data_files={'train': 'train.txt'}, sample_by='paragraph', sep='\t') dataset['train']['tokens'][0] >>> ['Distribution', 'and', 'dynamics', ... ] dataset['train']['ner_tags'][0] >>> ['O', 'O', 'O', ... ] ```
2023-01-31T14:50:18Z
5,486
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-31T10:39:53Z
https://api.github.com/repos/huggingface/datasets/issues/5486/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5486/timeline
Adding `sep` to TextConfig
https://api.github.com/repos/huggingface/datasets/issues/5486/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/29576434?v=4", "events_url": "https://api.github.com/users/omar-araboghli/events{/privacy}", "followers_url": "https://api.github.com/users/omar-araboghli/followers", "following_url": "https://api.github.com/users/omar-araboghli/following{/other_user}", "gists_url": "https://api.github.com/users/omar-araboghli/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/omar-araboghli", "id": 29576434, "login": "omar-araboghli", "node_id": "MDQ6VXNlcjI5NTc2NDM0", "organizations_url": "https://api.github.com/users/omar-araboghli/orgs", "received_events_url": "https://api.github.com/users/omar-araboghli/received_events", "repos_url": "https://api.github.com/users/omar-araboghli/repos", "site_admin": false, "starred_url": "https://api.github.com/users/omar-araboghli/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/omar-araboghli/subscriptions", "type": "User", "url": "https://api.github.com/users/omar-araboghli" }
[]
null
null
NONE
null
null
I_kwDODunzps5dOahl
[ "Hi @omar-araboghli, thanks for your proposal.\r\n\r\nHave you tried to use \"csv\" loader instead of \"text\"? That already has a `sep` argument.", "Hi @albertvillanova, thanks for the quick response!\r\n\r\nIndeed, I have been trying to use `csv` instead of `text`. However I am still not able to define range of...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5486/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5486
https://github.com/huggingface/datasets/issues/5486
false
1,563,002,829
https://api.github.com/repos/huggingface/datasets/issues/5485/labels{/name}
Introduces an `IterableDataset` and how to access it in the tutorial section. It also adds a brief next step section at the end to provide a path for users who want more explanation and a path for users who want something more practical and learn how to preprocess these dataset types. It'll complement the awesome new doc introduced in: - #5410
2023-02-01T18:15:38Z
5,485
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-30T18:43:04Z
https://api.github.com/repos/huggingface/datasets/issues/5485/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5485/timeline
Add section in tutorial for IterableDataset
https://api.github.com/repos/huggingface/datasets/issues/5485/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
null
null
MEMBER
2023-02-01T18:08:46Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5485.diff", "html_url": "https://github.com/huggingface/datasets/pull/5485", "merged_at": "2023-02-01T18:08:46Z", "patch_url": "https://github.com/huggingface/datasets/pull/5485.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5485" }
PR_kwDODunzps5I2ER2
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5485/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5485
https://github.com/huggingface/datasets/pull/5485
true
1,562,877,070
https://api.github.com/repos/huggingface/datasets/issues/5484/labels{/name}
This PR will fix the issue mentioned in #5461. Here is brief overview, ## Bug: Discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-side comparison, ![image](https://user-images.githubusercontent.com/36858976/214381162-1d9582c2-6750-4114-a01a-61ca1cd5f872.png) ## Fix: When I first loaded the datasets from HF I noticed it was 30GB but in DenseDepth data is only 4GB with dtype=uint8. This means data from fast-depth (before loading to HF) must have high precision. So when I tried to dig deeper by directly loading depth_map from `h5py`, I found depth_map from `h5py` came with `float32`. But when the data is processed in HF with `datasets.Image()` it was directly converted to `uint8` from `float32` hence the **discretized** depth map. https://github.com/huggingface/datasets/blob/c78559cacbb0ca6e0bc8bfc313cc0359f8c23ead/src/datasets/features/image.py#L91-L93 cc: @sayakpaul @lhoestq
2023-09-29T06:43:11Z
5,484
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-30T17:37:08Z
https://api.github.com/repos/huggingface/datasets/issues/5484/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5484/timeline
Update docs for `nyu_depth_v2` dataset
https://api.github.com/repos/huggingface/datasets/issues/5484/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36858976?v=4", "events_url": "https://api.github.com/users/awsaf49/events{/privacy}", "followers_url": "https://api.github.com/users/awsaf49/followers", "following_url": "https://api.github.com/users/awsaf49/following{/other_user}", "gists_url": "https://api.github.com/users/awsaf49/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/awsaf49", "id": 36858976, "login": "awsaf49", "node_id": "MDQ6VXNlcjM2ODU4OTc2", "organizations_url": "https://api.github.com/users/awsaf49/orgs", "received_events_url": "https://api.github.com/users/awsaf49/received_events", "repos_url": "https://api.github.com/users/awsaf49/repos", "site_admin": false, "starred_url": "https://api.github.com/users/awsaf49/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/awsaf49/subscriptions", "type": "User", "url": "https://api.github.com/users/awsaf49" }
[]
null
null
CONTRIBUTOR
2023-02-05T14:15:04Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5484.diff", "html_url": "https://github.com/huggingface/datasets/pull/5484", "merged_at": "2023-02-05T14:15:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/5484.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5484" }
PR_kwDODunzps5I1oaq
[ "I think I need to create another PR on https://huggingface.co/datasets/huggingface/documentation-images/tree/main/datasets for hosting the images there?", "_The documentation is not available anymore as the PR was closed or merged._", "Thanks for the update @awsaf49 !", "> Thanks a lot for the updates!\r\n> ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5484/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5484
https://github.com/huggingface/datasets/pull/5484
true
1,560,894,690
https://api.github.com/repos/huggingface/datasets/issues/5483/labels{/name}
### Describe the bug Uploading a simple dataset ends with an exception ### Steps to reproduce the bug I created a new conda env with python 3.10, pip installed datasets and: ```python >>> from datasets import load_dataset, load_from_disk, Dataset >>> d = Dataset.from_dict({"text": ["hello"] * 2}) >>> d.push_to_hub("ttt111") /home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/utils/_hf_folder.py:92: UserWarning: A token has been found in `/a/home/cc/students/cs/kirstain/.huggingface/token`. This is the old path where tokens were stored. The new location is `/home/olab/kirstain/.cache/huggingface/token` which is configurable using `HF_HOME` environment variable. Your token has been copied to this new location. You can now safely delete the old token file manually or use `huggingface-cli logout`. warnings.warn( Creating parquet from Arrow format: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:00<00:00, 279.94ba/s] Upload 1 LFS files: 0%| | 0/1 [00:02<?, ?it/s] Pushing dataset shards to the dataset hub: 0%| | 0/1 [00:04<?, ?it/s] Traceback (most recent call last): File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/utils/_errors.py", line 264, in hf_raise_for_status response.raise_for_status() File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/requests/models.py", line 1021, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://s3.us-east-1.amazonaws.com/lfs.huggingface.co/repos/cf/0c/cf0c5ab8a3f729e5f57a8b79a36ecea64a31126f13218591c27ed9a1c7bd9b41/ece885a4bb6bbc8c1bb51b45542b805283d74590f72cd4c45d3ba76628570386?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA4N7VTDGO27GPWFUO%2F20230128%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20230128T151640Z&X-Amz-Expires=900&X-Amz-Signature=89e78e9a9d70add7ed93d453334f4f93c6f29d889d46750a1f2da04af73978db&X-Amz-SignedHeaders=host&x-amz-storage-class=INTELLIGENT_TIERING&x-id=PutObject The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/_commit_api.py", line 334, in _inner_upload_lfs_object return _upload_lfs_object( File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/_commit_api.py", line 391, in _upload_lfs_object lfs_upload( File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/lfs.py", line 273, in lfs_upload _upload_single_part( File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/lfs.py", line 305, in _upload_single_part hf_raise_for_status(upload_res) File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/utils/_errors.py", line 318, in hf_raise_for_status raise HfHubHTTPError(str(e), response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: 403 Client Error: Forbidden for url: https://s3.us-east-1.amazonaws.com/lfs.huggingface.co/repos/cf/0c/cf0c5ab8a3f729e5f57a8b79a36ecea64a31126f13218591c27ed9a1c7bd9b41/ece885a4bb6bbc8c1bb51b45542b805283d74590f72cd4c45d3ba76628570386?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA4N7VTDGO27GPWFUO%2F20230128%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20230128T151640Z&X-Amz-Expires=900&X-Amz-Signature=89e78e9a9d70add7ed93d453334f4f93c6f29d889d46750a1f2da04af73978db&X-Amz-SignedHeaders=host&x-amz-storage-class=INTELLIGENT_TIERING&x-id=PutObject The above exception was the direct cause of the following exception: Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 4909, in push_to_hub repo_id, split, uploaded_size, dataset_nbytes, repo_files, deleted_size = self._push_parquet_shards_to_hub( File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 4804, in _push_parquet_shards_to_hub _retry( File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/datasets/utils/file_utils.py", line 281, in _retry return func(*func_args, **func_kwargs) File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 124, in _inner_fn return fn(*args, **kwargs) File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/hf_api.py", line 2537, in upload_file commit_info = self.create_commit( File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 124, in _inner_fn return fn(*args, **kwargs) File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/hf_api.py", line 2346, in create_commit upload_lfs_files( File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 124, in _inner_fn return fn(*args, **kwargs) File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/_commit_api.py", line 346, in upload_lfs_files thread_map( File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/tqdm/contrib/concurrent.py", line 94, in thread_map return _executor_map(ThreadPoolExecutor, fn, *iterables, **tqdm_kwargs) File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/tqdm/contrib/concurrent.py", line 76, in _executor_map return list(tqdm_class(ex.map(fn, *iterables, **map_args), **kwargs)) File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/tqdm/std.py", line 1195, in __iter__ for obj in iterable: File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/concurrent/futures/_base.py", line 621, in result_iterator yield _result_or_cancel(fs.pop()) File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/concurrent/futures/_base.py", line 319, in _result_or_cancel return fut.result(timeout) File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/concurrent/futures/_base.py", line 458, in result return self.__get_result() File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/concurrent/futures/_base.py", line 403, in __get_result raise self._exception File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/concurrent/futures/thread.py", line 58, in run result = self.fn(*self.args, **self.kwargs) File "/home/olab/kirstain/anaconda3/envs/datasets/lib/python3.10/site-packages/huggingface_hub/_commit_api.py", line 338, in _inner_upload_lfs_object raise RuntimeError( RuntimeError: Error while uploading 'data/train-00000-of-00001-6df93048e66df326.parquet' to the Hub. ``` ### Expected behavior The dataset should be uploaded without any exceptions ### Environment info - `datasets` version: 2.9.0 - Platform: Linux-4.15.0-65-generic-x86_64-with-glibc2.27 - Python version: 3.10.9 - PyArrow version: 11.0.0 - Pandas version: 1.5.3
2023-01-29T08:09:49Z
5,483
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-28T15:18:26Z
https://api.github.com/repos/huggingface/datasets/issues/5483/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5483/timeline
Unable to upload dataset
https://api.github.com/repos/huggingface/datasets/issues/5483/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/57996478?v=4", "events_url": "https://api.github.com/users/yuvalkirstain/events{/privacy}", "followers_url": "https://api.github.com/users/yuvalkirstain/followers", "following_url": "https://api.github.com/users/yuvalkirstain/following{/other_user}", "gists_url": "https://api.github.com/users/yuvalkirstain/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/yuvalkirstain", "id": 57996478, "login": "yuvalkirstain", "node_id": "MDQ6VXNlcjU3OTk2NDc4", "organizations_url": "https://api.github.com/users/yuvalkirstain/orgs", "received_events_url": "https://api.github.com/users/yuvalkirstain/received_events", "repos_url": "https://api.github.com/users/yuvalkirstain/repos", "site_admin": false, "starred_url": "https://api.github.com/users/yuvalkirstain/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yuvalkirstain/subscriptions", "type": "User", "url": "https://api.github.com/users/yuvalkirstain" }
[]
null
completed
NONE
2023-01-29T08:09:49Z
null
I_kwDODunzps5dCVzi
[ "Seems to work now, perhaps it was something internal with our university's network." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5483/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5483
https://github.com/huggingface/datasets/issues/5483
false
1,560,853,137
https://api.github.com/repos/huggingface/datasets/issues/5482/labels{/name}
The idea would be to allow this : ```python ds.to_parquet("my_dataset/ds.parquet") reloaded = load_dataset("my_dataset") assert ds.features == reloaded.features ``` And it should also work with Image and Audio types (right now they're reloaded as a dict type) This can be implemented by storing and reading the feature types in the parquet metadata, as we do for arrow files.
2023-02-12T15:57:02Z
5,482
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "BDE59C", "default": fals...
2023-01-28T13:12:31Z
https://api.github.com/repos/huggingface/datasets/issues/5482/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/6368040?v=4", "events_url": "https://api.github.com/users/MFreidank/events{/privacy}", "followers_url": "https://api.github.com/users/MFreidank/followers", "following_url": "https://api.github.com/users/MFreidank/following{/other_user}", "gists_url": "https://api.github.com/users/MFreidank/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/MFreidank", "id": 6368040, "login": "MFreidank", "node_id": "MDQ6VXNlcjYzNjgwNDA=", "organizations_url": "https://api.github.com/users/MFreidank/orgs", "received_events_url": "https://api.github.com/users/MFreidank/received_events", "repos_url": "https://api.github.com/users/MFreidank/repos", "site_admin": false, "starred_url": "https://api.github.com/users/MFreidank/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MFreidank/subscriptions", "type": "User", "url": "https://api.github.com/users/MFreidank" }
https://api.github.com/repos/huggingface/datasets/issues/5482/timeline
Reload features from Parquet metadata
https://api.github.com/repos/huggingface/datasets/issues/5482/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/6368040?v=4", "events_url": "https://api.github.com/users/MFreidank/events{/privacy}", "followers_url": "https://api.github.com/users/MFreidank/followers", "following_url": "https://api.github.com/users/MFreidank/following{/other_user}", "...
null
completed
MEMBER
2023-02-12T15:57:02Z
null
I_kwDODunzps5dCLqR
[ "I'd be happy to have a look, if nobody else has started working on this yet @lhoestq. \r\n\r\nIt seems to me that for the `arrow` format features are currently attached as metadata [in `datasets.arrow_writer`](https://github.com/huggingface/datasets/blob/5f810b7011a8a4ab077a1847c024d2d9e267b065/src/datasets/arrow_...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5482/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5482
https://github.com/huggingface/datasets/issues/5482
false
1,560,468,195
https://api.github.com/repos/huggingface/datasets/issues/5481/labels{/name}
The idea would be to allow something like ```python ds = load_dataset("c4", "en", as_iterable=True) ``` To be used to train models. It would load an IterableDataset from the cached Arrow files. Cc @stas00 Edit : from the discussions we may load from cache when streaming=True
2023-06-26T10:48:53Z
5,481
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "BDE59C", "default": fals...
2023-01-27T21:43:51Z
https://api.github.com/repos/huggingface/datasets/issues/5481/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5481/timeline
Load a cached dataset as iterable
https://api.github.com/repos/huggingface/datasets/issues/5481/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
null
null
I_kwDODunzps5dAtrj
[ "Can I work on this issue? I am pretty new to this.", "Hi ! Sure :) you can comment `#self-assign` to assign yourself to this issue.\r\n\r\nI can give you some pointers to get started:\r\n\r\n`load_dataset` works roughly this way:\r\n1. it instantiate a dataset builder using `load_dataset_builder()`\r\n2. the bui...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 5, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 5, "url": "https://api.github.com/repos/huggingface/datasets/issues/5481/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5481
https://github.com/huggingface/datasets/issues/5481
false
1,560,364,866
https://api.github.com/repos/huggingface/datasets/issues/5480/labels{/name}
Close #5474 and #5468.
2023-02-13T11:10:13Z
5,480
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-27T20:06:16Z
https://api.github.com/repos/huggingface/datasets/issues/5480/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5480/timeline
Select columns of Dataset or DatasetDict
https://api.github.com/repos/huggingface/datasets/issues/5480/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/9336514?v=4", "events_url": "https://api.github.com/users/daskol/events{/privacy}", "followers_url": "https://api.github.com/users/daskol/followers", "following_url": "https://api.github.com/users/daskol/following{/other_user}", "gists_url": "https://api.github.com/users/daskol/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/daskol", "id": 9336514, "login": "daskol", "node_id": "MDQ6VXNlcjkzMzY1MTQ=", "organizations_url": "https://api.github.com/users/daskol/orgs", "received_events_url": "https://api.github.com/users/daskol/received_events", "repos_url": "https://api.github.com/users/daskol/repos", "site_admin": false, "starred_url": "https://api.github.com/users/daskol/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/daskol/subscriptions", "type": "User", "url": "https://api.github.com/users/daskol" }
[]
null
null
CONTRIBUTOR
2023-02-13T09:59:35Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5480.diff", "html_url": "https://github.com/huggingface/datasets/pull/5480", "merged_at": "2023-02-13T09:59:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/5480.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5480" }
PR_kwDODunzps5ItY2y
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5480/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5480
https://github.com/huggingface/datasets/pull/5480
true
1,560,357,590
https://api.github.com/repos/huggingface/datasets/issues/5479/labels{/name}
### Describe the bug I'm using a custom audio dataset (400+ audio files) in the correct format for audiofolder. Although loading the dataset with audiofolder works in one local setup, it doesn't in a remote one (it just creates an empty dataset). I have both ffmpeg and libndfile installed on both computers, what could be missing/need to be updated in the one that doesn't work? On the remote env, libsndfile is 1.0.28 and ffmpeg is 4.2.1. from datasets import load_dataset ds = load_dataset("audiofolder", data_dir="...") Here is the output (should be generating 400+ rows): Downloading and preparing dataset audiofolder/default to ... Downloading data files: 0%| | 0/2 [00:00<?, ?it/s] Downloading data files: 0it [00:00, ?it/s] Extracting data files: 0it [00:00, ?it/s] Generating train split: 0 examples [00:00, ? examples/s] Dataset audiofolder downloaded and prepared to ... Subsequent calls will reuse this data. 0%| | 0/1 [00:00<?, ?it/s] DatasetDict({ train: Dataset({ features: ['audio', 'transcription'], num_rows: 1 }) }) Here is my pip environment in the one that doesn't work (uses torch 1.11.a0 from shared env): Package Version ------------------- ------------------- aiofiles 22.1.0 aiohttp 3.8.3 aiosignal 1.3.1 altair 4.2.1 anyio 3.6.2 appdirs 1.4.4 argcomplete 2.0.0 argon2-cffi 20.1.0 astunparse 1.6.3 async-timeout 4.0.2 attrs 21.2.0 audioread 3.0.0 backcall 0.2.0 bleach 4.0.0 certifi 2021.10.8 cffi 1.14.6 charset-normalizer 2.0.12 click 8.1.3 contourpy 1.0.7 cycler 0.11.0 datasets 2.9.0 debugpy 1.4.1 decorator 5.0.9 defusedxml 0.7.1 dill 0.3.6 distlib 0.3.4 entrypoints 0.3 evaluate 0.4.0 expecttest 0.1.3 fastapi 0.89.1 ffmpy 0.3.0 filelock 3.6.0 fonttools 4.38.0 frozenlist 1.3.3 fsspec 2023.1.0 future 0.18.2 gradio 3.16.2 h11 0.14.0 httpcore 0.16.3 httpx 0.23.3 huggingface-hub 0.12.0 idna 3.3 ipykernel 6.2.0 ipython 7.26.0 ipython-genutils 0.2.0 ipywidgets 7.6.3 jedi 0.18.0 Jinja2 3.0.1 jiwer 2.5.1 joblib 1.2.0 jsonschema 3.2.0 jupyter 1.0.0 jupyter-client 6.1.12 jupyter-console 6.4.0 jupyter-core 4.7.1 jupyterlab-pygments 0.1.2 jupyterlab-widgets 1.0.0 kiwisolver 1.4.4 Levenshtein 0.20.2 librosa 0.9.2 linkify-it-py 1.0.3 llvmlite 0.39.1 markdown-it-py 2.1.0 MarkupSafe 2.0.1 matplotlib 3.6.3 matplotlib-inline 0.1.2 mdit-py-plugins 0.3.3 mdurl 0.1.2 mistune 0.8.4 multidict 6.0.4 multiprocess 0.70.14 nbclient 0.5.4 nbconvert 6.1.0 nbformat 5.1.3 nest-asyncio 1.5.1 notebook 6.4.3 numba 0.56.4 numpy 1.20.3 orjson 3.8.5 packaging 21.0 pandas 1.5.3 pandocfilters 1.4.3 parso 0.8.2 pexpect 4.8.0 pickleshare 0.7.5 Pillow 9.4.0 pip 22.3.1 pipx 1.1.0 platformdirs 2.5.2 pooch 1.6.0 prometheus-client 0.11.0 prompt-toolkit 3.0.19 psutil 5.9.0 ptyprocess 0.7.0 pyarrow 10.0.1 pycparser 2.20 pycryptodome 3.16.0 pydantic 1.10.4 pydub 0.25.1 Pygments 2.10.0 pyparsing 2.4.7 pyrsistent 0.18.0 python-dateutil 2.8.2 python-multipart 0.0.5 pytz 2022.7.1 PyYAML 6.0 pyzmq 22.2.1 qtconsole 5.1.1 QtPy 1.10.0 rapidfuzz 2.13.7 regex 2022.10.31 requests 2.27.1 resampy 0.4.2 responses 0.18.0 rfc3986 1.5.0 scikit-learn 1.2.1 scipy 1.6.3 Send2Trash 1.8.0 setuptools 65.5.1 shiboken6 6.3.1 shiboken6-generator 6.3.1 six 1.16.0 sniffio 1.3.0 soundfile 0.11.0 starlette 0.22.0 terminado 0.11.0 testpath 0.5.0 threadpoolctl 3.1.0 tokenizers 0.13.2 toolz 0.12.0 torch 1.11.0a0+gitunknown tornado 6.1 tqdm 4.64.1 traitlets 5.0.5 transformers 4.27.0.dev0 types-dataclasses 0.6.4 typing_extensions 4.1.1 uc-micro-py 1.0.1 urllib3 1.26.9 userpath 1.8.0 uvicorn 0.20.0 virtualenv 20.14.1 wcwidth 0.2.5 webencodings 0.5.1 websockets 10.4 wheel 0.37.1 widgetsnbextension 3.5.1 xxhash 3.2.0 yarl 1.8.2 ### Steps to reproduce the bug Create a pip environment with the packages listed above (make sure ffmpeg and libsndfile is installed with same versions listed above). Create a custom audio dataset and load it in with load_dataset("audiofolder", ...) ### Expected behavior load_dataset should create a dataset with 400+ rows. ### Environment info - `datasets` version: 2.9.0 - Platform: Linux-3.10.0-1160.80.1.el7.x86_64-x86_64-with-glibc2.17 - Python version: 3.9.0 - PyArrow version: 10.0.1 - Pandas version: 1.5.3
2023-01-29T05:23:14Z
5,479
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-27T20:01:22Z
https://api.github.com/repos/huggingface/datasets/issues/5479/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5479/timeline
audiofolder works on local env, but creates empty dataset in a remote one, what dependencies could I be missing/outdated
https://api.github.com/repos/huggingface/datasets/issues/5479/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/107211437?v=4", "events_url": "https://api.github.com/users/jcho19/events{/privacy}", "followers_url": "https://api.github.com/users/jcho19/followers", "following_url": "https://api.github.com/users/jcho19/following{/other_user}", "gists_url": "https://api.github.com/users/jcho19/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jcho19", "id": 107211437, "login": "jcho19", "node_id": "U_kgDOBmPqrQ", "organizations_url": "https://api.github.com/users/jcho19/orgs", "received_events_url": "https://api.github.com/users/jcho19/received_events", "repos_url": "https://api.github.com/users/jcho19/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jcho19/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jcho19/subscriptions", "type": "User", "url": "https://api.github.com/users/jcho19" }
[]
null
completed
NONE
2023-01-29T05:23:14Z
null
I_kwDODunzps5dASrW
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5479/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5479
https://github.com/huggingface/datasets/issues/5479
false
1,560,357,583
https://api.github.com/repos/huggingface/datasets/issues/5478/labels{/name}
From this [feedback](https://discuss.huggingface.co/t/nonmatchingsplitssizeserror/30033) on the forum, thought I'd include a tip for recomputing the metadata numbers if it is your own dataset.
2023-01-30T19:22:21Z
5,478
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-27T20:01:22Z
https://api.github.com/repos/huggingface/datasets/issues/5478/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5478/timeline
Tip for recomputing metadata
https://api.github.com/repos/huggingface/datasets/issues/5478/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
null
null
MEMBER
2023-01-30T19:15:26Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5478.diff", "html_url": "https://github.com/huggingface/datasets/pull/5478", "merged_at": "2023-01-30T19:15:26Z", "patch_url": "https://github.com/huggingface/datasets/pull/5478.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5478" }
PR_kwDODunzps5ItXQG
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5478/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5478
https://github.com/huggingface/datasets/pull/5478
true
1,559,909,892
https://api.github.com/repos/huggingface/datasets/issues/5477/labels{/name}
Once the source issue is fixed: - pandas-dev/pandas#51015 we should revert the pin introduced in: - #5476
2024-01-26T14:50:45Z
5,477
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-27T15:01:55Z
https://api.github.com/repos/huggingface/datasets/issues/5477/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5477/timeline
Unpin sqlalchemy once issue is fixed
https://api.github.com/repos/huggingface/datasets/issues/5477/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
completed
MEMBER
2024-01-26T14:50:45Z
null
I_kwDODunzps5c-lYE
[ "@albertvillanova It looks like that issue has been fixed so I made a PR to unpin sqlalchemy! ", "The source issue:\r\n- https://github.com/pandas-dev/pandas/issues/40686\r\n\r\nhas been fixed:\r\n- https://github.com/pandas-dev/pandas/pull/48576\r\n\r\nThe fix was released yesterday (2023-04-03) only in `pandas-...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5477/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5477
https://github.com/huggingface/datasets/issues/5477
false
1,559,594,684
https://api.github.com/repos/huggingface/datasets/issues/5476/labels{/name}
since sqlalchemy update to 2.0.0 the CI started to fail: https://github.com/huggingface/datasets/actions/runs/4023742457/jobs/6914976514 the error comes from pandas: https://github.com/pandas-dev/pandas/issues/51015
2023-01-27T12:06:51Z
5,476
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-27T11:26:38Z
https://api.github.com/repos/huggingface/datasets/issues/5476/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5476/timeline
Pin sqlalchemy
https://api.github.com/repos/huggingface/datasets/issues/5476/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-01-27T11:57:48Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5476.diff", "html_url": "https://github.com/huggingface/datasets/pull/5476", "merged_at": "2023-01-27T11:57:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/5476.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5476" }
PR_kwDODunzps5IqwC_
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5476/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5476
https://github.com/huggingface/datasets/pull/5476
true
1,559,030,149
https://api.github.com/repos/huggingface/datasets/issues/5475/labels{/name}
### Describe the bug I'm basically running the same scanning experiment from the tutorials https://huggingface.co/course/chapter5/4?fw=pt except now I'm comparing to a native pyarrow version. I'm finding that the native pyarrow approach is much faster (2 orders of magnitude). Is there something I'm missing that explains this phenomenon? ### Steps to reproduce the bug https://colab.research.google.com/drive/11EtHDaGAf1DKCpvYnAPJUW-LFfAcDzHY?usp=sharing ### Expected behavior I expect scan times to be on par with using pyarrow directly. ### Environment info standard colab environment
2023-01-30T16:17:11Z
5,475
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-27T01:32:25Z
https://api.github.com/repos/huggingface/datasets/issues/5475/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5475/timeline
Dataset scan time is much slower than using native arrow
https://api.github.com/repos/huggingface/datasets/issues/5475/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/121845112?v=4", "events_url": "https://api.github.com/users/jonny-cyberhaven/events{/privacy}", "followers_url": "https://api.github.com/users/jonny-cyberhaven/followers", "following_url": "https://api.github.com/users/jonny-cyberhaven/following{/other_user}", "gists_url": "https://api.github.com/users/jonny-cyberhaven/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jonny-cyberhaven", "id": 121845112, "login": "jonny-cyberhaven", "node_id": "U_kgDOB0M1eA", "organizations_url": "https://api.github.com/users/jonny-cyberhaven/orgs", "received_events_url": "https://api.github.com/users/jonny-cyberhaven/received_events", "repos_url": "https://api.github.com/users/jonny-cyberhaven/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jonny-cyberhaven/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonny-cyberhaven/subscriptions", "type": "User", "url": "https://api.github.com/users/jonny-cyberhaven" }
[]
null
completed
CONTRIBUTOR
2023-01-30T16:17:11Z
null
I_kwDODunzps5c7OmF
[ "Hi ! In your code you only iterate on the Arrow buffers - you don't actually load the data as python objects. For a fair comparison, you can modify your code using:\r\n```diff\r\n- for _ in range(0, len(table), bsz):\r\n- _ = {k:table[k][_ : _ + bsz] for k in cols}\r\n+ for _ in range(0, len(table)...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5475/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5475
https://github.com/huggingface/datasets/issues/5475
false
1,558,827,155
https://api.github.com/repos/huggingface/datasets/issues/5474/labels{/name}
### Feature request There is no operation to select a subset of columns of original dataset. Expected API follows. ```python a = Dataset.from_dict({ 'int': [0, 1, 2] 'char': ['a', 'b', 'c'], 'none': [None] * 3, }) b = a.project('int', 'char') # usually, .select() print(a.column_names) # stdout: ['int', 'char', 'none'] print(b.column_names) # stdout: ['int', 'char'] ``` Method project can easily accept not only column names (as a `str)` but univariant function applied to corresponding column as an example. Or keyword arguments can be used in order to rename columns in advance (see `pandas`, `pyspark`, `pyarrow`, and SQL).. ### Motivation Projection is a typical operation in every data processing library. And it is a basic block of a well-known data manipulation language like SQL. Without this operation `datasets.Dataset` interface is not complete. ### Your contribution Not sure. Some of my PRs are still open and some do not have any discussions.
2023-02-13T09:59:37Z
5,474
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "cfd3d7", "default": true, "description": "This issue or pull request already exists", "id": 1935892865, "name": "duplicate", "node_id": "MDU6TGFiZWwxOTM1ODkyODY1", "url": "https://api.github.com/repos/huggingface/datasets/labels/duplicate" }, { "color": "a2eeef", ...
2023-01-26T21:47:53Z
https://api.github.com/repos/huggingface/datasets/issues/5474/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5474/timeline
Column project operation on `datasets.Dataset`
https://api.github.com/repos/huggingface/datasets/issues/5474/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/9336514?v=4", "events_url": "https://api.github.com/users/daskol/events{/privacy}", "followers_url": "https://api.github.com/users/daskol/followers", "following_url": "https://api.github.com/users/daskol/following{/other_user}", "gists_url": "https://api.github.com/users/daskol/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/daskol", "id": 9336514, "login": "daskol", "node_id": "MDQ6VXNlcjkzMzY1MTQ=", "organizations_url": "https://api.github.com/users/daskol/orgs", "received_events_url": "https://api.github.com/users/daskol/received_events", "repos_url": "https://api.github.com/users/daskol/repos", "site_admin": false, "starred_url": "https://api.github.com/users/daskol/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/daskol/subscriptions", "type": "User", "url": "https://api.github.com/users/daskol" }
[]
null
completed
CONTRIBUTOR
2023-02-13T09:59:37Z
null
I_kwDODunzps5c6dCT
[ "Hi ! This would be a nice addition indeed :) This sounds like a duplicate of https://github.com/huggingface/datasets/issues/5468\r\n\r\n> Not sure. Some of my PRs are still open and some do not have any discussions.\r\n\r\nSorry to hear that, feel free to ping me on those PRs" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5474/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5474
https://github.com/huggingface/datasets/issues/5474
false
1,558,668,197
https://api.github.com/repos/huggingface/datasets/issues/5473/labels{/name}
null
2023-01-26T19:47:34Z
5,473
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-26T19:34:44Z
https://api.github.com/repos/huggingface/datasets/issues/5473/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5473/timeline
Set dev version
https://api.github.com/repos/huggingface/datasets/issues/5473/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-01-26T19:38:30Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5473.diff", "html_url": "https://github.com/huggingface/datasets/pull/5473", "merged_at": "2023-01-26T19:38:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/5473.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5473" }
PR_kwDODunzps5Inm9h
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5473/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5473
https://github.com/huggingface/datasets/pull/5473
true
1,558,662,251
https://api.github.com/repos/huggingface/datasets/issues/5472/labels{/name}
null
2023-01-26T19:40:44Z
5,472
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-26T19:29:42Z
https://api.github.com/repos/huggingface/datasets/issues/5472/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5472/timeline
Release: 2.9.0
https://api.github.com/repos/huggingface/datasets/issues/5472/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-01-26T19:33:00Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5472.diff", "html_url": "https://github.com/huggingface/datasets/pull/5472", "merged_at": "2023-01-26T19:33:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/5472.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5472" }
PR_kwDODunzps5Inlp8
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5472/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5472
https://github.com/huggingface/datasets/pull/5472
true
1,558,557,545
https://api.github.com/repos/huggingface/datasets/issues/5471/labels{/name}
`to_tf_dataset` calls can be very costly because of the number of test batches drawn during `_get_output_signature`. The test batches are draw in order to estimate the shapes when creating the tensorflow dataset. This is necessary when the shapes can be irregular, but not in cases when the tensor shapes are the same across all samples. This PR adds an option to change the number of batches drawn, so the user can speed this conversion up. Running the following, and modifying `num_test_batches` ``` import time from datasets import load_dataset from transformers import DefaultDataCollator data_collator = DefaultDataCollator() dataset = load_dataset("beans") dataset = dataset["train"].with_format("np") start = time.time() dataset = dataset.to_tf_dataset( columns=["image"], label_cols=["label"], batch_size=8, collate_fn=data_collator, num_test_batches=NUM_TEST_BATCHES, ) end = time.time() print(end - start) ``` NUM_TEST_BATCHES=200: 0.8197s NUM_TEST_BATCHES=50: 0.3070s NUM_TEST_BATCHES=2: 0.1417s NUM_TEST_BATCHES=1: 0.1352s
2023-01-27T18:16:45Z
5,471
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-26T18:09:40Z
https://api.github.com/repos/huggingface/datasets/issues/5471/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5471/timeline
Add num_test_batches option
https://api.github.com/repos/huggingface/datasets/issues/5471/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/22614925?v=4", "events_url": "https://api.github.com/users/amyeroberts/events{/privacy}", "followers_url": "https://api.github.com/users/amyeroberts/followers", "following_url": "https://api.github.com/users/amyeroberts/following{/other_user}", "gists_url": "https://api.github.com/users/amyeroberts/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/amyeroberts", "id": 22614925, "login": "amyeroberts", "node_id": "MDQ6VXNlcjIyNjE0OTI1", "organizations_url": "https://api.github.com/users/amyeroberts/orgs", "received_events_url": "https://api.github.com/users/amyeroberts/received_events", "repos_url": "https://api.github.com/users/amyeroberts/repos", "site_admin": false, "starred_url": "https://api.github.com/users/amyeroberts/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amyeroberts/subscriptions", "type": "User", "url": "https://api.github.com/users/amyeroberts" }
[]
null
null
CONTRIBUTOR
2023-01-27T18:08:36Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5471.diff", "html_url": "https://github.com/huggingface/datasets/pull/5471", "merged_at": "2023-01-27T18:08:36Z", "patch_url": "https://github.com/huggingface/datasets/pull/5471.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5471" }
PR_kwDODunzps5InPA7
[ "_The documentation is not available anymore as the PR was closed or merged._", "I thought this issue was resolved in my parallel `to_tf_dataset` PR! I changed the default `num_test_batches` in `_get_output_signature` to 20 and used a test batch size of 1 to maximize variance to detect shorter samples. I think it...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5471/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5471
https://github.com/huggingface/datasets/pull/5471
true
1,558,542,611
https://api.github.com/repos/huggingface/datasets/issues/5470/labels{/name}
Encourages users to create a dataset card on the Hub directly with the new metadata ui + import dataset card template instead of telling users to manually create and upload one.
2023-01-27T16:27:00Z
5,470
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-26T17:57:51Z
https://api.github.com/repos/huggingface/datasets/issues/5470/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5470/timeline
Update dataset card creation
https://api.github.com/repos/huggingface/datasets/issues/5470/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
null
null
MEMBER
2023-01-27T16:20:10Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5470.diff", "html_url": "https://github.com/huggingface/datasets/pull/5470", "merged_at": "2023-01-27T16:20:10Z", "patch_url": "https://github.com/huggingface/datasets/pull/5470.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5470" }
PR_kwDODunzps5InLw9
[ "_The documentation is not available anymore as the PR was closed or merged._", "The CI failure is unrelated to your PR - feel free to merge :)", "Haha thanks, you read my mind :)", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n##...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5470/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5470
https://github.com/huggingface/datasets/pull/5470
true
1,558,346,906
https://api.github.com/repos/huggingface/datasets/issues/5469/labels{/name}
The docstrings say that it was supposed to be deprecated since version 2.4.0, can we remove it?
2023-01-26T17:37:51Z
5,469
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-26T15:40:56Z
https://api.github.com/repos/huggingface/datasets/issues/5469/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5469/timeline
Remove deprecated `shard_size` arg from `.push_to_hub()`
https://api.github.com/repos/huggingface/datasets/issues/5469/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
null
null
CONTRIBUTOR
2023-01-26T17:30:59Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5469.diff", "html_url": "https://github.com/huggingface/datasets/pull/5469", "merged_at": "2023-01-26T17:30:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/5469.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5469" }
PR_kwDODunzps5Imhk2
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5469/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5469
https://github.com/huggingface/datasets/pull/5469
true
1,558,066,625
https://api.github.com/repos/huggingface/datasets/issues/5468/labels{/name}
### Feature request In this blog post https://huggingface.co/blog/audio-datasets, I noticed the following code: ```python COLUMNS_TO_KEEP = ["text", "audio"] all_columns = gigaspeech["train"].column_names columns_to_remove = set(all_columns) - set(COLUMNS_TO_KEEP) gigaspeech = gigaspeech.remove_columns(columns_to_remove) ``` This kind of thing happens a lot when you don't need to keep all columns from the dataset. It would be more convenient (and less error prone) if you could just write: ```python gigaspeech = gigaspeech.keep_columns(["text", "audio"]) ``` Internally, `keep_columns` could still call `remove_columns`, but it expresses more clearly what the user's intent is. ### Motivation Less code to write for the user of the dataset. ### Your contribution -
2023-02-13T09:59:38Z
5,468
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "7057ff", "default": true...
2023-01-26T12:28:09Z
https://api.github.com/repos/huggingface/datasets/issues/5468/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5468/timeline
Allow opposite of remove_columns on Dataset and DatasetDict
https://api.github.com/repos/huggingface/datasets/issues/5468/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/346853?v=4", "events_url": "https://api.github.com/users/hollance/events{/privacy}", "followers_url": "https://api.github.com/users/hollance/followers", "following_url": "https://api.github.com/users/hollance/following{/other_user}", "gists_url": "https://api.github.com/users/hollance/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hollance", "id": 346853, "login": "hollance", "node_id": "MDQ6VXNlcjM0Njg1Mw==", "organizations_url": "https://api.github.com/users/hollance/orgs", "received_events_url": "https://api.github.com/users/hollance/received_events", "repos_url": "https://api.github.com/users/hollance/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hollance/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hollance/subscriptions", "type": "User", "url": "https://api.github.com/users/hollance" }
[]
null
completed
NONE
2023-02-13T09:59:38Z
null
I_kwDODunzps5c3jXB
[ "Hi! I agree it would be nice to have a method like that. Instead of `keep_columns`, we can name it `select_columns` to be more aligned with PyArrow's naming convention (`pa.Table.select`).", "Hi, I am a newbie to open source and would like to contribute. @mariosasko can I take up this issue ?", "Hey, I also wa...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5468/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5468
https://github.com/huggingface/datasets/issues/5468
false
1,557,898,273
https://api.github.com/repos/huggingface/datasets/issues/5467/labels{/name}
The [conda forge channel](https://anaconda.org/conda-forge/datasets) is lagging behind (as of right now, only 2.7.1 is available), we should recommend using the [Hugging face channel](https://anaconda.org/HuggingFace/datasets) that we are maintaining ``` conda install -c huggingface datasets ```
2023-09-24T10:06:59Z
5,467
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-26T10:03:01Z
https://api.github.com/repos/huggingface/datasets/issues/5467/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5467/timeline
Fix conda command in readme
https://api.github.com/repos/huggingface/datasets/issues/5467/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-01-26T18:29:37Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5467.diff", "html_url": "https://github.com/huggingface/datasets/pull/5467", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5467.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5467" }
PR_kwDODunzps5IlAlk
[ "ah didn't read well - it's all good", "or maybe it isn't ? `-c huggingface -c conda-forge` installs from HF or from conda-forge ?", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5467/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5467
https://github.com/huggingface/datasets/pull/5467
true
1,557,584,845
https://api.github.com/repos/huggingface/datasets/issues/5466/labels{/name}
Pathlib will convert "//" to "/" which causes retry errors when downloading from cloud storage
2023-01-26T17:08:57Z
5,466
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-26T03:25:45Z
https://api.github.com/repos/huggingface/datasets/issues/5466/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5466/timeline
remove pathlib.Path with URIs
https://api.github.com/repos/huggingface/datasets/issues/5466/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/121845112?v=4", "events_url": "https://api.github.com/users/jonny-cyberhaven/events{/privacy}", "followers_url": "https://api.github.com/users/jonny-cyberhaven/followers", "following_url": "https://api.github.com/users/jonny-cyberhaven/following{/other_user}", "gists_url": "https://api.github.com/users/jonny-cyberhaven/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jonny-cyberhaven", "id": 121845112, "login": "jonny-cyberhaven", "node_id": "U_kgDOB0M1eA", "organizations_url": "https://api.github.com/users/jonny-cyberhaven/orgs", "received_events_url": "https://api.github.com/users/jonny-cyberhaven/received_events", "repos_url": "https://api.github.com/users/jonny-cyberhaven/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jonny-cyberhaven/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonny-cyberhaven/subscriptions", "type": "User", "url": "https://api.github.com/users/jonny-cyberhaven" }
[]
null
null
CONTRIBUTOR
2023-01-26T16:59:11Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5466.diff", "html_url": "https://github.com/huggingface/datasets/pull/5466", "merged_at": "2023-01-26T16:59:11Z", "patch_url": "https://github.com/huggingface/datasets/pull/5466.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5466" }
PR_kwDODunzps5Ij-z1
[ "Thanks !\r\n`os.path.join` will use a backslash `\\` on windows which will also fail. You can use this instead in `load_from_disk`:\r\n```python\r\nfrom .filesystems import is_remote_filesystem\r\n\r\nis_local = not is_remote_filesystem(fs)\r\npath_join = os.path.join if is_local else posixpath.join\r\n```", "Th...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5466/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5466
https://github.com/huggingface/datasets/pull/5466
true
1,557,510,618
https://api.github.com/repos/huggingface/datasets/issues/5465/labels{/name}
### Describe the bug The structure of my dataset folder called "my_dataset" is : data metadata.csv The data folder consists of all mp3 files and metadata.csv consist of file locations like 'data/...mp3 and transcriptions. There's 400+ mp3 files and corresponding transcriptions for my dataset. When I run the following: ds = load_dataset("audiofolder", data_dir="my_dataset") I get: Using custom data configuration default-... Downloading and preparing dataset audiofolder/default to /... Downloading data files: 0%| | 0/2 [00:00<?, ?it/s] Downloading data files: 0it [00:00, ?it/s] Extracting data files: 0it [00:00, ?it/s] Generating train split: 0 examples [00:00, ? examples/s] Dataset audiofolder downloaded and prepared to /.... Subsequent calls will reuse this data. 0%| | 0/1 [00:00<?, ?it/s] DatasetDict({ train: Dataset({ features: ['audio', 'transcription'], num_rows: 1 }) }) ### Steps to reproduce the bug Create a dataset folder called 'my_dataset' with a subfolder called 'data' that has mp3 files. Also, create metadata.csv that has file locations like 'data/...mp3' and their corresponding transcription. Run: ds = load_dataset("audiofolder", data_dir="my_dataset") ### Expected behavior It should generate a dataset with numerous rows. ### Environment info Run on Jupyter notebook
2023-01-26T08:48:45Z
5,465
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-26T01:45:45Z
https://api.github.com/repos/huggingface/datasets/issues/5465/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5465/timeline
audiofolder creates empty dataset even though the dataset passed in follows the correct structure
https://api.github.com/repos/huggingface/datasets/issues/5465/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/107211437?v=4", "events_url": "https://api.github.com/users/jcho19/events{/privacy}", "followers_url": "https://api.github.com/users/jcho19/followers", "following_url": "https://api.github.com/users/jcho19/following{/other_user}", "gists_url": "https://api.github.com/users/jcho19/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jcho19", "id": 107211437, "login": "jcho19", "node_id": "U_kgDOBmPqrQ", "organizations_url": "https://api.github.com/users/jcho19/orgs", "received_events_url": "https://api.github.com/users/jcho19/received_events", "repos_url": "https://api.github.com/users/jcho19/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jcho19/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jcho19/subscriptions", "type": "User", "url": "https://api.github.com/users/jcho19" }
[]
null
completed
NONE
2023-01-26T08:48:45Z
null
I_kwDODunzps5c1bna
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5465/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5465
https://github.com/huggingface/datasets/issues/5465
false
1,557,462,104
https://api.github.com/repos/huggingface/datasets/issues/5464/labels{/name}
### Describe the bug The checksum of the file has likely changed on the remote host. ### Steps to reproduce the bug `dataset = nlp.load_dataset("hendrycks_test", "anatomy")` ### Expected behavior no error thrown ### Environment info - `datasets` version: 2.2.1 - Platform: macOS-13.1-arm64-arm-64bit - Python version: 3.9.13 - PyArrow version: 9.0.0 - Pandas version: 1.5.1
2023-01-27T05:44:31Z
5,464
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-26T00:43:23Z
https://api.github.com/repos/huggingface/datasets/issues/5464/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5464/timeline
NonMatchingChecksumError for hendrycks_test
https://api.github.com/repos/huggingface/datasets/issues/5464/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8027676?v=4", "events_url": "https://api.github.com/users/sarahwie/events{/privacy}", "followers_url": "https://api.github.com/users/sarahwie/followers", "following_url": "https://api.github.com/users/sarahwie/following{/other_user}", "gists_url": "https://api.github.com/users/sarahwie/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sarahwie", "id": 8027676, "login": "sarahwie", "node_id": "MDQ6VXNlcjgwMjc2NzY=", "organizations_url": "https://api.github.com/users/sarahwie/orgs", "received_events_url": "https://api.github.com/users/sarahwie/received_events", "repos_url": "https://api.github.com/users/sarahwie/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sarahwie/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sarahwie/subscriptions", "type": "User", "url": "https://api.github.com/users/sarahwie" }
[]
null
completed
NONE
2023-01-26T07:41:58Z
null
I_kwDODunzps5c1PxY
[ "Thanks for reporting, @sarahwie.\r\n\r\nPlease note this issue was already fixed in `datasets` 2.6.0 version:\r\n- #5040\r\n\r\nIf you update your `datasets` version, you will be able to load the dataset:\r\n```\r\npip install -U datasets\r\n```", "Oops, missed that I needed to upgrade. Thanks!" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5464/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5464
https://github.com/huggingface/datasets/issues/5464
false
1,557,021,041
https://api.github.com/repos/huggingface/datasets/issues/5463/labels{/name}
null
2023-01-25T18:33:35Z
5,463
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-25T17:24:01Z
https://api.github.com/repos/huggingface/datasets/issues/5463/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5463/timeline
Imagefolder docs: mention support of CSV and ZIP
https://api.github.com/repos/huggingface/datasets/issues/5463/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-01-25T18:26:15Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5463.diff", "html_url": "https://github.com/huggingface/datasets/pull/5463", "merged_at": "2023-01-25T18:26:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/5463.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5463" }
PR_kwDODunzps5IiGWb
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5463/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5463
https://github.com/huggingface/datasets/pull/5463
true
1,556,572,144
https://api.github.com/repos/huggingface/datasets/issues/5462/labels{/name}
Allow to concatenate on axis 1 two tables made of misaligned blocks. For example if the first table has 2 row blocks of 3 rows each, and the second table has 3 row blocks or 2 rows each. To do that, I slice the row blocks to re-align the blocks. Fix https://github.com/huggingface/datasets/issues/5413
2023-01-26T09:37:00Z
5,462
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-25T12:33:22Z
https://api.github.com/repos/huggingface/datasets/issues/5462/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5462/timeline
Concatenate on axis=1 with misaligned blocks
https://api.github.com/repos/huggingface/datasets/issues/5462/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-01-26T09:27:19Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5462.diff", "html_url": "https://github.com/huggingface/datasets/pull/5462", "merged_at": "2023-01-26T09:27:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/5462.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5462" }
PR_kwDODunzps5Iglqu
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5462/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5462
https://github.com/huggingface/datasets/pull/5462
true
1,555,532,719
https://api.github.com/repos/huggingface/datasets/issues/5461/labels{/name}
### Describe the bug I think there is a discrepancy between depth map of `nyu_depth_v2` dataset [here](https://huggingface.co/docs/datasets/main/en/depth_estimation) and actual depth map. Depth values somehow got **discretized/clipped** resulting in depth maps that are different from actual ones. Here is a side-by-side comparison, ![image](https://user-images.githubusercontent.com/36858976/214381162-1d9582c2-6750-4114-a01a-61ca1cd5f872.png) I tried to find the origin of this issue but sadly as I mentioned in tensorflow/datasets/issues/4674, the download link from `fast-depth` doesn't work anymore hence couldn't verify if the error originated there or during porting data from there to HF. Hi @sayakpaul, as you worked on huggingface/datasets/issues/5255, if you still have access to that data could you please share the data or perhaps checkout this issue? ### Steps to reproduce the bug This [notebook](https://colab.research.google.com/drive/1K3ZU8XUPRDOYD38MQS9nreQXJYitlKSW?usp=sharing#scrollTo=UEW7QSh0jf0i) from @sayakpaul could be used to generate depth maps and actual ground truths could be checked from this [dataset](https://www.kaggle.com/datasets/awsaf49/nyuv2-bts-dataset) from BTS repo. > Note: BTS dataset has only 36K data compared to the train-test 50K. They sampled the data as adjacent frames look quite the same ### Expected behavior Expected depth maps should be smooth rather than discrete/clipped. ### Environment info - `datasets` version: 2.8.1.dev0 - Platform: Linux-5.10.147+-x86_64-with-glibc2.29 - Python version: 3.8.10 - PyArrow version: 9.0.0 - Pandas version: 1.3.5
2023-02-06T20:52:00Z
5,461
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-24T19:15:46Z
https://api.github.com/repos/huggingface/datasets/issues/5461/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5461/timeline
Discrepancy in `nyu_depth_v2` dataset
https://api.github.com/repos/huggingface/datasets/issues/5461/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/36858976?v=4", "events_url": "https://api.github.com/users/awsaf49/events{/privacy}", "followers_url": "https://api.github.com/users/awsaf49/followers", "following_url": "https://api.github.com/users/awsaf49/following{/other_user}", "gists_url": "https://api.github.com/users/awsaf49/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/awsaf49", "id": 36858976, "login": "awsaf49", "node_id": "MDQ6VXNlcjM2ODU4OTc2", "organizations_url": "https://api.github.com/users/awsaf49/orgs", "received_events_url": "https://api.github.com/users/awsaf49/received_events", "repos_url": "https://api.github.com/users/awsaf49/repos", "site_admin": false, "starred_url": "https://api.github.com/users/awsaf49/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/awsaf49/subscriptions", "type": "User", "url": "https://api.github.com/users/awsaf49" }
[]
null
null
CONTRIBUTOR
null
null
I_kwDODunzps5ct4uv
[ "Ccing @dwofk (the author of `fast-depth`). \r\n\r\nThanks, @awsaf49 for reporting this. I believe this is because the NYU Depth V2 shipped from `fast-depth` is already preprocessed. \r\n\r\nIf you think it might be better to have the NYU Depth V2 dataset from BTS [here](https://huggingface.co/datasets/sayakpaul/ny...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5461/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5461
https://github.com/huggingface/datasets/issues/5461
false
1,555,387,532
https://api.github.com/repos/huggingface/datasets/issues/5460/labels{/name}
null
2023-01-25T16:11:10Z
5,460
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-24T17:33:38Z
https://api.github.com/repos/huggingface/datasets/issues/5460/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5460/timeline
Document that removing all the columns returns an empty document and the num_row is lost
https://api.github.com/repos/huggingface/datasets/issues/5460/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "events_url": "https://api.github.com/users/thomasw21/events{/privacy}", "followers_url": "https://api.github.com/users/thomasw21/followers", "following_url": "https://api.github.com/users/thomasw21/following{/other_user}", "gists_url": "https://api.github.com/users/thomasw21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thomasw21", "id": 24695242, "login": "thomasw21", "node_id": "MDQ6VXNlcjI0Njk1MjQy", "organizations_url": "https://api.github.com/users/thomasw21/orgs", "received_events_url": "https://api.github.com/users/thomasw21/received_events", "repos_url": "https://api.github.com/users/thomasw21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thomasw21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomasw21/subscriptions", "type": "User", "url": "https://api.github.com/users/thomasw21" }
[]
null
null
CONTRIBUTOR
2023-01-25T16:04:03Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5460.diff", "html_url": "https://github.com/huggingface/datasets/pull/5460", "merged_at": "2023-01-25T16:04:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/5460.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5460" }
PR_kwDODunzps5Icn9C
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5460/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5460
https://github.com/huggingface/datasets/pull/5460
true
1,555,367,504
https://api.github.com/repos/huggingface/datasets/issues/5459/labels{/name}
The library `aiohttp` performs a requoting of redirection URLs that unquotes the single quotation mark character: `%27` => `'` This is a problem for our Hugging Face Hub, which requires exact URL from location header. Specifically, in the query component of the URL (`https://netloc/path?query`), the value for `response-content-disposition` contains `%27`: ``` response-content-disposition=attachment%3B+filename*%3DUTF-8%27%27sample.jsonl.gz%3B+filename%3D%22sample.jsonl.gz%22%3B ``` and after the requoting, the `%27` characters get unquoted to `'`: ``` response-content-disposition=attachment%3B+filename*%3DUTF-8''sample.jsonl.gz%3B+filename%3D%22sample.jsonl.gz%22%3B ``` This PR disables the `aiohttp` requoting of redirection URLs.
2023-02-01T08:45:33Z
5,459
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-24T17:18:59Z
https://api.github.com/repos/huggingface/datasets/issues/5459/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5459/timeline
Disable aiohttp requoting of redirection URL
https://api.github.com/repos/huggingface/datasets/issues/5459/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-31T08:37:54Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5459.diff", "html_url": "https://github.com/huggingface/datasets/pull/5459", "merged_at": "2023-01-31T08:37:54Z", "patch_url": "https://github.com/huggingface/datasets/pull/5459.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5459" }
PR_kwDODunzps5Icjwe
[ "_The documentation is not available anymore as the PR was closed or merged._", "Comment by @lhoestq:\r\n> Do you think we need this in `datasets` if it's fixed on the moon landing side ? In the aiohttp doc they consider those symbols as \"non-safe\" ", "The lib `requests` does not perform that requote on redir...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5459/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5459
https://github.com/huggingface/datasets/pull/5459
true
1,555,054,737
https://api.github.com/repos/huggingface/datasets/issues/5458/labels{/name}
### Describe the bug When using the `load_dataset` function with streaming set to True, slicing splits is apparently not supported. Did I miss this in the documentation? ### Steps to reproduce the bug `load_dataset("lhoestq/demo1",revision=None, streaming=True, split="train[:3]")` causes ValueError: Bad split: train[:3]. Available splits: ['train', 'test'] in builder.py, line 1213, in as_streaming_dataset ### Expected behavior The first 3 entries of the dataset as a stream ### Environment info - `datasets` version: 2.8.0 - Platform: Windows-10-10.0.19045-SP0 - Python version: 3.10.9 - PyArrow version: 10.0.1 - Pandas version: 1.5.2
2023-01-24T15:11:47Z
5,458
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-24T14:08:17Z
https://api.github.com/repos/huggingface/datasets/issues/5458/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5458/timeline
slice split while streaming
https://api.github.com/repos/huggingface/datasets/issues/5458/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/122370631?v=4", "events_url": "https://api.github.com/users/SvenDS9/events{/privacy}", "followers_url": "https://api.github.com/users/SvenDS9/followers", "following_url": "https://api.github.com/users/SvenDS9/following{/other_user}", "gists_url": "https://api.github.com/users/SvenDS9/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SvenDS9", "id": 122370631, "login": "SvenDS9", "node_id": "U_kgDOB0s6Rw", "organizations_url": "https://api.github.com/users/SvenDS9/orgs", "received_events_url": "https://api.github.com/users/SvenDS9/received_events", "repos_url": "https://api.github.com/users/SvenDS9/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SvenDS9/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SvenDS9/subscriptions", "type": "User", "url": "https://api.github.com/users/SvenDS9" }
[]
null
completed
NONE
2023-01-24T15:11:47Z
null
I_kwDODunzps5csECR
[ "Hi! Yes, that's correct. When `streaming` is `True`, only split names can be specified as `split`, and for slicing, you have to use `.skip`/`.take` instead.\r\n\r\nE.g. \r\n`load_dataset(\"lhoestq/demo1\",revision=None, streaming=True, split=\"train[:3]\")`\r\n\r\nrewritten with `.skip`/`.take`:\r\n`load_dataset(\...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5458/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5458
https://github.com/huggingface/datasets/issues/5458
false
1,554,171,264
https://api.github.com/repos/huggingface/datasets/issues/5457/labels{/name}
### Describe the bug I pre-built the dataset: ``` python -c 'import sys; from datasets import load_dataset; ds=load_dataset(sys.argv[1])' HuggingFaceM4/general-pmd-synthetic-testing ``` and it can be used just fine. now I wipe out `downloads/extracted` and it no longer works. ``` rm -r ~/.cache/huggingface/datasets/downloads ``` That is I can still load it: ``` python -c 'import sys; from datasets import load_dataset; ds=load_dataset(sys.argv[1])' HuggingFaceM4/general-pmd-synthetic-testing No config specified, defaulting to: general-pmd-synthetic-testing/100.unique Found cached dataset general-pmd-synthetic-testing (/home/stas/.cache/huggingface/datasets/HuggingFaceM4___general-pmd-synthetic-testing/100.unique/1.1.1/86bc445e3e48cb5ef79de109eb4e54ff85b318cd55c3835c4ee8f86eae33d9d2) ``` but if I try to use it: ``` E stderr: Traceback (most recent call last): E stderr: File "/mnt/nvme0/code/huggingface/m4-master-6/m4/training/main.py", line 116, in <module> E stderr: train_loader, val_loader = get_dataloaders( E stderr: File "/mnt/nvme0/code/huggingface/m4-master-6/m4/training/dataset.py", line 170, in get_dataloaders E stderr: train_loader = get_dataloader_from_config( E stderr: File "/mnt/nvme0/code/huggingface/m4-master-6/m4/training/dataset.py", line 443, in get_dataloader_from_config E stderr: dataloader = get_dataloader( E stderr: File "/mnt/nvme0/code/huggingface/m4-master-6/m4/training/dataset.py", line 264, in get_dataloader E stderr: is_pmd = "meta" in hf_dataset[0] and "source" in hf_dataset[0] E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/arrow_dataset.py", line 2601, in __getitem__ E stderr: return self._getitem( E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/arrow_dataset.py", line 2586, in _getitem E stderr: formatted_output = format_table( E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/formatting/formatting.py", line 634, in format_table E stderr: return formatter(pa_table, query_type=query_type) E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/formatting/formatting.py", line 406, in __call__ E stderr: return self.format_row(pa_table) E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/formatting/formatting.py", line 442, in format_row E stderr: row = self.python_features_decoder.decode_row(row) E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/formatting/formatting.py", line 225, in decode_row E stderr: return self.features.decode_example(row) if self.features else row E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/features/features.py", line 1846, in decode_example E stderr: return { E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/features/features.py", line 1847, in <dictcomp> E stderr: column_name: decode_nested_example(feature, value, token_per_repo_id=token_per_repo_id) E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/features/features.py", line 1304, in decode_nested_example E stderr: return decode_nested_example([schema.feature], obj) E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/features/features.py", line 1296, in decode_nested_example E stderr: if decode_nested_example(sub_schema, first_elmt) != first_elmt: E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/features/features.py", line 1309, in decode_nested_example E stderr: return schema.decode_example(obj, token_per_repo_id=token_per_repo_id) E stderr: File "/mnt/nvme0/code/huggingface/datasets-master/src/datasets/features/image.py", line 144, in decode_example E stderr: image = PIL.Image.open(path) E stderr: File "/home/stas/anaconda3/envs/py38-pt113/lib/python3.8/site-packages/PIL/Image.py", line 3092, in open E stderr: fp = builtins.open(filename, "rb") E stderr: FileNotFoundError: [Errno 2] No such file or directory: '/mnt/nvme0/code/data/cache/huggingface/datasets/downloads/extracted/134227b9b94c4eccf19b205bf3021d4492d0227b9be6c2ddb6bf517d8d55a8cb/data/101/images_01.jpg' ``` Only if I wipe out the cached dir and rebuild then it starts working as `download/extracted` is back again with extracted files. ``` rm -r ~/.cache/huggingface/datasets/HuggingFaceM4___general-pmd-synthetic-testing python -c 'import sys; from datasets import load_dataset; ds=load_dataset(sys.argv[1])' HuggingFaceM4/general-pmd-synthetic-testing ``` I think there are 2 issues here: 1. why does it still rely on extracted files after `arrow` files were printed - did I do something incorrectly when creating this dataset? 2. why doesn't the dataset know that it has been gutted and loads just fine? If it has a dependency on `download/extracted` then `load_dataset` should check if it's there and fail or force rebuilding. I am sure this could be a very expensive operation, so probably really solving #1 will not require this check. and this second item is probably an overkill. Other than perhaps if it had an optional `check_consistency` flag to do that. ### Environment info datasets@main
2023-01-24T18:14:10Z
5,457
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-24T02:09:32Z
https://api.github.com/repos/huggingface/datasets/issues/5457/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5457/timeline
prebuilt dataset relies on `downloads/extracted`
https://api.github.com/repos/huggingface/datasets/issues/5457/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stas00", "id": 10676103, "login": "stas00", "node_id": "MDQ6VXNlcjEwNjc2MTAz", "organizations_url": "https://api.github.com/users/stas00/orgs", "received_events_url": "https://api.github.com/users/stas00/received_events", "repos_url": "https://api.github.com/users/stas00/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "type": "User", "url": "https://api.github.com/users/stas00" }
[]
null
null
CONTRIBUTOR
null
null
I_kwDODunzps5cosWA
[ "Hi! \r\n\r\nThis issue is due to our audio/image datasets not being self-contained. This allows us to save disk space (files are written only once) but also leads to the issues like this one. We plan to make all our datasets self-contained in Datasets 3.0.\r\n\r\nIn the meantime, you can run the following map to e...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5457/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5457
https://github.com/huggingface/datasets/issues/5457
false
1,553,905,148
https://api.github.com/repos/huggingface/datasets/issues/5456/labels{/name}
As described in #5418 I noticed also that the `to_json` function supports multi-workers whereas `to_parquet`, is that not possible/not needed with Parquet or something that hasn't been implemented yet?
2023-01-24T11:26:47Z
5,456
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-23T22:05:38Z
https://api.github.com/repos/huggingface/datasets/issues/5456/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5456/timeline
feat: tqdm for `to_parquet`
https://api.github.com/repos/huggingface/datasets/issues/5456/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/33707069?v=4", "events_url": "https://api.github.com/users/zanussbaum/events{/privacy}", "followers_url": "https://api.github.com/users/zanussbaum/followers", "following_url": "https://api.github.com/users/zanussbaum/following{/other_user}", "gists_url": "https://api.github.com/users/zanussbaum/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zanussbaum", "id": 33707069, "login": "zanussbaum", "node_id": "MDQ6VXNlcjMzNzA3MDY5", "organizations_url": "https://api.github.com/users/zanussbaum/orgs", "received_events_url": "https://api.github.com/users/zanussbaum/received_events", "repos_url": "https://api.github.com/users/zanussbaum/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zanussbaum/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zanussbaum/subscriptions", "type": "User", "url": "https://api.github.com/users/zanussbaum" }
[]
null
null
CONTRIBUTOR
2023-01-24T11:17:12Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5456.diff", "html_url": "https://github.com/huggingface/datasets/pull/5456", "merged_at": "2023-01-24T11:17:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/5456.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5456" }
PR_kwDODunzps5IXq92
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5456/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5456
https://github.com/huggingface/datasets/pull/5456
true
1,553,040,080
https://api.github.com/repos/huggingface/datasets/issues/5455/labels{/name}
Use the "shard generator approach with periodic progress updates" (used in `save_to_disk` and multi-proc `load_dataset`) in `Dataset.map` to enable having a single TQDM progress bar in the multi-proc mode. Closes https://github.com/huggingface/datasets/issues/771, closes https://github.com/huggingface/datasets/issues/3177 TODO: - [x] cleaner refactor of the `_map_single` decorators now that they also have to wrap generator functions (decorate `map` instead of `map_single` with the `transmit_` decorators and predict the shards' fingerprint in `map`)
2023-02-13T20:23:34Z
5,455
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-23T12:49:40Z
https://api.github.com/repos/huggingface/datasets/issues/5455/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5455/timeline
Single TQDM bar in multi-proc map
https://api.github.com/repos/huggingface/datasets/issues/5455/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2023-02-13T20:16:38Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5455.diff", "html_url": "https://github.com/huggingface/datasets/pull/5455", "merged_at": "2023-02-13T20:16:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/5455.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5455" }
PR_kwDODunzps5IUvAZ
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 1, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/5455/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5455
https://github.com/huggingface/datasets/pull/5455
true
1,552,890,419
https://api.github.com/repos/huggingface/datasets/issues/5454/labels{/name}
It would be nice when using `datasets` with a PyTorch DataLoader to be able to resume a training from a DataLoader state (e.g. to resume a training that crashed) What I have in mind (but lmk if you have other ideas or comments): For map-style datasets, this requires to have a PyTorch Sampler state that can be saved and reloaded per node and worker. For iterable datasets, this requires to save the state of the dataset iterator, which includes: - the current shard idx and row position in the current shard - the epoch number - the rng state - the shuffle buffer Right now you can already resume the data loading of an iterable dataset by using `IterableDataset.skip` but it takes a lot of time because it re-iterates on all the past data until it reaches the resuming point. cc @stas00 @sgugger
2024-03-19T02:50:44Z
5,454
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "c5def5", "default": fals...
2023-01-23T10:58:54Z
https://api.github.com/repos/huggingface/datasets/issues/5454/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5454/timeline
Save and resume the state of a DataLoader
https://api.github.com/repos/huggingface/datasets/issues/5454/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
null
null
I_kwDODunzps5cjzoz
[ "Something that'd be nice to have is \"manual update of state\". One of the learning from training LLMs is the ability to skip some batches whenever we notice huge spike might be handy.", "Your outline spec is very sound and clear, @lhoestq - thank you!\r\n\r\n@thomasw21, indeed that would be a wonderful extra fe...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 5, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 6, "url": "https://api.github.com/repos/huggingface/datasets/issues/5454/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5454
https://github.com/huggingface/datasets/issues/5454
false
1,552,727,425
https://api.github.com/repos/huggingface/datasets/issues/5453/labels{/name}
This PR fixes the extraction of insecure TAR files by changing the base path against which TAR members are compared: - from: "." - to: `output_path` This PR also adds tests for extracting insecure TAR files. Related to: - #5441 - #5452 @stas00 please note this PR addresses just one of the issues you pointed out: the use of the cwd by the extractor. The other issues (actionable error messages, raise instead of log error) should be addressed in other PRs.
2023-01-24T01:34:20Z
5,453
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-23T08:57:40Z
https://api.github.com/repos/huggingface/datasets/issues/5453/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5453/timeline
Fix base directory while extracting insecure TAR files
https://api.github.com/repos/huggingface/datasets/issues/5453/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-23T10:10:42Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5453.diff", "html_url": "https://github.com/huggingface/datasets/pull/5453", "merged_at": "2023-01-23T10:10:42Z", "patch_url": "https://github.com/huggingface/datasets/pull/5453.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5453" }
PR_kwDODunzps5ITraa
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5453/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5453
https://github.com/huggingface/datasets/pull/5453
true
1,552,655,939
https://api.github.com/repos/huggingface/datasets/issues/5452/labels{/name}
The log messages do not match their if-condition. This PR swaps them. Found while investigating: - #5441 CC: @lhoestq
2023-01-23T09:40:55Z
5,452
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-23T07:53:38Z
https://api.github.com/repos/huggingface/datasets/issues/5452/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5452/timeline
Swap log messages for symbolic/hard links in tar extractor
https://api.github.com/repos/huggingface/datasets/issues/5452/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-23T08:31:17Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5452.diff", "html_url": "https://github.com/huggingface/datasets/pull/5452", "merged_at": "2023-01-23T08:31:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/5452.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5452" }
PR_kwDODunzps5ITcA3
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5452/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5452
https://github.com/huggingface/datasets/pull/5452
true
1,552,336,300
https://api.github.com/repos/huggingface/datasets/issues/5451/labels{/name}
### Describe the bug I'm getting the following exception: ``` lib/python3.10/zipfile.py:1353 in _RealGetContents β”‚ β”‚ β”‚ β”‚ 1350 β”‚ β”‚ # self.start_dir: Position of start of central directory β”‚ β”‚ 1351 β”‚ β”‚ self.start_dir = offset_cd + concat β”‚ β”‚ 1352 β”‚ β”‚ if self.start_dir < 0: β”‚ β”‚ ❱ 1353 β”‚ β”‚ β”‚ raise BadZipFile("Bad offset for central directory") β”‚ β”‚ 1354 β”‚ β”‚ fp.seek(self.start_dir, 0) β”‚ β”‚ 1355 β”‚ β”‚ data = fp.read(size_cd) β”‚ β”‚ 1356 β”‚ β”‚ fp = io.BytesIO(data) β”‚ ╰──────────────────────────────────────────────────────────────────────────────────────────────────╯ BadZipFile: Bad offset for central directory Extracting data files: 35%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 38572/110812 [00:10<00:20, 3576.26it/s] ``` ### Steps to reproduce the bug ``` load_dataset( args.dataset_name, args.dataset_config_name, cache_dir=args.cache_dir, ), ``` ### Expected behavior loads the dataset ### Environment info datasets==2.8.0 Python 3.10.8 Linux 129-146-3-202 5.15.0-52-generic #58~20.04.1-Ubuntu SMP Thu Oct 13 13:09:46 UTC 2022 x86_64 x86_64 x86_64 GNU/Linux
2023-05-23T10:35:48Z
5,451
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-22T23:50:12Z
https://api.github.com/repos/huggingface/datasets/issues/5451/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5451/timeline
ImageFolder BadZipFile: Bad offset for central directory
https://api.github.com/repos/huggingface/datasets/issues/5451/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1524208?v=4", "events_url": "https://api.github.com/users/hmartiro/events{/privacy}", "followers_url": "https://api.github.com/users/hmartiro/followers", "following_url": "https://api.github.com/users/hmartiro/following{/other_user}", "gists_url": "https://api.github.com/users/hmartiro/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hmartiro", "id": 1524208, "login": "hmartiro", "node_id": "MDQ6VXNlcjE1MjQyMDg=", "organizations_url": "https://api.github.com/users/hmartiro/orgs", "received_events_url": "https://api.github.com/users/hmartiro/received_events", "repos_url": "https://api.github.com/users/hmartiro/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hmartiro/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hmartiro/subscriptions", "type": "User", "url": "https://api.github.com/users/hmartiro" }
[]
null
completed
NONE
2023-02-10T16:31:36Z
null
I_kwDODunzps5chsWs
[ "Hi ! Could you share the full stack trace ? Which dataset did you try to load ?\r\n\r\nit may be related to https://github.com/huggingface/datasets/pull/5640", "The `BadZipFile` error means the ZIP file is corrupted, so I'm closing this issue as it's not directly related to `datasets`.", "For others that find ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5451/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5451
https://github.com/huggingface/datasets/issues/5451
false
1,551,109,365
https://api.github.com/repos/huggingface/datasets/issues/5450/labels{/name}
### Describe the bug This will make more sense if you take a look at [a Colab notebook that reproduces this issue.](https://colab.research.google.com/drive/1rxyeciQFWJTI0WrZ5aojp4Ls1ut18fNH?usp=sharing) Briefly, there are several datasets that, when you iterate over them with `to_tf_dataset` **and** a data collator that returns `tf` tensors, become very slow. We haven't been able to figure this one out - it can be intermittent, and we have no idea what could possibly cause it. The weirdest thing is that **the slowdown affects other attempts to access the underlying dataset**. If you try to iterate over the `tf.data.Dataset`, then interrupt execution, and then try to iterate over the original dataset, the original dataset is now also very slow! This is true even if the dataset format is not set to `tf` - the iteration is slow even though it's not calling TF at all! There is a simple workaround for this - we can simply get our data collators to return `np` tensors. When we do this, the bug is never triggered and everything is fine. In general, `np` is preferred for this kind of preprocessing work anyway, when the preprocessing is not going to be compiled into a pure `tf.data` pipeline! However, the issue is fascinating, and the TF team were wondering if anyone in datasets (cc @lhoestq @mariosasko) might have an idea of what could cause this. ### Steps to reproduce the bug Run the attached Colab. ### Expected behavior The slowdown should go away, or at least not persist after we stop iterating over the `tf.data.Dataset` ### Environment info The issue occurs on multiple versions of Python and TF, both on local machines and on Colab. All testing was done using the latest versions of `transformers` and `datasets` from `main`
2023-02-13T14:13:34Z
5,450
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-20T16:08:37Z
https://api.github.com/repos/huggingface/datasets/issues/5450/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5450/timeline
to_tf_dataset with a TF collator causes bizarrely persistent slowdown
https://api.github.com/repos/huggingface/datasets/issues/5450/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4", "events_url": "https://api.github.com/users/Rocketknight1/events{/privacy}", "followers_url": "https://api.github.com/users/Rocketknight1/followers", "following_url": "https://api.github.com/users/Rocketknight1/following{/other_user}", "gists_url": "https://api.github.com/users/Rocketknight1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Rocketknight1", "id": 12866554, "login": "Rocketknight1", "node_id": "MDQ6VXNlcjEyODY2NTU0", "organizations_url": "https://api.github.com/users/Rocketknight1/orgs", "received_events_url": "https://api.github.com/users/Rocketknight1/received_events", "repos_url": "https://api.github.com/users/Rocketknight1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Rocketknight1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Rocketknight1/subscriptions", "type": "User", "url": "https://api.github.com/users/Rocketknight1" }
[]
null
completed
MEMBER
2023-02-13T14:13:34Z
null
I_kwDODunzps5cdAz1
[ "wtf", "Couldn't find what's causing this, this will need more investigation", "A possible hint: The function it seems to be spending a lot of time in (when iterating over the original dataset) is `_get_mp` in the PIL JPEG decoder: \r\n![image](https://user-images.githubusercontent.com/12866554/214057267-c889f0...
{ "+1": 0, "-1": 0, "confused": 1, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5450/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5450
https://github.com/huggingface/datasets/issues/5450
false
1,550,801,453
https://api.github.com/repos/huggingface/datasets/issues/5449/labels{/name}
Support fsspec 2023.1.0 in CI. In the 2023.1.0 fsspec release, they replaced the type of `fsspec.registry`: - from `ReadOnlyRegistry`, with an attribute called `target` - to `MappingProxyType`, without that attribute Consequently, we need to change our `mock_fsspec` fixtures, that were using the `target` attribute. Fix #5448.
2023-01-20T13:32:50Z
5,449
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-20T12:53:17Z
https://api.github.com/repos/huggingface/datasets/issues/5449/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5449/timeline
Support fsspec 2023.1.0 in CI
https://api.github.com/repos/huggingface/datasets/issues/5449/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-20T13:26:03Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5449.diff", "html_url": "https://github.com/huggingface/datasets/pull/5449", "merged_at": "2023-01-20T13:26:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/5449.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5449" }
PR_kwDODunzps5INgD9
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5449/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5449
https://github.com/huggingface/datasets/pull/5449
true
1,550,618,514
https://api.github.com/repos/huggingface/datasets/issues/5448/labels{/name}
Once we find out the root cause of: - #5445 we should revert the temporary pin on fsspec introduced by: - #5447
2023-01-20T13:26:05Z
5,448
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-01-20T10:26:31Z
https://api.github.com/repos/huggingface/datasets/issues/5448/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5448/timeline
Support fsspec 2023.1.0 in CI
https://api.github.com/repos/huggingface/datasets/issues/5448/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
MEMBER
2023-01-20T13:26:05Z
null
I_kwDODunzps5cbI-S
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5448/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5448
https://github.com/huggingface/datasets/issues/5448
false
1,550,599,193
https://api.github.com/repos/huggingface/datasets/issues/5447/labels{/name}
Temporarily pin fsspec < 2023.1.0 Fix #5445.
2023-01-20T10:38:13Z
5,447
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-20T10:11:02Z
https://api.github.com/repos/huggingface/datasets/issues/5447/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5447/timeline
Fix CI by temporarily pinning fsspec < 2023.1.0
https://api.github.com/repos/huggingface/datasets/issues/5447/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-20T10:28:43Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5447.diff", "html_url": "https://github.com/huggingface/datasets/pull/5447", "merged_at": "2023-01-20T10:28:43Z", "patch_url": "https://github.com/huggingface/datasets/pull/5447.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5447" }
PR_kwDODunzps5IM0Nu
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5447/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5447
https://github.com/huggingface/datasets/pull/5447
true
1,550,591,588
https://api.github.com/repos/huggingface/datasets/issues/5446/labels{/name}
DO NOT MERGE. Only to test the CI. cc @lhoestq @albertvillanova
2023-01-20T10:43:22Z
5,446
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-20T10:05:19Z
https://api.github.com/repos/huggingface/datasets/issues/5446/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5446/timeline
test v0.12.0.rc0
https://api.github.com/repos/huggingface/datasets/issues/5446/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/11801849?v=4", "events_url": "https://api.github.com/users/Wauplin/events{/privacy}", "followers_url": "https://api.github.com/users/Wauplin/followers", "following_url": "https://api.github.com/users/Wauplin/following{/other_user}", "gists_url": "https://api.github.com/users/Wauplin/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Wauplin", "id": 11801849, "login": "Wauplin", "node_id": "MDQ6VXNlcjExODAxODQ5", "organizations_url": "https://api.github.com/users/Wauplin/orgs", "received_events_url": "https://api.github.com/users/Wauplin/received_events", "repos_url": "https://api.github.com/users/Wauplin/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Wauplin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wauplin/subscriptions", "type": "User", "url": "https://api.github.com/users/Wauplin" }
[]
null
null
CONTRIBUTOR
2023-01-20T10:13:48Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5446.diff", "html_url": "https://github.com/huggingface/datasets/pull/5446", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5446.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5446" }
PR_kwDODunzps5IMyka
[ "_The documentation is not available anymore as the PR was closed or merged._", "@Wauplin I was testing it in a dedicated branch without opening a PR: https://github.com/huggingface/datasets/commits/test-hfh-0.12.0rc0", "Oops, sorry @albertvillanova. I thought for next time I'll start the CIs before pinging eve...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5446/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5446
https://github.com/huggingface/datasets/pull/5446
true
1,550,588,703
https://api.github.com/repos/huggingface/datasets/issues/5445/labels{/name}
CI tests are broken, raising `AttributeError: 'mappingproxy' object has no attribute 'target'`. See: https://github.com/huggingface/datasets/actions/runs/3966497597/jobs/6797384185 ``` ... ERROR tests/test_streaming_download_manager.py::TestxPath::test_xpath_rglob[mock://top_level-date=2019-10-0[1-4]/*-expected_paths4] - AttributeError: 'mappingproxy' object has no attribute 'target' ===== 2076 passed, 19 skipped, 15 warnings, 47 errors in 115.54s (0:01:55) ===== ```
2023-01-20T10:28:44Z
5,445
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2023-01-20T10:03:10Z
https://api.github.com/repos/huggingface/datasets/issues/5445/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5445/timeline
CI tests are broken: AttributeError: 'mappingproxy' object has no attribute 'target'
https://api.github.com/repos/huggingface/datasets/issues/5445/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
MEMBER
2023-01-20T10:28:44Z
null
I_kwDODunzps5cbBsf
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5445/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5445
https://github.com/huggingface/datasets/issues/5445
false
1,550,185,071
https://api.github.com/repos/huggingface/datasets/issues/5444/labels{/name}
### Describe the bug Code in `datasets` is using `logger.warning` when it should be using `logger.info`. Some of these are probably a matter of opinion, but I think anything starting with `logger.warning(f"Loading chached` clearly falls into the info category. Definitions from the Python docs for reference: * INFO: Confirmation that things are working as expected. * WARNING: An indication that something unexpected happened, or indicative of some problem in the near future (e.g. β€˜disk space low’). The software is still working as expected. In theory, a user should be able to resolve things such that there are no warnings. ### Steps to reproduce the bug Load any dataset that's already cached. ### Expected behavior No output when log level is at the default WARNING level. ### Environment info - `datasets` version: 2.8.0 - Platform: Linux-5.10.102.1-microsoft-standard-WSL2-x86_64-with-glibc2.31 - Python version: 3.10.8 - PyArrow version: 9.0.0 - Pandas version: 1.5.2
2023-07-12T17:19:31Z
5,444
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-20T01:19:18Z
https://api.github.com/repos/huggingface/datasets/issues/5444/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5444/timeline
info messages logged as warnings
https://api.github.com/repos/huggingface/datasets/issues/5444/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/4443482?v=4", "events_url": "https://api.github.com/users/davidgilbertson/events{/privacy}", "followers_url": "https://api.github.com/users/davidgilbertson/followers", "following_url": "https://api.github.com/users/davidgilbertson/following{/other_user}", "gists_url": "https://api.github.com/users/davidgilbertson/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/davidgilbertson", "id": 4443482, "login": "davidgilbertson", "node_id": "MDQ6VXNlcjQ0NDM0ODI=", "organizations_url": "https://api.github.com/users/davidgilbertson/orgs", "received_events_url": "https://api.github.com/users/davidgilbertson/received_events", "repos_url": "https://api.github.com/users/davidgilbertson/repos", "site_admin": false, "starred_url": "https://api.github.com/users/davidgilbertson/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/davidgilbertson/subscriptions", "type": "User", "url": "https://api.github.com/users/davidgilbertson" }
[]
null
completed
NONE
2023-07-12T17:19:31Z
null
I_kwDODunzps5cZfJv
[ "Looks like a duplicate of https://github.com/huggingface/datasets/issues/1948. \r\n\r\nI also think these should be logged as INFO messages, but let's see what @lhoestq thinks.", "It can be considered unexpected to see a `map` function return instantaneously. The warning is here to explain this case by mentionin...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5444/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5444
https://github.com/huggingface/datasets/issues/5444
false
1,550,178,914
https://api.github.com/repos/huggingface/datasets/issues/5443/labels{/name}
Based on feedback from discussion #5423, this PR updates the sharing tutorial with a mention of writing your own dataset loading script to support more advanced dataset creation options like multiple configs. I'll open a separate PR to update the *Create a Dataset card* with the new Hub metadata UI update πŸ˜„
2023-01-20T15:44:45Z
5,443
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-20T01:09:14Z
https://api.github.com/repos/huggingface/datasets/issues/5443/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5443/timeline
Update share tutorial
https://api.github.com/repos/huggingface/datasets/issues/5443/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu" }
[]
null
null
MEMBER
2023-01-20T15:37:30Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5443.diff", "html_url": "https://github.com/huggingface/datasets/pull/5443", "merged_at": "2023-01-20T15:37:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/5443.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5443" }
PR_kwDODunzps5ILbk8
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5443/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5443
https://github.com/huggingface/datasets/pull/5443
true
1,550,084,450
https://api.github.com/repos/huggingface/datasets/issues/5442/labels{/name}
### Feature request First of all , I would like to thank all community who are developed DataSet storage and make it free available How to integrate our Onedrive account or any other possible storage clouds (like google drive,...) with the **HF** datasets section. For example, if I have **50GB** on my **Onedrive** account and I want to move between drive and Hugging face repo or vis versa ### Motivation make the dataset section more flexible with other possible storage like the integration between Google Collab and Google drive the storage ### Your contribution Can be done using Hugging face CLI
2023-02-24T16:17:51Z
5,442
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-01-19T23:12:08Z
https://api.github.com/repos/huggingface/datasets/issues/5442/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5442/timeline
OneDrive Integrations with HF Datasets
https://api.github.com/repos/huggingface/datasets/issues/5442/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/59222637?v=4", "events_url": "https://api.github.com/users/Mohammed20201991/events{/privacy}", "followers_url": "https://api.github.com/users/Mohammed20201991/followers", "following_url": "https://api.github.com/users/Mohammed20201991/following{/other_user}", "gists_url": "https://api.github.com/users/Mohammed20201991/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Mohammed20201991", "id": 59222637, "login": "Mohammed20201991", "node_id": "MDQ6VXNlcjU5MjIyNjM3", "organizations_url": "https://api.github.com/users/Mohammed20201991/orgs", "received_events_url": "https://api.github.com/users/Mohammed20201991/received_events", "repos_url": "https://api.github.com/users/Mohammed20201991/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Mohammed20201991/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Mohammed20201991/subscriptions", "type": "User", "url": "https://api.github.com/users/Mohammed20201991" }
[]
null
completed
NONE
2023-02-24T16:17:51Z
null
I_kwDODunzps5cZGli
[ "Hi! \r\n\r\nWe use [`fsspec`](https://github.com/fsspec/filesystem_spec) to integrate with storage providers. You can find more info (and the usage examples) in [our docs](https://huggingface.co/docs/datasets/v2.8.0/filesystems#download-and-prepare-a-dataset-into-a-cloud-storage).\r\n\r\n[`gdrivefs`](https://githu...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5442/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5442
https://github.com/huggingface/datasets/issues/5442
false
1,548,417,594
https://api.github.com/repos/huggingface/datasets/issues/5441/labels{/name}
ok, every so often, I have been getting a strange failure on dataset install: ``` $ python -c 'import sys; from datasets import load_dataset; ds=load_dataset(sys.argv[1])' HuggingFaceM4/general-pmd-synthetic-testing No config specified, defaulting to: general-pmd-synthetic-testing/100.unique Downloading and preparing dataset general-pmd-synthetic-testing/100.unique (download: 3.21 KiB, generated: 16.01 MiB, post-processed: Unknown size, total: 16.02 MiB) to /home/stas/.cache/huggingface/datasets/HuggingFaceM4___general-pmd-synthetic-testing/100.unique/1.1.1/86bc445e3e48cb5ef79de109eb4e54ff85b318cd55c3835c4ee8f86eae33d9d2... Extraction of data is blocked (illegal path) Extraction of data/1 is blocked (illegal path) Extraction of data/1/text.null is blocked (illegal path) [...] ``` I had no idea what to do with that - what in the world does **illegal path** mean? I started looking at the code in `TarExtractor` and added a debug print of `base` so that told me that there was a problem with the current directory - which was a clone of one of the hf repos. This particular dataset extracts into a directory `data` and the current dir I was running the tests from already had `data` in it which was a symbolic link to another partition and somehow all that `badpath` code was blowing up there. https://github.com/huggingface/datasets/blob/80eb8db74f49b7ee9c0f73a819c22177fabd61db/src/datasets/utils/extract.py#L113-L114 I tried hard to come up with a repro, but no matter what I tried it only fails in that particular clone directory that has a `data` symlink and not anywhere else. In any case, in this PR I'm proposing to at least give a user a hint of what seems to be an issue. I'm not at all happy with the info I got with this proposed change, but at least it gave me a hint that `TarExtractor` tries to extract into the current directory without any respect to pre-existing files. Say what? https://github.com/huggingface/datasets/blob/80eb8db74f49b7ee9c0f73a819c22177fabd61db/src/datasets/utils/extract.py#L110 why won't it use the `datasets` designated directory for that? There would never be a problem if it were to do that. I had to look at all those `resolved`, `badpath` calls and see what it did and why it failed, since it was far from obvious. It appeared like it resolved a symlink and compared it to the original path which of course wasn't matching. So perhaps you have a better solution than what I proposed in this PR. I think that code line I quoted is the one that should be fixed instead. But if you can't think of a better solution let's merge this at least so that the user will have a clue that the current dir is somehow involved. p.s. I double checked that if I remove the pre-existing `data` symlink in the current dir I'm running the dataset install command from, the problem goes away too. Thanks.
2023-01-20T16:49:22Z
5,441
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-19T02:17:21Z
https://api.github.com/repos/huggingface/datasets/issues/5441/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5441/timeline
resolving a weird tar extract issue
https://api.github.com/repos/huggingface/datasets/issues/5441/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "events_url": "https://api.github.com/users/stas00/events{/privacy}", "followers_url": "https://api.github.com/users/stas00/followers", "following_url": "https://api.github.com/users/stas00/following{/other_user}", "gists_url": "https://api.github.com/users/stas00/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stas00", "id": 10676103, "login": "stas00", "node_id": "MDQ6VXNlcjEwNjc2MTAz", "organizations_url": "https://api.github.com/users/stas00/orgs", "received_events_url": "https://api.github.com/users/stas00/received_events", "repos_url": "https://api.github.com/users/stas00/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stas00/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stas00/subscriptions", "type": "User", "url": "https://api.github.com/users/stas00" }
[]
null
null
CONTRIBUTOR
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/5441.diff", "html_url": "https://github.com/huggingface/datasets/pull/5441", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5441.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5441" }
PR_kwDODunzps5IFeCW
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5441/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5441
https://github.com/huggingface/datasets/pull/5441
true
1,538,361,143
https://api.github.com/repos/huggingface/datasets/issues/5440/labels{/name}
null
2023-01-18T17:57:29Z
5,440
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-18T17:04:27Z
https://api.github.com/repos/huggingface/datasets/issues/5440/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5440/timeline
Fix documentation about batch samplers
https://api.github.com/repos/huggingface/datasets/issues/5440/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/24695242?v=4", "events_url": "https://api.github.com/users/thomasw21/events{/privacy}", "followers_url": "https://api.github.com/users/thomasw21/followers", "following_url": "https://api.github.com/users/thomasw21/following{/other_user}", "gists_url": "https://api.github.com/users/thomasw21/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thomasw21", "id": 24695242, "login": "thomasw21", "node_id": "MDQ6VXNlcjI0Njk1MjQy", "organizations_url": "https://api.github.com/users/thomasw21/orgs", "received_events_url": "https://api.github.com/users/thomasw21/received_events", "repos_url": "https://api.github.com/users/thomasw21/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thomasw21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thomasw21/subscriptions", "type": "User", "url": "https://api.github.com/users/thomasw21" }
[]
null
null
CONTRIBUTOR
2023-01-18T17:50:04Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5440.diff", "html_url": "https://github.com/huggingface/datasets/pull/5440", "merged_at": "2023-01-18T17:50:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/5440.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5440" }
PR_kwDODunzps5HpRbF
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5440/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5440
https://github.com/huggingface/datasets/pull/5440
true
1,537,973,564
https://api.github.com/repos/huggingface/datasets/issues/5439/labels{/name}
### Feature request Please add the common voice 12_0 datasets. Apart from English, a significant amount of audio-data has been added to the other minor-language datasets. ### Motivation The dataset link: https://commonvoice.mozilla.org/en/datasets
2023-07-21T14:26:10Z
5,439
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-01-18T13:07:05Z
https://api.github.com/repos/huggingface/datasets/issues/5439/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
https://api.github.com/repos/huggingface/datasets/issues/5439/timeline
[dataset request] Add Common Voice 12.0
https://api.github.com/repos/huggingface/datasets/issues/5439/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/31034499?v=4", "events_url": "https://api.github.com/users/MohammedRakib/events{/privacy}", "followers_url": "https://api.github.com/users/MohammedRakib/followers", "following_url": "https://api.github.com/users/MohammedRakib/following{/other_user}", "gists_url": "https://api.github.com/users/MohammedRakib/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/MohammedRakib", "id": 31034499, "login": "MohammedRakib", "node_id": "MDQ6VXNlcjMxMDM0NDk5", "organizations_url": "https://api.github.com/users/MohammedRakib/orgs", "received_events_url": "https://api.github.com/users/MohammedRakib/received_events", "repos_url": "https://api.github.com/users/MohammedRakib/repos", "site_admin": false, "starred_url": "https://api.github.com/users/MohammedRakib/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MohammedRakib/subscriptions", "type": "User", "url": "https://api.github.com/users/MohammedRakib" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_use...
null
completed
NONE
2023-07-21T14:26:09Z
null
I_kwDODunzps5bq508
[ "@polinaeterna any tentative date on when the Common Voice 12.0 dataset will be added ?", "This dataset is now hosted on the Hub here: https://huggingface.co/datasets/mozilla-foundation/common_voice_12_0" ]
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/5439/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5439
https://github.com/huggingface/datasets/issues/5439
false
1,537,489,730
https://api.github.com/repos/huggingface/datasets/issues/5438/labels{/name}
This PR updates the "checkout" GitHub Action to its latest version, as previous ones are deprecated: https://github.blog/changelog/2022-09-22-github-actions-all-actions-will-begin-running-on-node16-instead-of-node12/
2023-01-18T13:49:51Z
5,438
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-18T06:53:15Z
https://api.github.com/repos/huggingface/datasets/issues/5438/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5438/timeline
Update actions/checkout in CD Conda release
https://api.github.com/repos/huggingface/datasets/issues/5438/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-18T13:42:49Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5438.diff", "html_url": "https://github.com/huggingface/datasets/pull/5438", "merged_at": "2023-01-18T13:42:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/5438.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5438" }
PR_kwDODunzps5HmWA8
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5438/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5438
https://github.com/huggingface/datasets/pull/5438
true
1,536,837,144
https://api.github.com/repos/huggingface/datasets/issues/5437/labels{/name}
I try to create dataset which contains about 9000 png images 64x64 in size, and they are all 4-channel (RGBA). When trying to use load_dataset() then a dataset is created from only 2 images. What exactly interferes I can not understand.![Screenshot_20230117_212213.jpg](https://user-images.githubusercontent.com/41611046/212980147-9aa68e30-76e9-4b61-a937-c2fdabd56564.jpg)
2023-01-18T20:20:15Z
5,437
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-17T18:22:27Z
https://api.github.com/repos/huggingface/datasets/issues/5437/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5437/timeline
Can't load png dataset with 4 channel (RGBA)
https://api.github.com/repos/huggingface/datasets/issues/5437/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/41611046?v=4", "events_url": "https://api.github.com/users/WiNE-iNEFF/events{/privacy}", "followers_url": "https://api.github.com/users/WiNE-iNEFF/followers", "following_url": "https://api.github.com/users/WiNE-iNEFF/following{/other_user}", "gists_url": "https://api.github.com/users/WiNE-iNEFF/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/WiNE-iNEFF", "id": 41611046, "login": "WiNE-iNEFF", "node_id": "MDQ6VXNlcjQxNjExMDQ2", "organizations_url": "https://api.github.com/users/WiNE-iNEFF/orgs", "received_events_url": "https://api.github.com/users/WiNE-iNEFF/received_events", "repos_url": "https://api.github.com/users/WiNE-iNEFF/repos", "site_admin": false, "starred_url": "https://api.github.com/users/WiNE-iNEFF/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/WiNE-iNEFF/subscriptions", "type": "User", "url": "https://api.github.com/users/WiNE-iNEFF" }
[]
null
completed
NONE
2023-01-18T20:20:15Z
null
I_kwDODunzps5bmkYY
[ "Hi! Can you please share the directory structure of your image folder and the `load_dataset` call? We decode images with Pillow, and Pillow supports RGBA PNGs, so this shouldn't be a problem.\r\n\r\n", "> Hi! Can you please share the directory structure of your image folder and the `load_dataset` call? We decode...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5437/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5437
https://github.com/huggingface/datasets/issues/5437
false
1,536,633,173
https://api.github.com/repos/huggingface/datasets/issues/5436/labels{/name}
Closes #5433, reverts #5432, and also: * Uses [ghcr.io container images](https://cml.dev/doc/self-hosted-runners/#docker-images) for extra speed * Updates `actions/checkout` to `v3` (note that `v2` is [deprecated](https://github.blog/changelog/2022-09-22-github-actions-all-actions-will-begin-running-on-node16-instead-of-node12/)) * Follows the new naming convention for environment variables introduced with [iterative/cml#1272](https://github.com/iterative/cml/pull/1272)
2023-01-18T09:05:49Z
5,436
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-17T15:59:50Z
https://api.github.com/repos/huggingface/datasets/issues/5436/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5436/timeline
Revert container image pin in CI benchmarks
https://api.github.com/repos/huggingface/datasets/issues/5436/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/11387611?v=4", "events_url": "https://api.github.com/users/0x2b3bfa0/events{/privacy}", "followers_url": "https://api.github.com/users/0x2b3bfa0/followers", "following_url": "https://api.github.com/users/0x2b3bfa0/following{/other_user}", "gists_url": "https://api.github.com/users/0x2b3bfa0/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/0x2b3bfa0", "id": 11387611, "login": "0x2b3bfa0", "node_id": "MDQ6VXNlcjExMzg3NjEx", "organizations_url": "https://api.github.com/users/0x2b3bfa0/orgs", "received_events_url": "https://api.github.com/users/0x2b3bfa0/received_events", "repos_url": "https://api.github.com/users/0x2b3bfa0/repos", "site_admin": false, "starred_url": "https://api.github.com/users/0x2b3bfa0/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/0x2b3bfa0/subscriptions", "type": "User", "url": "https://api.github.com/users/0x2b3bfa0" }
[]
null
null
CONTRIBUTOR
2023-01-18T06:29:06Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5436.diff", "html_url": "https://github.com/huggingface/datasets/pull/5436", "merged_at": "2023-01-18T06:29:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/5436.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5436" }
PR_kwDODunzps5Hjh4v
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 3, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 3, "url": "https://api.github.com/repos/huggingface/datasets/issues/5436/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5436
https://github.com/huggingface/datasets/pull/5436
true
1,536,099,300
https://api.github.com/repos/huggingface/datasets/issues/5435/labels{/name}
### Describe the bug In the [Split your dataset with take and skip](https://huggingface.co/docs/datasets/v1.10.2/dataset_streaming.html#split-your-dataset-with-take-and-skip), it states: > Using take (or skip) prevents future calls to shuffle from shuffling the dataset shards order, otherwise the taken examples could come from other shards. In this case it only uses the shuffle buffer. Therefore it is advised to shuffle the dataset before splitting using take or skip. See more details in the [Shuffling the dataset: shuffle](https://huggingface.co/docs/datasets/v1.10.2/dataset_streaming.html#iterable-dataset-shuffling) section.` >> \# You can also create splits from a shuffled dataset >> train_dataset = shuffled_dataset.skip(1000) >> eval_dataset = shuffled_dataset.take(1000) Where the shuffled dataset comes from: `shuffled_dataset = dataset.shuffle(buffer_size=10_000, seed=42)` At least in Tensorflow 2.9/2.10/2.11, [docs](https://www.tensorflow.org/api_docs/python/tf/data/Dataset#shuffle) states the `reshuffle_each_iteration` argument is `True` by default. This means the dataset would be shuffled after each epoch, and as a result **the validation data would leak into training test**. ### Steps to reproduce the bug N/A ### Expected behavior The `reshuffle_each_iteration` argument should be set to `False`. ### Environment info Tensorflow 2.9/2.10/2.11
2023-01-19T09:56:03Z
5,435
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-17T10:04:16Z
https://api.github.com/repos/huggingface/datasets/issues/5435/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5435/timeline
Wrong statement in "Load a Dataset in Streaming mode" leads to data leakage
https://api.github.com/repos/huggingface/datasets/issues/5435/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/80093591?v=4", "events_url": "https://api.github.com/users/DanielYang59/events{/privacy}", "followers_url": "https://api.github.com/users/DanielYang59/followers", "following_url": "https://api.github.com/users/DanielYang59/following{/other_user}", "gists_url": "https://api.github.com/users/DanielYang59/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/DanielYang59", "id": 80093591, "login": "DanielYang59", "node_id": "MDQ6VXNlcjgwMDkzNTkx", "organizations_url": "https://api.github.com/users/DanielYang59/orgs", "received_events_url": "https://api.github.com/users/DanielYang59/received_events", "repos_url": "https://api.github.com/users/DanielYang59/repos", "site_admin": false, "starred_url": "https://api.github.com/users/DanielYang59/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DanielYang59/subscriptions", "type": "User", "url": "https://api.github.com/users/DanielYang59" }
[]
null
completed
NONE
2023-01-19T09:56:03Z
null
I_kwDODunzps5bjwPk
[ "Just for your information, Tensorflow confirmed this issue [here.](https://github.com/tensorflow/tensorflow/issues/59279)", "Thanks for reporting, @HaoyuYang59.\r\n\r\nPlease note that these are different \"dataset\" objects: our docs refer to Hugging Face `datasets.Dataset` and not to TensorFlow `tf.data.Datase...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5435/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5435
https://github.com/huggingface/datasets/issues/5435
false
1,536,090,042
https://api.github.com/repos/huggingface/datasets/issues/5434/labels{/name}
null
2023-01-19T13:52:12Z
5,434
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-17T09:57:54Z
https://api.github.com/repos/huggingface/datasets/issues/5434/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5434/timeline
sample_dataset module not found
https://api.github.com/repos/huggingface/datasets/issues/5434/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/15816213?v=4", "events_url": "https://api.github.com/users/nickums/events{/privacy}", "followers_url": "https://api.github.com/users/nickums/followers", "following_url": "https://api.github.com/users/nickums/following{/other_user}", "gists_url": "https://api.github.com/users/nickums/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nickums", "id": 15816213, "login": "nickums", "node_id": "MDQ6VXNlcjE1ODE2MjEz", "organizations_url": "https://api.github.com/users/nickums/orgs", "received_events_url": "https://api.github.com/users/nickums/received_events", "repos_url": "https://api.github.com/users/nickums/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nickums/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nickums/subscriptions", "type": "User", "url": "https://api.github.com/users/nickums" }
[]
null
completed
NONE
2023-01-19T07:55:11Z
null
I_kwDODunzps5bjt-6
[ "Hi! Can you describe what the actual error is?", "working on the setfit example script\r\n\r\n from setfit import SetFitModel, SetFitTrainer, sample_dataset\r\n\r\nImportError: cannot import name 'sample_dataset' from 'setfit' (C:\\Python\\Python38\\lib\\site-packages\\setfit\\__init__.py)\r\n\r\n apart from t...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5434/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5434
https://github.com/huggingface/datasets/issues/5434
false
1,536,017,901
https://api.github.com/repos/huggingface/datasets/issues/5433/labels{/name}
Once we find out the root cause of: - #5431 we should revert the temporary pin on the Docker image version introduced by: - #5432
2023-01-18T06:29:08Z
5,433
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-01-17T09:06:08Z
https://api.github.com/repos/huggingface/datasets/issues/5433/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5433/timeline
Support latest Docker image in CI benchmarks
https://api.github.com/repos/huggingface/datasets/issues/5433/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
completed
MEMBER
2023-01-18T06:29:08Z
null
I_kwDODunzps5bjcXt
[ "Sorry, it was us:[^1] https://github.com/iterative/cml/pull/1317 & https://github.com/iterative/cml/issues/1319#issuecomment-1385599559; should be fixed with [v0.18.17](https://github.com/iterative/cml/releases/tag/v0.18.17).\r\n\r\n[^1]: More or less, see https://github.com/yargs/yargs/issues/873.", "Opened htt...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/5433/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5433
https://github.com/huggingface/datasets/issues/5433
false
1,535,893,019
https://api.github.com/repos/huggingface/datasets/issues/5432/labels{/name}
This PR fixes CI benchmarks, by temporarily pinning Docker image version, instead of "latest" tag. It also updates deprecated `cml-send-comment` command and using `cml comment create` instead. Fix #5431.
2023-01-17T08:58:22Z
5,432
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-17T07:15:31Z
https://api.github.com/repos/huggingface/datasets/issues/5432/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5432/timeline
Fix CI benchmarks by temporarily pinning Docker image version
https://api.github.com/repos/huggingface/datasets/issues/5432/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-17T08:51:17Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5432.diff", "html_url": "https://github.com/huggingface/datasets/pull/5432", "merged_at": "2023-01-17T08:51:17Z", "patch_url": "https://github.com/huggingface/datasets/pull/5432.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5432" }
PR_kwDODunzps5HhEA8
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5432/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5432
https://github.com/huggingface/datasets/pull/5432
true
1,535,862,621
https://api.github.com/repos/huggingface/datasets/issues/5431/labels{/name}
Our CI benchmarks are broken, raising `Unknown arguments` error: https://github.com/huggingface/datasets/actions/runs/3932397079/jobs/6724905161 ``` Unknown arguments: runnerPath, path ``` Stack trace: ``` 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 500/500 [00:01<00:00, 338.98ba/s] Updating lock file 'dvc.lock' To track the changes with git, run: git add dvc.lock To enable auto staging, run: dvc config core.autostage true Use `dvc push` to send your updates to remote storage. cml send-comment <markdown file> Global Options: --log Logging verbosity [string] [choices: "error", "warn", "info", "debug"] [default: "info"] --driver Git provider where the repository is hosted [string] [choices: "github", "gitlab", "bitbucket"] [default: infer from the environment] --repo Repository URL or slug [string] [default: infer from the environment] --driver-token, --token CI driver personal/project access token (PAT) [string] [default: infer from the environment] --help Show help [boolean] Options: --target Comment type (`commit`, `pr`, `commit/f00bar`, `pr/42`, `issue/1337`),default is automatic (`pr` but fallback to `commit`). [string] --watch Watch for changes and automatically update the comment [boolean] --publish Upload any local images found in the Markdown report [boolean] [default: true] --publish-url Self-hosted image server URL [string] [default: "https://asset.cml.dev/"] --publish-native, --native Uses driver's native capabilities to upload assets instead of CML's storage; not available on GitHub [boolean] --watermark-title Hidden comment marker (used for targeting in subsequent `cml comment update`); "{workflow}" & "{run}" are auto-replaced [string] [default: ""] Unknown arguments: runnerPath, path Error: Process completed with exit code 1. ``` Issue reported to iterative/cml: - iterative/cml#1319
2023-01-18T06:33:24Z
5,431
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d4c5f9", "default": false, "description": "Maintenance tasks", "id": 4296013012, "name": "maintenance", "node_id": "LA_kwDODunzps8AAAABAA_01A", "url": "https://api.github.com/repos/huggingface/datasets/labels/maintenance" } ]
2023-01-17T06:49:57Z
https://api.github.com/repos/huggingface/datasets/issues/5431/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5431/timeline
CI benchmarks are broken: Unknown arguments: runnerPath, path
https://api.github.com/repos/huggingface/datasets/issues/5431/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
MEMBER
2023-01-17T08:51:18Z
null
I_kwDODunzps5bi2dd
[]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5431/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5431
https://github.com/huggingface/datasets/issues/5431
false
1,535,856,503
https://api.github.com/repos/huggingface/datasets/issues/5430/labels{/name}
Once we find out the root cause of: - #5426 we should revert the temporary pin on apache-beam introduced by: - #5429
2024-02-06T19:24:21Z
5,430
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-01-17T06:42:12Z
https://api.github.com/repos/huggingface/datasets/issues/5430/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5430/timeline
Support Apache Beam >= 2.44.0
https://api.github.com/repos/huggingface/datasets/issues/5430/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
MEMBER
2024-02-06T19:24:21Z
null
I_kwDODunzps5bi093
[ "Some of the shard files now have 0 number of rows.\r\n\r\nWe have opened an issue in the Apache Beam repo:\r\n- https://github.com/apache/beam/issues/25041" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5430/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5430
https://github.com/huggingface/datasets/issues/5430
false
1,535,192,687
https://api.github.com/repos/huggingface/datasets/issues/5429/labels{/name}
Temporarily pin apache-beam < 2.44.0 Fix #5426.
2023-01-16T16:51:42Z
5,429
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-16T16:20:09Z
https://api.github.com/repos/huggingface/datasets/issues/5429/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5429/timeline
Fix CI by temporarily pinning apache-beam < 2.44.0
https://api.github.com/repos/huggingface/datasets/issues/5429/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-16T16:49:03Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5429.diff", "html_url": "https://github.com/huggingface/datasets/pull/5429", "merged_at": "2023-01-16T16:49:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/5429.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5429" }
PR_kwDODunzps5HeuyT
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5429/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5429
https://github.com/huggingface/datasets/pull/5429
true
1,535,166,139
https://api.github.com/repos/huggingface/datasets/issues/5428/labels{/name}
### Feature request From what I understand `faiss` already support this [link](https://github.com/facebookresearch/faiss/wiki/Index-IO,-cloning-and-hyper-parameter-tuning#generic-io-support) I would like to use a stream as input to `Dataset.load_faiss_index` and `Dataset.save_faiss_index`. ### Motivation In my case, I'm saving faiss index in cloud storage and use `fsspec` to load them. It would be ideal if I could send the stream directly instead of copying the file locally (or mounting the bucket) and then load the index. ### Your contribution I can submit the PR
2023-03-27T15:18:22Z
5,428
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-01-16T16:08:12Z
https://api.github.com/repos/huggingface/datasets/issues/5428/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5428/timeline
Load/Save FAISS index using fsspec
https://api.github.com/repos/huggingface/datasets/issues/5428/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8976546?v=4", "events_url": "https://api.github.com/users/Dref360/events{/privacy}", "followers_url": "https://api.github.com/users/Dref360/followers", "following_url": "https://api.github.com/users/Dref360/following{/other_user}", "gists_url": "https://api.github.com/users/Dref360/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Dref360", "id": 8976546, "login": "Dref360", "node_id": "MDQ6VXNlcjg5NzY1NDY=", "organizations_url": "https://api.github.com/users/Dref360/orgs", "received_events_url": "https://api.github.com/users/Dref360/received_events", "repos_url": "https://api.github.com/users/Dref360/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Dref360/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Dref360/subscriptions", "type": "User", "url": "https://api.github.com/users/Dref360" }
[]
null
completed
CONTRIBUTOR
2023-03-27T15:18:22Z
null
I_kwDODunzps5bgMa7
[ "Hi! Sure, feel free to submit a PR. Maybe if we want to be consistent with the existing API, it would be cleaner to directly add support for `fsspec` paths in `Dataset.load_faiss_index`/`Dataset.save_faiss_index` in the same manner as it was done in `Dataset.load_from_disk`/`Dataset.save_to_disk`.", "That's a gr...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5428/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5428
https://github.com/huggingface/datasets/issues/5428
false
1,535,162,889
https://api.github.com/repos/huggingface/datasets/issues/5427/labels{/name}
### Describe the bug I tried to download dataset `id_clickbait`, but receive this error message. ``` FileNotFoundError: Couldn't find file at https://md-datasets-cache-zipfiles-prod.s3.eu-west-1.amazonaws.com/k42j7x2kpn-1.zip ``` When i open the link using browser, i got this XML data. ```xml <?xml version="1.0" encoding="UTF-8"?> <Error><Code>NoSuchBucket</Code><Message>The specified bucket does not exist</Message><BucketName>md-datasets-cache-zipfiles-prod</BucketName><RequestId>NVRM6VEEQD69SD00</RequestId><HostId>W/SPDxLGvlCGi0OD6d7mSDvfOAUqLAfvs9nTX50BkJrjMny+X9Jnqp/Li2lG9eTUuT4MUkAA2jjTfCrCiUmu7A==</HostId></Error> ``` ### Steps to reproduce the bug Code snippet: ``` from datasets import load_dataset load_dataset('id_clickbait', 'annotated') load_dataset('id_clickbait', 'raw') ``` Link to Kaggle notebook: https://www.kaggle.com/code/ilosvigil/bug-check-on-id-clickbait-dataset ### Expected behavior Successfully download and load `id_newspaper` dataset. ### Environment info - `datasets` version: 2.8.0 - Platform: Linux-5.15.65+-x86_64-with-debian-bullseye-sid - Python version: 3.7.12 - PyArrow version: 8.0.0 - Pandas version: 1.3.5
2023-01-18T09:51:28Z
5,427
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-16T16:05:36Z
https://api.github.com/repos/huggingface/datasets/issues/5427/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5427/timeline
Unable to download dataset id_clickbait
https://api.github.com/repos/huggingface/datasets/issues/5427/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/45941585?v=4", "events_url": "https://api.github.com/users/ilos-vigil/events{/privacy}", "followers_url": "https://api.github.com/users/ilos-vigil/followers", "following_url": "https://api.github.com/users/ilos-vigil/following{/other_user}", "gists_url": "https://api.github.com/users/ilos-vigil/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ilos-vigil", "id": 45941585, "login": "ilos-vigil", "node_id": "MDQ6VXNlcjQ1OTQxNTg1", "organizations_url": "https://api.github.com/users/ilos-vigil/orgs", "received_events_url": "https://api.github.com/users/ilos-vigil/received_events", "repos_url": "https://api.github.com/users/ilos-vigil/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ilos-vigil/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ilos-vigil/subscriptions", "type": "User", "url": "https://api.github.com/users/ilos-vigil" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
NONE
2023-01-18T09:25:19Z
null
I_kwDODunzps5bgLoJ
[ "Thanks for reporting, @ilos-vigil.\r\n\r\nWe have transferred this issue to the corresponding dataset on the Hugging Face Hub: https://huggingface.co/datasets/id_clickbait/discussions/1 " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5427/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5427
https://github.com/huggingface/datasets/issues/5427
false
1,535,158,555
https://api.github.com/repos/huggingface/datasets/issues/5426/labels{/name}
CI test (unit, ubuntu-latest, deps-minimum) is broken, raising a `SchemaInferenceError`: see https://github.com/huggingface/datasets/actions/runs/3930901593/jobs/6721492004 ``` FAILED tests/test_beam.py::BeamBuilderTest::test_download_and_prepare_sharded - datasets.arrow_writer.SchemaInferenceError: Please pass `features` or at least one example when writing data ``` Stack trace: ``` ______________ BeamBuilderTest.test_download_and_prepare_sharded _______________ [gw1] linux -- Python 3.7.15 /opt/hostedtoolcache/Python/3.7.15/x64/bin/python self = <tests.test_beam.BeamBuilderTest testMethod=test_download_and_prepare_sharded> @require_beam def test_download_and_prepare_sharded(self): import apache_beam as beam original_write_parquet = beam.io.parquetio.WriteToParquet expected_num_examples = len(get_test_dummy_examples()) with tempfile.TemporaryDirectory() as tmp_cache_dir: builder = DummyBeamDataset(cache_dir=tmp_cache_dir, beam_runner="DirectRunner") with patch("apache_beam.io.parquetio.WriteToParquet") as write_parquet_mock: write_parquet_mock.side_effect = partial(original_write_parquet, num_shards=2) > builder.download_and_prepare() tests/test_beam.py:97: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ /opt/hostedtoolcache/Python/3.7.15/x64/lib/python3.7/site-packages/datasets/builder.py:864: in download_and_prepare **download_and_prepare_kwargs, /opt/hostedtoolcache/Python/3.7.15/x64/lib/python3.7/site-packages/datasets/builder.py:1976: in _download_and_prepare num_examples, num_bytes = beam_writer.finalize(metrics.query(m_filter)) /opt/hostedtoolcache/Python/3.7.15/x64/lib/python3.7/site-packages/datasets/arrow_writer.py:694: in finalize shard_num_bytes, _ = parquet_to_arrow(source, destination) /opt/hostedtoolcache/Python/3.7.15/x64/lib/python3.7/site-packages/datasets/arrow_writer.py:740: in parquet_to_arrow num_bytes, num_examples = writer.finalize() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <datasets.arrow_writer.ArrowWriter object at 0x7f6dcbb3e810> close_stream = True def finalize(self, close_stream=True): self.write_rows_on_file() # In case current_examples < writer_batch_size, but user uses finalize() if self._check_duplicates: self.check_duplicate_keys() # Re-intializing to empty list for next batch self.hkey_record = [] self.write_examples_on_file() # If schema is known, infer features even if no examples were written if self.pa_writer is None and self.schema: self._build_writer(self.schema) if self.pa_writer is not None: self.pa_writer.close() self.pa_writer = None if close_stream: self.stream.close() else: if close_stream: self.stream.close() > raise SchemaInferenceError("Please pass `features` or at least one example when writing data") E datasets.arrow_writer.SchemaInferenceError: Please pass `features` or at least one example when writing data /opt/hostedtoolcache/Python/3.7.15/x64/lib/python3.7/site-packages/datasets/arrow_writer.py:593: SchemaInferenceError ```
2023-06-02T06:40:32Z
5,426
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2023-01-16T16:02:07Z
https://api.github.com/repos/huggingface/datasets/issues/5426/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5426/timeline
CI tests are broken: SchemaInferenceError
https://api.github.com/repos/huggingface/datasets/issues/5426/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
MEMBER
2023-01-16T16:49:04Z
null
I_kwDODunzps5bgKkb
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5426/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5426
https://github.com/huggingface/datasets/issues/5426
false
1,534,581,850
https://api.github.com/repos/huggingface/datasets/issues/5425/labels{/name}
### Feature request From discussion on forum: https://discuss.huggingface.co/t/datasets-dataset-sort-does-not-preserve-ordering/29065/1 `sort()` does not preserve ordering, and it does not support sorting on multiple columns, nor a key function. The suggested solution: > ... having something similar to pandas and be able to specify multiple columns for sorting. We’re already using pandas under the hood to do the sorting in datasets. The suggested workaround: > convert your dataset to pandas and use `df.sort_values()` ### Motivation Preserved ordering when sorting is very handy when one needs to sort on multiple columns, A and B, so that e.g. whenever A is equal for two or more rows, B is kept sorted. Having a parameter to do this in πŸ€—datasets would be cleaner than going through pandas and back, and it wouldn't add much complexity to the library. Alternatives: - the possibility to specify multiple keys to sort by with decreasing priority (suggested solution), - the ability to provide a key function for sorting, so that one can manually specify the sorting criteria. ### Your contribution I'll be happy to contribute by submitting a PR. Will get documented on `CONTRIBUTING.MD`. Would love to get thoughts on this, if anyone has anything to add.
2023-02-24T16:15:11Z
5,425
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "7057ff", "default": true...
2023-01-16T09:22:26Z
https://api.github.com/repos/huggingface/datasets/issues/5425/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5425/timeline
Sort on multiple keys with datasets.Dataset.sort()
https://api.github.com/repos/huggingface/datasets/issues/5425/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/101344863?v=4", "events_url": "https://api.github.com/users/rocco-fortuna/events{/privacy}", "followers_url": "https://api.github.com/users/rocco-fortuna/followers", "following_url": "https://api.github.com/users/rocco-fortuna/following{/other_user}", "gists_url": "https://api.github.com/users/rocco-fortuna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/rocco-fortuna", "id": 101344863, "login": "rocco-fortuna", "node_id": "U_kgDOBgpmXw", "organizations_url": "https://api.github.com/users/rocco-fortuna/orgs", "received_events_url": "https://api.github.com/users/rocco-fortuna/received_events", "repos_url": "https://api.github.com/users/rocco-fortuna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/rocco-fortuna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rocco-fortuna/subscriptions", "type": "User", "url": "https://api.github.com/users/rocco-fortuna" }
[]
null
completed
NONE
2023-02-24T16:15:11Z
null
I_kwDODunzps5bd9xa
[ "Hi! \r\n\r\n`Dataset.sort` calls `df.sort_values` internally, and `df.sort_values` brings all the \"sort\" columns in memory, so sorting on multiple keys could be very expensive. This makes me think that maybe we can replace `df.sort_values` with `pyarrow.compute.sort_indices` - the latter can also sort on multipl...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5425/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5425
https://github.com/huggingface/datasets/issues/5425
false
1,534,394,756
https://api.github.com/repos/huggingface/datasets/issues/5424/labels{/name}
### Describe the bug I am loading datasets from custom `tsv` files stored locally and applying split instructions for each split. Although the ReadInstruction is being applied correctly and I was expecting it to be `DatasetDict` but instead it is a list of `Dataset`. ### Steps to reproduce the bug Steps to reproduce the behaviour: 1. Import `from datasets import load_dataset, ReadInstruction` 2. Instruction to load the dataset ``` instructions = [ ReadInstruction(split_name="train", from_=0, to=10, unit='%', rounding='closest'), ReadInstruction(split_name="dev", from_=0, to=10, unit='%', rounding='closest'), ReadInstruction(split_name="test", from_=0, to=5, unit='%', rounding='closest') ] ``` 3. Load `dataset = load_dataset('csv', data_dir="data/", data_files={"train":"train.tsv", "dev":"dev.tsv", "test":"test.tsv"}, delimiter="\t", split=instructions)` ### Expected behavior **Current behaviour** ![Screenshot from 2023-01-16 10-45-27](https://user-images.githubusercontent.com/25720695/212614754-306898d8-8c27-4475-9bb8-0321bd939561.png) : **Expected behaviour** ![Screenshot from 2023-01-16 10-45-42](https://user-images.githubusercontent.com/25720695/212614813-0d336bf7-5266-482e-bb96-ef51f64de204.png) ### Environment info ``datasets==2.8.0 `` `Python==3.8.5 ` `Platform - Ubuntu 20.04.4 LTS`
2023-02-24T16:19:00Z
5,424
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-16T06:54:28Z
https://api.github.com/repos/huggingface/datasets/issues/5424/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5424/timeline
When applying `ReadInstruction` to custom load it's not DatasetDict but list of Dataset?
https://api.github.com/repos/huggingface/datasets/issues/5424/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/25720695?v=4", "events_url": "https://api.github.com/users/macabdul9/events{/privacy}", "followers_url": "https://api.github.com/users/macabdul9/followers", "following_url": "https://api.github.com/users/macabdul9/following{/other_user}", "gists_url": "https://api.github.com/users/macabdul9/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/macabdul9", "id": 25720695, "login": "macabdul9", "node_id": "MDQ6VXNlcjI1NzIwNjk1", "organizations_url": "https://api.github.com/users/macabdul9/orgs", "received_events_url": "https://api.github.com/users/macabdul9/received_events", "repos_url": "https://api.github.com/users/macabdul9/repos", "site_admin": false, "starred_url": "https://api.github.com/users/macabdul9/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/macabdul9/subscriptions", "type": "User", "url": "https://api.github.com/users/macabdul9" }
[]
null
completed
NONE
2023-02-24T16:19:00Z
null
I_kwDODunzps5bdQGE
[ "Hi! You can get a `DatasetDict` if you pass a dictionary with read instructions as follows:\r\n```python\r\ninstructions = [\r\n ReadInstruction(split_name=\"train\", from_=0, to=10, unit='%', rounding='closest'),\r\n ReadInstruction(split_name=\"dev\", from_=0, to=10, unit='%', rounding='closest'),\r\n R...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5424/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5424
https://github.com/huggingface/datasets/issues/5424
false
1,533,385,239
https://api.github.com/repos/huggingface/datasets/issues/5422/labels{/name}
### Describe the bug Loading a previously downloaded & saved dataset as described in the HuggingFace course: issues_dataset = load_dataset("json", data_files="issues/datasets-issues.jsonl", split="train") Gives this error: datasets.builder.DatasetGenerationError: An error occurred while generating the dataset A work-around I found was to use streaming. ### Steps to reproduce the bug Reproduce by executing the code provided: https://huggingface.co/course/chapter5/5?fw=pt From the heading: 'let’s create a function that can download all the issues from a GitHub repository' ### Expected behavior No error ### Environment info Datasets version 2.8.0. Note that version 2.6.1 gives the same error (related to null timestamp). **[EDIT]** This is the complete error trace confirming the issue is related to the timestamp (`Couldn't cast array of type timestamp[s] to null`) ``` Using custom data configuration default-950028611d2860c8 Downloading and preparing dataset json/default to [...]/.cache/huggingface/datasets/json/default-950028611d2860c8/0.0.0/0f7e3662623656454fcd2b650f34e886a7db4b9104504885bd462096cc7a9f51... Downloading data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:00<?, ?it/s] Extracting data files: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:00<00:00, 500.63it/s] Generating train split: 2619 examples [00:00, 7155.72 examples/s]Traceback (most recent call last): File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\builder.py", line 1831, in _prepare_split_single writer.write_table(table) File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\arrow_writer.py", line 567, in write_table pa_table = table_cast(pa_table, self._schema) File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 2282, in table_cast return cast_table_to_schema(table, schema) File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 2241, in cast_table_to_schema arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 2241, in <listcomp> arrays = [cast_array_to_feature(table[name], feature) for name, feature in features.items()] File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 1807, in wrapper return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 1807, in <listcomp> return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 2035, in cast_array_to_feature arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()] File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 2035, in <listcomp> arrays = [_c(array.field(name), subfeature) for name, subfeature in feature.items()] File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 1809, in wrapper return func(array, *args, **kwargs) File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 2101, in cast_array_to_feature return array_cast(array, feature(), allow_number_to_str=allow_number_to_str) File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 1809, in wrapper return func(array, *args, **kwargs) File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\table.py", line 1990, in array_cast raise TypeError(f"Couldn't cast array of type {array.type} to {pa_type}") TypeError: Couldn't cast array of type timestamp[s] to null The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\Program Files\JetBrains\PyCharm 2022.1.3\plugins\python\helpers\pydev\pydevconsole.py", line 364, in runcode coro = func() File "<input>", line 1, in <module> File "C:\Program Files\JetBrains\PyCharm 2022.1.3\plugins\python\helpers\pydev\_pydev_bundle\pydev_umd.py", line 198, in runfile pydev_imports.execfile(filename, global_vars, local_vars) # execute the script File "C:\Program Files\JetBrains\PyCharm 2022.1.3\plugins\python\helpers\pydev\_pydev_imps\_pydev_execfile.py", line 18, in execfile exec(compile(contents+"\n", file, 'exec'), glob, loc) File "[...]\PycharmProjects\TransformersTesting\dataset_issues.py", line 20, in <module> issues_dataset = load_dataset("json", data_files="issues/datasets-issues.jsonl", split="train") File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\load.py", line 1757, in load_dataset builder_instance.download_and_prepare( File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\builder.py", line 860, in download_and_prepare self._download_and_prepare( File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\builder.py", line 953, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\builder.py", line 1706, in _prepare_split for job_id, done, content in self._prepare_split_single( File "[...]\miniconda3\envs\HuggingFace\lib\site-packages\datasets\builder.py", line 1849, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.builder.DatasetGenerationError: An error occurred while generating the dataset Generating train split: 2619 examples [00:19, 7155.72 examples/s] ```
2023-09-14T11:39:57Z
5,422
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-14T17:29:38Z
https://api.github.com/repos/huggingface/datasets/issues/5422/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5422/timeline
Datasets load error for saved github issues
https://api.github.com/repos/huggingface/datasets/issues/5422/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7360564?v=4", "events_url": "https://api.github.com/users/folterj/events{/privacy}", "followers_url": "https://api.github.com/users/folterj/followers", "following_url": "https://api.github.com/users/folterj/following{/other_user}", "gists_url": "https://api.github.com/users/folterj/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/folterj", "id": 7360564, "login": "folterj", "node_id": "MDQ6VXNlcjczNjA1NjQ=", "organizations_url": "https://api.github.com/users/folterj/orgs", "received_events_url": "https://api.github.com/users/folterj/received_events", "repos_url": "https://api.github.com/users/folterj/repos", "site_admin": false, "starred_url": "https://api.github.com/users/folterj/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/folterj/subscriptions", "type": "User", "url": "https://api.github.com/users/folterj" }
[]
null
null
NONE
null
null
I_kwDODunzps5bZZoX
[ "I can confirm that the error exists!\r\nI'm trying to read 3 parquet files locally:\r\n```python\r\nfrom datasets import load_dataset, Features, Value, ClassLabel\r\n\r\nreview_dataset = load_dataset(\r\n \"parquet\",\r\n data_files={\r\n \"train\": os.path.join(sentiment_analysis_data_path, \"train.p...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5422/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5422
https://github.com/huggingface/datasets/issues/5422
false
1,532,278,307
https://api.github.com/repos/huggingface/datasets/issues/5421/labels{/name}
### Feature request The dataset name on the Hub is case-insensitive (see https://github.com/huggingface/moon-landing/pull/2399, internal issue), i.e., https://huggingface.co/datasets/GLUE redirects to https://huggingface.co/datasets/glue. Ideally, we could load the glue dataset using the following: ``` from datasets import load_dataset load_dataset('GLUE', 'cola') ``` It breaks because the loading script `GLUE.py` does not exist (`glue.py` should be selected instead). Minor additional comment: in other cases without a loading script, we can load the dataset, but the automatically generated config name depends on the casing: - `load_dataset('severo/danish-wit')` generates the config name `severo--danish-wit-e6fda5b070deb133`, while - `load_dataset('severo/danish-WIT')` generates the config name `severo--danish-WIT-e6fda5b070deb133` ### Motivation To follow the same UX on the Hub and in the datasets library. ### Your contribution ...
2023-01-13T20:12:32Z
5,421
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-01-13T13:07:07Z
https://api.github.com/repos/huggingface/datasets/issues/5421/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5421/timeline
Support case-insensitive Hub dataset name in load_dataset
https://api.github.com/repos/huggingface/datasets/issues/5421/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[]
null
completed
CONTRIBUTOR
2023-01-13T20:12:32Z
null
I_kwDODunzps5bVLYj
[ "Closing as case-insensitivity should be only for URL redirection on the Hub. In the APIs, we will only support the canonical name (https://github.com/huggingface/moon-landing/pull/2399#issuecomment-1382085611)" ]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5421/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5421
https://github.com/huggingface/datasets/issues/5421
false
1,532,265,742
https://api.github.com/repos/huggingface/datasets/issues/5420/labels{/name}
add-dataset is not needed anymore since the "canonical" datasets are on the Hub. And dataset-viewer is managed within the datasets-server project. See https://github.com/huggingface/datasets/issues/new/choose <img width="1245" alt="Capture d’écran 2023-01-13 aΜ€ 13 59 58" src="https://user-images.githubusercontent.com/1676121/212325813-2d4c30e2-343e-4aa2-8cce-b2b77f45628e.png">
2023-01-13T13:36:00Z
5,420
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-13T12:58:43Z
https://api.github.com/repos/huggingface/datasets/issues/5420/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5420/timeline
ci: 🎑 remove two obsolete issue templates
https://api.github.com/repos/huggingface/datasets/issues/5420/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[]
null
null
CONTRIBUTOR
2023-01-13T13:29:01Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5420.diff", "html_url": "https://github.com/huggingface/datasets/pull/5420", "merged_at": "2023-01-13T13:29:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/5420.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5420" }
PR_kwDODunzps5HVAhL
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5420/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5420
https://github.com/huggingface/datasets/pull/5420
true
1,531,999,850
https://api.github.com/repos/huggingface/datasets/issues/5419/labels{/name}
### Describe the bug When preparing a dataset for a task using `datasets.TextClassification`, the output feature is named `labels`. When preparing the trainer using the `transformers.DataCollator` the default column name is `label` if binary or `label_ids` if multi-class problem. It is required to rename the column accordingly to the expected name : `label` or `label_ids` ### Steps to reproduce the bug ```python from datasets import TextClassification, AutoTokenized, DataCollatorWithPadding ds_prepared = my_dataset.prepare_for_task(TextClassification(text_column='TEXT', label_column='MY_LABEL_COLUMN_1_OR_0')) print(ds_prepared) tokenizer = AutoTokenizer.from_pretrained("distilbert-base-uncased") ds_tokenized = ds_prepared.map(lambda x: tokenizer(x['text'], truncation=True), batched=True) print(ds_tokenized) data_collator = DataCollatorWithPadding(tokenizer=tokenizer, return_tensors="tf") tf_data = model.prepare_tf_dataset(ds_tokenized, shuffle=True, batch_size=16, collate_fn=data_collator) print(tf_data) ``` ### Expected behavior Without renaming the the column, the target column is not in the final tf_data since it is not in the column name expected by the data_collator. To correct this, we have to rename the column: ```python ds_prepared = my_dataset.prepare_for_task(TextClassification(text_column='TEXT', label_column='MY_LABEL_COLUMN_1_OR_0')).rename_column('labels', 'label') ``` ### Environment info - `datasets` version: 2.8.0 - Platform: Linux-5.15.79.1-microsoft-standard-WSL2-x86_64-with-glibc2.35 - Python version: 3.10.6 - PyArrow version: 10.0.1 - Pandas version: 1.5.2 - `transformers` version: 4.26.0.dev0 - Platform: Linux-5.15.79.1-microsoft-standard-WSL2-x86_64-with-glibc2.35 - Python version: 3.10.6 - Huggingface_hub version: 0.11.1 - PyTorch version (GPU?): not installed (NA) - Tensorflow version (GPU?): 2.11.0 (True) - Flax version (CPU?/GPU?/TPU?): not installed (NA) - Jax version: not installed - JaxLib version: not installed - Using GPU in script?: <fill in> - Using distributed or parallel set-up in script?: <fill in>
2023-07-21T14:27:08Z
5,419
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-13T09:40:07Z
https://api.github.com/repos/huggingface/datasets/issues/5419/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5419/timeline
label_column='labels' in datasets.TextClassification and 'label' or 'label_ids' in transformers.DataColator
https://api.github.com/repos/huggingface/datasets/issues/5419/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/172385?v=4", "events_url": "https://api.github.com/users/CreatixEA/events{/privacy}", "followers_url": "https://api.github.com/users/CreatixEA/followers", "following_url": "https://api.github.com/users/CreatixEA/following{/other_user}", "gists_url": "https://api.github.com/users/CreatixEA/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CreatixEA", "id": 172385, "login": "CreatixEA", "node_id": "MDQ6VXNlcjE3MjM4NQ==", "organizations_url": "https://api.github.com/users/CreatixEA/orgs", "received_events_url": "https://api.github.com/users/CreatixEA/received_events", "repos_url": "https://api.github.com/users/CreatixEA/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CreatixEA/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CreatixEA/subscriptions", "type": "User", "url": "https://api.github.com/users/CreatixEA" }
[]
null
completed
NONE
2023-07-21T14:27:08Z
null
I_kwDODunzps5bUHZq
[ "Hi! Thanks for pointing out this inconsistency. Changing the default value at this point is probably not worth it, considering we've started discussing the state of the task API internally - we will most likely deprecate the current one and replace it with a more robust solution that relies on the `train_eval_inde...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5419/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5419
https://github.com/huggingface/datasets/issues/5419
false
1,530,111,184
https://api.github.com/repos/huggingface/datasets/issues/5418/labels{/name}
### Feature request Add a progress bar for `Dataset.to_parquet`, similar to how `to_json` works. ### Motivation It's a bit frustrating to not know how long a dataset will take to write to file and if it's stuck or not without a progress bar ### Your contribution Sure I can help if needed
2023-01-24T18:18:24Z
5,418
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-01-12T05:06:20Z
https://api.github.com/repos/huggingface/datasets/issues/5418/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/33707069?v=4", "events_url": "https://api.github.com/users/zanussbaum/events{/privacy}", "followers_url": "https://api.github.com/users/zanussbaum/followers", "following_url": "https://api.github.com/users/zanussbaum/following{/other_user}", "gists_url": "https://api.github.com/users/zanussbaum/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zanussbaum", "id": 33707069, "login": "zanussbaum", "node_id": "MDQ6VXNlcjMzNzA3MDY5", "organizations_url": "https://api.github.com/users/zanussbaum/orgs", "received_events_url": "https://api.github.com/users/zanussbaum/received_events", "repos_url": "https://api.github.com/users/zanussbaum/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zanussbaum/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zanussbaum/subscriptions", "type": "User", "url": "https://api.github.com/users/zanussbaum" }
https://api.github.com/repos/huggingface/datasets/issues/5418/timeline
Add ProgressBar for `to_parquet`
https://api.github.com/repos/huggingface/datasets/issues/5418/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/33707069?v=4", "events_url": "https://api.github.com/users/zanussbaum/events{/privacy}", "followers_url": "https://api.github.com/users/zanussbaum/followers", "following_url": "https://api.github.com/users/zanussbaum/following{/other_user}", "gists_url": "https://api.github.com/users/zanussbaum/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/zanussbaum", "id": 33707069, "login": "zanussbaum", "node_id": "MDQ6VXNlcjMzNzA3MDY5", "organizations_url": "https://api.github.com/users/zanussbaum/orgs", "received_events_url": "https://api.github.com/users/zanussbaum/received_events", "repos_url": "https://api.github.com/users/zanussbaum/repos", "site_admin": false, "starred_url": "https://api.github.com/users/zanussbaum/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zanussbaum/subscriptions", "type": "User", "url": "https://api.github.com/users/zanussbaum" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/33707069?v=4", "events_url": "https://api.github.com/users/zanussbaum/events{/privacy}", "followers_url": "https://api.github.com/users/zanussbaum/followers", "following_url": "https://api.github.com/users/zanussbaum/following{/other_user}", ...
null
completed
CONTRIBUTOR
2023-01-24T18:18:24Z
null
I_kwDODunzps5bM6TQ
[ "Thanks for your proposal, @zanussbaum. Yes, I agree that would definitely be a nice feature to have!", "@albertvillanova I’m happy to make a quick PR for the feature! let me know ", "That would be awesome ! You can comment `#self-assign` to assign you to this issue and open a PR :) Will be happy to review", ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5418/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5418
https://github.com/huggingface/datasets/issues/5418
false
1,526,988,113
https://api.github.com/repos/huggingface/datasets/issues/5416/labels{/name}
This PR fixes the RuntimeError: Sharding is ambiguous for this dataset. The error for ambiguous sharding will be raised only if num_proc > 1. Fix #5415, fix #5414. Fix https://huggingface.co/datasets/ami/discussions/3.
2023-01-18T17:12:17Z
5,416
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-10T08:43:19Z
https://api.github.com/repos/huggingface/datasets/issues/5416/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5416/timeline
Fix RuntimeError: Sharding is ambiguous for this dataset
https://api.github.com/repos/huggingface/datasets/issues/5416/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-18T14:09:02Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5416.diff", "html_url": "https://github.com/huggingface/datasets/pull/5416", "merged_at": "2023-01-18T14:09:02Z", "patch_url": "https://github.com/huggingface/datasets/pull/5416.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5416" }
PR_kwDODunzps5HDLmR
[ "_The documentation is not available anymore as the PR was closed or merged._", "By the way, do we know how many datasets are impacted by this issue?\r\n\r\nMaybe we should do a patch release with this fix.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated be...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5416/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5416
https://github.com/huggingface/datasets/pull/5416
true
1,526,904,861
https://api.github.com/repos/huggingface/datasets/issues/5415/labels{/name}
### Describe the bug When loading some datasets, a RuntimeError is raised. For example, for "ami" dataset: https://huggingface.co/datasets/ami/discussions/3 ``` .../huggingface/datasets/src/datasets/builder.py in _prepare_split(self, split_generator, check_duplicate_keys, file_format, num_proc, max_shard_size) 1415 fpath = path_join(self._output_dir, fname) 1416 -> 1417 num_input_shards = _number_of_shards_in_gen_kwargs(split_generator.gen_kwargs) 1418 if num_input_shards <= 1 and num_proc is not None: 1419 logger.warning( .../huggingface/datasets/src/datasets/utils/sharding.py in _number_of_shards_in_gen_kwargs(gen_kwargs) 10 lists_lengths = {key: len(value) for key, value in gen_kwargs.items() if isinstance(value, list)} 11 if len(set(lists_lengths.values())) > 1: ---> 12 raise RuntimeError( 13 ( 14 "Sharding is ambiguous for this dataset: " RuntimeError: Sharding is ambiguous for this dataset: we found several data sources lists of different lengths, and we don't know over which list we should parallelize: - key samples_paths has length 6 - key ids has length 7 - key verification_ids has length 6 To fix this, check the 'gen_kwargs' and make sure to use lists only for data sources, and use tuples otherwise. In the end there should only be one single list, or several lists with the same length. ``` This behavior was introduced when implementing multiprocessing by PR: - #5107 ### Steps to reproduce the bug ```python ds = load_dataset("ami", "microphone-single", split="train", revision="2d7620bb7c3f1aab9f329615c3bdb598069d907a") ``` ### Expected behavior No error raised. ### Environment info Since datasets 2.7.0
2023-01-18T14:09:04Z
5,415
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-10T07:36:11Z
https://api.github.com/repos/huggingface/datasets/issues/5415/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5415/timeline
RuntimeError: Sharding is ambiguous for this dataset
https://api.github.com/repos/huggingface/datasets/issues/5415/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
MEMBER
2023-01-18T14:09:03Z
null
I_kwDODunzps5bArgd
[]
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5415/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5415
https://github.com/huggingface/datasets/issues/5415
false
1,525,733,818
https://api.github.com/repos/huggingface/datasets/issues/5414/labels{/name}
### Describe the bug Loading the German Multilingual LibriSpeech dataset results in a RuntimeError regarding sharding with the following stacktrace: ``` Downloading and preparing dataset multilingual_librispeech/german to /home/nithin/datadrive/cache/huggingface/datasets/facebook___multilingual_librispeech/german/2.1.0/1904af50f57a5c370c9364cc337699cfe496d4e9edcae6648a96be23086362d0... Downloading data files: 100% 3/3 [00:00<00:00, 107.23it/s] Downloading data files: 100% 1/1 [00:00<00:00, 35.08it/s] Downloading data files: 100% 6/6 [00:00<00:00, 303.36it/s] Downloading data files: 100% 3/3 [00:00<00:00, 130.37it/s] Downloading data files: 100% 1049/1049 [00:00<00:00, 4491.40it/s] Downloading data files: 100% 37/37 [00:00<00:00, 1096.78it/s] Downloading data files: 100% 40/40 [00:00<00:00, 1003.93it/s] Extracting data files: 100% 3/3 [00:11<00:00, 2.62s/it] Generating train split: 469942/0 [34:13<00:00, 273.21 examples/s] Output exceeds the size limit. Open the full output data in a text editor --------------------------------------------------------------------------- RuntimeError Traceback (most recent call last) <ipython-input-14-74fa6d092bdc> in <module> ----> 1 mls = load_dataset(MLS_DATASET, 2 LANGUAGE, 3 cache_dir="~/datadrive/cache/huggingface/datasets", 4 ignore_verifications=True) /anaconda/envs/py38_default/lib/python3.8/site-packages/datasets/load.py in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, num_proc, **config_kwargs) 1755 1756 # Download and prepare data -> 1757 builder_instance.download_and_prepare( 1758 download_config=download_config, 1759 download_mode=download_mode, /anaconda/envs/py38_default/lib/python3.8/site-packages/datasets/builder.py in download_and_prepare(self, output_dir, download_config, download_mode, ignore_verifications, try_from_hf_gcs, dl_manager, base_path, use_auth_token, file_format, max_shard_size, num_proc, storage_options, **download_and_prepare_kwargs) 858 if num_proc is not None: 859 prepare_split_kwargs["num_proc"] = num_proc --> 860 self._download_and_prepare( 861 dl_manager=dl_manager, 862 verify_infos=verify_infos, /anaconda/envs/py38_default/lib/python3.8/site-packages/datasets/builder.py in _download_and_prepare(self, dl_manager, verify_infos, **prepare_splits_kwargs) 1609 1610 def _download_and_prepare(self, dl_manager, verify_infos, **prepare_splits_kwargs): ... RuntimeError: Sharding is ambiguous for this dataset: we found several data sources lists of different lengths, and we don't know over which list we should parallelize: - key audio_archives has length 1049 - key local_extracted_archive has length 1049 - key limited_ids_paths has length 1 To fix this, check the 'gen_kwargs' and make sure to use lists only for data sources, and use tuples otherwise. In the end there should only be one single list, or several lists with the same length. ``` ### Steps to reproduce the bug Here is the code to reproduce it: ```python from datasets import load_dataset MLS_DATASET = "facebook/multilingual_librispeech" LANGUAGE = "german" mls = load_dataset(MLS_DATASET, LANGUAGE, cache_dir="~/datadrive/cache/huggingface/datasets", ignore_verifications=True) ``` ### Expected behavior The expected behaviour is that the dataset is successfully loaded. ### Environment info - `datasets` version: 2.8.0 - Platform: Linux-5.4.0-1094-azure-x86_64-with-glibc2.10 - Python version: 3.8.8 - PyArrow version: 10.0.1 - Pandas version: 1.2.4
2023-01-18T14:09:04Z
5,414
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-09T14:45:31Z
https://api.github.com/repos/huggingface/datasets/issues/5414/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5414/timeline
Sharding error with Multilingual LibriSpeech
https://api.github.com/repos/huggingface/datasets/issues/5414/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/19574344?v=4", "events_url": "https://api.github.com/users/Nithin-Holla/events{/privacy}", "followers_url": "https://api.github.com/users/Nithin-Holla/followers", "following_url": "https://api.github.com/users/Nithin-Holla/following{/other_user}", "gists_url": "https://api.github.com/users/Nithin-Holla/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Nithin-Holla", "id": 19574344, "login": "Nithin-Holla", "node_id": "MDQ6VXNlcjE5NTc0MzQ0", "organizations_url": "https://api.github.com/users/Nithin-Holla/orgs", "received_events_url": "https://api.github.com/users/Nithin-Holla/received_events", "repos_url": "https://api.github.com/users/Nithin-Holla/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Nithin-Holla/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Nithin-Holla/subscriptions", "type": "User", "url": "https://api.github.com/users/Nithin-Holla" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
NONE
2023-01-18T14:09:04Z
null
I_kwDODunzps5a8Nm6
[ "Thanks for reporting, @Nithin-Holla.\r\n\r\nThis is a known issue for multiple datasets and we are investigating it:\r\n- See e.g.: https://huggingface.co/datasets/ami/discussions/3", "Main issue:\r\n- #5415", "@albertvillanova Thanks! As a workaround for now, can I use the dataset in streaming mode?", "Yes,...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5414/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5414
https://github.com/huggingface/datasets/issues/5414
false
1,524,591,837
https://api.github.com/repos/huggingface/datasets/issues/5413/labels{/name}
### Describe the bug When using `concatenate_datasets([dataset1, dataset2], axis = 1)` to concatenate two datasets with shards > 1, it fails: ``` File "/home/xzg/anaconda3/envs/tri-transfer/lib/python3.9/site-packages/datasets/combine.py", line 182, in concatenate_datasets return _concatenate_map_style_datasets(dsets, info=info, split=split, axis=axis) File "/home/xzg/anaconda3/envs/tri-transfer/lib/python3.9/site-packages/datasets/arrow_dataset.py", line 5499, in _concatenate_map_style_datasets table = concat_tables([dset._data for dset in dsets], axis=axis) File "/home/xzg/anaconda3/envs/tri-transfer/lib/python3.9/site-packages/datasets/table.py", line 1778, in concat_tables return ConcatenationTable.from_tables(tables, axis=axis) File "/home/xzg/anaconda3/envs/tri-transfer/lib/python3.9/site-packages/datasets/table.py", line 1483, in from_tables blocks = _extend_blocks(blocks, table_blocks, axis=axis) File "/home/xzg/anaconda3/envs/tri-transfer/lib/python3.9/site-packages/datasets/table.py", line 1477, in _extend_blocks result[i].extend(row_blocks) IndexError: list index out of range ``` ### Steps to reproduce the bug dataset = concatenate_datasets([dataset1, dataset2], axis = 1) ### Expected behavior The datasets are correctly concatenated. ### Environment info datasets==2.8.0
2023-01-26T09:27:21Z
5,413
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-08T17:01:52Z
https://api.github.com/repos/huggingface/datasets/issues/5413/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
https://api.github.com/repos/huggingface/datasets/issues/5413/timeline
concatenate_datasets fails when two dataset with shards > 1 and unequal shard numbers
https://api.github.com/repos/huggingface/datasets/issues/5413/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/38279341?v=4", "events_url": "https://api.github.com/users/ZeguanXiao/events{/privacy}", "followers_url": "https://api.github.com/users/ZeguanXiao/followers", "following_url": "https://api.github.com/users/ZeguanXiao/following{/other_user}", "gists_url": "https://api.github.com/users/ZeguanXiao/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ZeguanXiao", "id": 38279341, "login": "ZeguanXiao", "node_id": "MDQ6VXNlcjM4Mjc5MzQx", "organizations_url": "https://api.github.com/users/ZeguanXiao/orgs", "received_events_url": "https://api.github.com/users/ZeguanXiao/received_events", "repos_url": "https://api.github.com/users/ZeguanXiao/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ZeguanXiao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZeguanXiao/subscriptions", "type": "User", "url": "https://api.github.com/users/ZeguanXiao" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists...
null
completed
NONE
2023-01-26T09:27:21Z
null
I_kwDODunzps5a32zd
[ "Hi ! Thanks for reporting :)\r\n\r\nI managed to reproduce the hub using\r\n```python\r\n\r\nfrom datasets import concatenate_datasets, Dataset, load_from_disk\r\n\r\nDataset.from_dict({\"a\": range(9)}).save_to_disk(\"tmp/ds1\")\r\nds1 = load_from_disk(\"tmp/ds1\")\r\nds1 = concatenate_datasets([ds1, ds1])\r\n\r\...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5413/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5413
https://github.com/huggingface/datasets/issues/5413
false
1,524,250,269
https://api.github.com/repos/huggingface/datasets/issues/5412/labels{/name}
### Describe the bug I have a custom local dataset in JSON form. I am trying to do multiple training runs in parallel. The first training run runs with no issue. However, when I start another run on another GPU, the following code throws this error. If there is a workaround to ignore the cache I think that would solve my problem too. I am using datasets version 2.8.0. ### Steps to reproduce the bug 1. Start training run of GPU 0 loading dataset from ``` load_dataset( "json", data_files=tr_dataset_path, split=f"train", download_mode="force_redownload", ) ``` 2. While GPU 0 is training, start an identical run on GPU 1. GPU 1 will produce the following error: ``` Traceback (most recent call last): File "/local-scratch1/data/mt/code/qq/train.py", line 198, in <module> main() File "/home/username/.local/lib/python3.8/site-packages/click/core.py", line 1130, in __call__ return self.main(*args, **kwargs) File "/home/username/.local/lib/python3.8/site-packages/click/core.py", line 1055, in main rv = self.invoke(ctx) File "/home/username/.local/lib/python3.8/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/home/username/.local/lib/python3.8/site-packages/click/core.py", line 760, in invoke return __callback(*args, **kwargs) File "/local-scratch1/data/mt/code/qq/train.py", line 113, in main load_dataset( File "/home/username/miniconda3/envs/qq3/lib/python3.8/site-packages/datasets/load.py", line 1734, in load_dataset builder_instance = load_dataset_builder( File "/home/username/miniconda3/envs/qq3/lib/python3.8/site-packages/datasets/load.py", line 1518, in load_dataset_builder builder_instance: DatasetBuilder = builder_cls( File "/home/username/miniconda3/envs/qq3/lib/python3.8/site-packages/datasets/builder.py", line 366, in __init__ self.info = DatasetInfo.from_directory(self._cache_dir) File "/home/username/miniconda3/envs/qq3/lib/python3.8/site-packages/datasets/info.py", line 313, in from_directory with fs.open(path_join(dataset_info_dir, config.DATASET_INFO_FILENAME), "r", encoding="utf-8") as f: File "/home/username/miniconda3/envs/qq3/lib/python3.8/site-packages/fsspec/spec.py", line 1094, in open self.open( File "/home/username/miniconda3/envs/qq3/lib/python3.8/site-packages/fsspec/spec.py", line 1106, in open f = self._open( File "/home/username/miniconda3/envs/qq3/lib/python3.8/site-packages/fsspec/implementations/local.py", line 175, in _open return LocalFileOpener(path, mode, fs=self, **kwargs) File "/home/username/miniconda3/envs/qq3/lib/python3.8/site-packages/fsspec/implementations/local.py", line 273, in __init__ self._open() File "/home/username/miniconda3/envs/qq3/lib/python3.8/site-packages/fsspec/implementations/local.py", line 278, in _open self.f = open(self.path, mode=self.mode) FileNotFoundError: [Errno 2] No such file or directory: '/home/username/.cache/huggingface/datasets/json/default-43d06a4aedb25e6d/0.0.0/0f7e3662623656454fcd2b650f34e886a7db4b9104504885bd462096cc7a9f51/dataset_info.json' ``` ### Expected behavior Expected behavior: 2nd GPU training run should run the same as 1st GPU training run. ### Environment info Copy-and-paste the text below in your GitHub issue. - `datasets` version: 2.8.0 - Platform: Linux-5.4.0-120-generic-x86_64-with-glibc2.10 - Python version: 3.8.15 - PyArrow version: 9.0.0 - Pandas version: 1.5.2
2023-01-19T20:28:43Z
5,412
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-08T00:44:32Z
https://api.github.com/repos/huggingface/datasets/issues/5412/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5412/timeline
load_dataset() cannot find dataset_info.json with multiple training runs in parallel
https://api.github.com/repos/huggingface/datasets/issues/5412/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7139344?v=4", "events_url": "https://api.github.com/users/mtoles/events{/privacy}", "followers_url": "https://api.github.com/users/mtoles/followers", "following_url": "https://api.github.com/users/mtoles/following{/other_user}", "gists_url": "https://api.github.com/users/mtoles/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mtoles", "id": 7139344, "login": "mtoles", "node_id": "MDQ6VXNlcjcxMzkzNDQ=", "organizations_url": "https://api.github.com/users/mtoles/orgs", "received_events_url": "https://api.github.com/users/mtoles/received_events", "repos_url": "https://api.github.com/users/mtoles/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mtoles/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mtoles/subscriptions", "type": "User", "url": "https://api.github.com/users/mtoles" }
[]
null
completed
NONE
2023-01-19T20:28:43Z
null
I_kwDODunzps5a2jad
[ "Hi ! It fails because the dataset is already being prepared by your first run. I'd encourage you to prepare your dataset before using it for multiple trainings.\r\n\r\nYou can also specify another cache directory by passing `cache_dir=` to `load_dataset()`.", "Thank you! What do you mean by prepare it beforehand...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5412/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5412
https://github.com/huggingface/datasets/issues/5412
false
1,523,297,786
https://api.github.com/repos/huggingface/datasets/issues/5411/labels{/name}
[s3fs has migrated to all async calls](https://github.com/fsspec/s3fs/commit/0de2c6fb3d87c08ea694de96dca0d0834034f8bf). Updating documentation to use `AioSession` while using s3fs for download manager as well as working with datasets
2023-01-18T11:18:59Z
5,411
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-06T23:19:17Z
https://api.github.com/repos/huggingface/datasets/issues/5411/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5411/timeline
Update docs of S3 filesystem with async aiobotocore
https://api.github.com/repos/huggingface/datasets/issues/5411/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/5677912?v=4", "events_url": "https://api.github.com/users/maheshpec/events{/privacy}", "followers_url": "https://api.github.com/users/maheshpec/followers", "following_url": "https://api.github.com/users/maheshpec/following{/other_user}", "gists_url": "https://api.github.com/users/maheshpec/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/maheshpec", "id": 5677912, "login": "maheshpec", "node_id": "MDQ6VXNlcjU2Nzc5MTI=", "organizations_url": "https://api.github.com/users/maheshpec/orgs", "received_events_url": "https://api.github.com/users/maheshpec/received_events", "repos_url": "https://api.github.com/users/maheshpec/repos", "site_admin": false, "starred_url": "https://api.github.com/users/maheshpec/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/maheshpec/subscriptions", "type": "User", "url": "https://api.github.com/users/maheshpec" }
[]
null
null
CONTRIBUTOR
2023-01-18T11:12:04Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5411.diff", "html_url": "https://github.com/huggingface/datasets/pull/5411", "merged_at": "2023-01-18T11:12:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/5411.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5411" }
PR_kwDODunzps5G23-T
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5411/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5411
https://github.com/huggingface/datasets/pull/5411
true
1,521,168,032
https://api.github.com/repos/huggingface/datasets/issues/5410/labels{/name}
Added `ds.to_iterable()` to get an iterable dataset from a map-style arrow dataset. It also has a `num_shards` argument to split the dataset before converting to an iterable dataset. Sharding is important to enable efficient shuffling and parallel loading of iterable datasets. TODO: - [x] tests - [x] docs Fix https://github.com/huggingface/datasets/issues/5265
2023-02-01T18:11:45Z
5,410
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-05T18:12:17Z
https://api.github.com/repos/huggingface/datasets/issues/5410/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5410/timeline
Map-style Dataset to IterableDataset
https://api.github.com/repos/huggingface/datasets/issues/5410/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2023-02-01T16:36:01Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5410.diff", "html_url": "https://github.com/huggingface/datasets/pull/5410", "merged_at": "2023-02-01T16:36:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/5410.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5410" }
PR_kwDODunzps5GvnJH
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5410/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5410
https://github.com/huggingface/datasets/pull/5410
true
1,520,374,219
https://api.github.com/repos/huggingface/datasets/issues/5409/labels{/name}
The `DatasetBuilder.download_and_prepare` argument `use_auth_token` was deprecated in: - #5302 However, `use_auth_token` is still passed to `download_and_prepare` in our built-in `io` readers (csv, json, parquet,...). This PR fixes it, so that no deprecation warning is raised. Fix #5407.
2023-01-06T11:06:16Z
5,409
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-05T09:10:58Z
https://api.github.com/repos/huggingface/datasets/issues/5409/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5409/timeline
Fix deprecation warning when use_auth_token passed to download_and_prepare
https://api.github.com/repos/huggingface/datasets/issues/5409/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-06T10:59:13Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5409.diff", "html_url": "https://github.com/huggingface/datasets/pull/5409", "merged_at": "2023-01-06T10:59:13Z", "patch_url": "https://github.com/huggingface/datasets/pull/5409.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5409" }
PR_kwDODunzps5Gs3nL
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5409/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5409
https://github.com/huggingface/datasets/pull/5409
true
1,519,890,752
https://api.github.com/repos/huggingface/datasets/issues/5408/labels{/name}
### Describe the bug I follow the [blog post](https://huggingface.co/blog/fine-tune-whisper#building-a-demo) to finetune a Cantonese transcribe model. When using map function to prepare dataset, following warning pop out: `common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names["train"], num_proc=1)` > Parameter 'function'=<function prepare_dataset at 0x000001D1D9D79A60> of the transform datasets.arrow_dataset.Dataset._map_single couldn't be hashed properly, a random hash was used instead. Make sure your transforms and parameters are serializable with pickle or dill for the dataset fingerprinting and caching to work. If you reuse this transform, the caching mechanism will consider it to be different from the previous calls and recompute everything. This warning is only showed once. Subsequent hashing failures won't be showed. I read https://github.com/huggingface/datasets/issues/4521 and https://github.com/huggingface/datasets/issues/3178 but cannot solve the issue. ### Steps to reproduce the bug ```python from datasets import load_dataset, DatasetDict common_voice = DatasetDict() common_voice["train"] = load_dataset("mozilla-foundation/common_voice_11_0", "zh-HK", split="train+validation") common_voice["test"] = load_dataset("mozilla-foundation/common_voice_11_0", "zh-HK", split="test") common_voice = common_voice.remove_columns(["accent", "age", "client_id", "down_votes", "gender", "locale", "path", "segment", "up_votes"]) from transformers import WhisperFeatureExtractor, WhisperTokenizer, WhisperProcessor feature_extractor = WhisperFeatureExtractor.from_pretrained("openai/whisper-small") tokenizer = WhisperTokenizer.from_pretrained("openai/whisper-small", language="chinese", task="transcribe") processor = WhisperProcessor.from_pretrained("openai/whisper-small", language="chinese", task="transcribe") from datasets import Audio common_voice = common_voice.cast_column("audio", Audio(sampling_rate=16000)) def prepare_dataset(batch): # load and resample audio data from 48 to 16kHz audio = batch["audio"] # compute log-Mel input features from input audio array batch["input_features"] = feature_extractor(audio["array"], sampling_rate=audio["sampling_rate"]).input_features[0] # encode target text to label ids batch["labels"] = tokenizer(batch["sentence"]).input_ids return batch common_voice = common_voice.map(prepare_dataset, remove_columns=common_voice.column_names["train"], num_proc=1) ``` ### Expected behavior Should be no warning shown. ### Environment info - `datasets` version: 2.7.0 - Platform: Windows-10-10.0.19045-SP0 - Python version: 3.9.12 - PyArrow version: 8.0.0 - Pandas version: 1.3.5 - dill version: 0.3.4 - multiprocess version: 0.70.12.2
2023-01-06T13:22:19Z
5,408
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-05T01:59:59Z
https://api.github.com/repos/huggingface/datasets/issues/5408/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5408/timeline
dataset map function could not be hash properly
https://api.github.com/repos/huggingface/datasets/issues/5408/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/68179274?v=4", "events_url": "https://api.github.com/users/Tungway1990/events{/privacy}", "followers_url": "https://api.github.com/users/Tungway1990/followers", "following_url": "https://api.github.com/users/Tungway1990/following{/other_user}", "gists_url": "https://api.github.com/users/Tungway1990/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Tungway1990", "id": 68179274, "login": "Tungway1990", "node_id": "MDQ6VXNlcjY4MTc5Mjc0", "organizations_url": "https://api.github.com/users/Tungway1990/orgs", "received_events_url": "https://api.github.com/users/Tungway1990/received_events", "repos_url": "https://api.github.com/users/Tungway1990/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Tungway1990/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Tungway1990/subscriptions", "type": "User", "url": "https://api.github.com/users/Tungway1990" }
[]
null
completed
NONE
2023-01-06T13:22:18Z
null
I_kwDODunzps5al7FA
[ "Hi ! On macos I tried with\r\n- py 3.9.11\r\n- datasets 2.8.0\r\n- transformers 4.25.1\r\n- dill 0.3.4\r\n\r\nand I was able to hash `prepare_dataset` correctly:\r\n```python\r\nfrom datasets.fingerprint import Hasher\r\nHasher.hash(prepare_dataset)\r\n```\r\n\r\nWhat version of transformers do you have ? Can you ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5408/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5408
https://github.com/huggingface/datasets/issues/5408
false
1,519,797,345
https://api.github.com/repos/huggingface/datasets/issues/5407/labels{/name}
### Describe the bug Calling `Datasets.from_sql()` generates a warning: `.../site-packages/datasets/builder.py:712: FutureWarning: 'use_auth_token' was deprecated in version 2.7.1 and will be removed in 3.0.0. Pass 'use_auth_token' to the initializer/'load_dataset_builder' instead.` ### Steps to reproduce the bug Any valid call to `Datasets.from_sql()` will produce the deprecation warning. ### Expected behavior No warning. The fix should be simply to remove the parameter `use_auth_token` from the call to `builder.download_and_prepare()` at line 43 of `io/sql.py` (it is set to `None` anyway, and is not needed). ### Environment info - `datasets` version: 2.8.0 - Platform: Linux-4.15.0-169-generic-x86_64-with-glibc2.27 - Python version: 3.9.15 - PyArrow version: 10.0.1 - Pandas version: 1.5.2
2023-01-06T10:59:14Z
5,407
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-05T00:43:17Z
https://api.github.com/repos/huggingface/datasets/issues/5407/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/5407/timeline
Datasets.from_sql() generates deprecation warning
https://api.github.com/repos/huggingface/datasets/issues/5407/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/21002157?v=4", "events_url": "https://api.github.com/users/msummerfield/events{/privacy}", "followers_url": "https://api.github.com/users/msummerfield/followers", "following_url": "https://api.github.com/users/msummerfield/following{/other_user}", "gists_url": "https://api.github.com/users/msummerfield/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/msummerfield", "id": 21002157, "login": "msummerfield", "node_id": "MDQ6VXNlcjIxMDAyMTU3", "organizations_url": "https://api.github.com/users/msummerfield/orgs", "received_events_url": "https://api.github.com/users/msummerfield/received_events", "repos_url": "https://api.github.com/users/msummerfield/repos", "site_admin": false, "starred_url": "https://api.github.com/users/msummerfield/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/msummerfield/subscriptions", "type": "User", "url": "https://api.github.com/users/msummerfield" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
NONE
2023-01-06T10:59:14Z
null
I_kwDODunzps5alkRh
[ "Thanks for reporting @msummerfield. We are fixing it." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5407/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5407
https://github.com/huggingface/datasets/issues/5407
false
1,519,140,544
https://api.github.com/repos/huggingface/datasets/issues/5406/labels{/name}
`datasets` 2.6.1 and 2.7.0 started to stop supporting datasets like IMDB, ConLL or MNIST datasets. When loading a dataset using 2.6.1 or 2.7.0, you may this error when loading certain datasets: ```python TypeError: can only concatenate str (not "int") to str ``` This is because we started to update the metadata of those datasets to a format that is not supported in 2.6.1 and 2.7.0 This change is required or those datasets won't be supported by the Hugging Face Hub. Therefore if you encounter this error or if you're using `datasets` 2.6.1 or 2.7.0, we encourage you to update to a newer version. For example, versions 2.6.2 and 2.7.1 patch this issue. ```python pip install -U datasets ``` All the datasets affected are the ones with a ClassLabel feature type and YAML "dataset_info" metadata. More info [here](https://github.com/huggingface/datasets/issues/5275). We apologize for the inconvenience.
2023-06-21T18:45:38Z
5,406
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-04T15:10:04Z
https://api.github.com/repos/huggingface/datasets/issues/5406/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5406/timeline
[2.6.1][2.7.0] Upgrade `datasets` to fix `TypeError: can only concatenate str (not "int") to str`
https://api.github.com/repos/huggingface/datasets/issues/5406/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
null
null
I_kwDODunzps5ajD7A
[ "I still get this error on 2.9.0\r\n<img width=\"1925\" alt=\"image\" src=\"https://user-images.githubusercontent.com/7208470/215597359-2f253c76-c472-4612-8099-d3a74d16eb29.png\">\r\n", "Hi ! I just tested locally and or colab and it works fine for 2.9 on `sst2`.\r\n\r\nAlso the code that is shown in your stack t...
{ "+1": 11, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 11, "url": "https://api.github.com/repos/huggingface/datasets/issues/5406/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5406
https://github.com/huggingface/datasets/issues/5406
false
1,517,879,386
https://api.github.com/repos/huggingface/datasets/issues/5405/labels{/name}
### Describe the bug Hi, it looks like whenever you pull a dataset and get size_in_bytes, it returns the same size for all splits (and that size is the combined size of all splits). It seems like this shouldn't be the intended behavior since it is misleading. Here's an example: ``` >>> from datasets import load_dataset >>> x = load_dataset("glue", "wnli") Found cached dataset glue (/Users/breakend/.cache/huggingface/datasets/glue/wnli/1.0.0/dacbe3125aa31d7f70367a07a8a9e72a5a0bfeb5fc42e75c9db75b96da6053ad) 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3/3 [00:00<00:00, 1097.70it/s] >>> x["train"].size_in_bytes 186159 >>> x["validation"].size_in_bytes 186159 >>> x["test"].size_in_bytes 186159 >>> ``` ### Steps to reproduce the bug ``` >>> from datasets import load_dataset >>> x = load_dataset("glue", "wnli") >>> x["train"].size_in_bytes 186159 >>> x["validation"].size_in_bytes 186159 >>> x["test"].size_in_bytes 186159 ``` ### Expected behavior The expected behavior is that it should return the separate sizes for all splits. ### Environment info - `datasets` version: 2.7.1 - Platform: macOS-12.5-arm64-arm-64bit - Python version: 3.10.8 - PyArrow version: 10.0.1 - Pandas version: 1.5.2
2023-01-04T09:22:59Z
5,405
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-03T20:25:48Z
https://api.github.com/repos/huggingface/datasets/issues/5405/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5405/timeline
size_in_bytes the same for all splits
https://api.github.com/repos/huggingface/datasets/issues/5405/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1609857?v=4", "events_url": "https://api.github.com/users/Breakend/events{/privacy}", "followers_url": "https://api.github.com/users/Breakend/followers", "following_url": "https://api.github.com/users/Breakend/following{/other_user}", "gists_url": "https://api.github.com/users/Breakend/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Breakend", "id": 1609857, "login": "Breakend", "node_id": "MDQ6VXNlcjE2MDk4NTc=", "organizations_url": "https://api.github.com/users/Breakend/orgs", "received_events_url": "https://api.github.com/users/Breakend/received_events", "repos_url": "https://api.github.com/users/Breakend/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Breakend/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Breakend/subscriptions", "type": "User", "url": "https://api.github.com/users/Breakend" }
[]
null
null
NONE
null
null
I_kwDODunzps5aeQBa
[ "Hi @Breakend,\r\n\r\nIndeed, the attribute `size_in_bytes` refers to the size of the entire dataset configuration, for all splits (size of downloaded files + Arrow files), not the specific split.\r\nThis is also the case for `download_size` (downloaded files) and `dataset_size` (Arrow files).\r\n\r\nThe size of th...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5405/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5405
https://github.com/huggingface/datasets/issues/5405
false
1,517,566,331
https://api.github.com/repos/huggingface/datasets/issues/5404/labels{/name}
### Feature request Ideally, it would be nice to have a maintained PyPI package for `bigbench`. ### Motivation We'd like to allow anyone to access, explore and use any task. ### Your contribution @lhoestq has opened an issue in their repo: - https://github.com/google/BIG-bench/issues/906
2023-02-09T20:30:26Z
5,404
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2023-01-03T15:37:57Z
https://api.github.com/repos/huggingface/datasets/issues/5404/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5404/timeline
Better integration of BIG-bench
https://api.github.com/repos/huggingface/datasets/issues/5404/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
null
null
I_kwDODunzps5adDl7
[ "Hi, I made my version : https://huggingface.co/datasets/tasksource/bigbench" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5404/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5404
https://github.com/huggingface/datasets/issues/5404
false
1,517,466,492
https://api.github.com/repos/huggingface/datasets/issues/5403/labels{/name}
This PR updates a code example for consistency across the docs based on [feedback from this comment](https://github.com/huggingface/transformers/pull/20925/files/9fda31634d203a47d3212e4e8d43d3267faf9808#r1058769500): "In terms of style we usually stay away from one-letter imports like this (even if the community uses them) as they are not always known by beginners and one letter is very undescriptive. Here it wouldn't change anything to use albumentations instead of A."
2023-01-03T15:06:18Z
5,403
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-03T14:26:32Z
https://api.github.com/repos/huggingface/datasets/issues/5403/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5403/timeline
Replace one letter import in docs
https://api.github.com/repos/huggingface/datasets/issues/5403/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1065417?v=4", "events_url": "https://api.github.com/users/MKhalusova/events{/privacy}", "followers_url": "https://api.github.com/users/MKhalusova/followers", "following_url": "https://api.github.com/users/MKhalusova/following{/other_user}", "gists_url": "https://api.github.com/users/MKhalusova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/MKhalusova", "id": 1065417, "login": "MKhalusova", "node_id": "MDQ6VXNlcjEwNjU0MTc=", "organizations_url": "https://api.github.com/users/MKhalusova/orgs", "received_events_url": "https://api.github.com/users/MKhalusova/received_events", "repos_url": "https://api.github.com/users/MKhalusova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/MKhalusova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MKhalusova/subscriptions", "type": "User", "url": "https://api.github.com/users/MKhalusova" }
[]
null
null
CONTRIBUTOR
2023-01-03T14:59:01Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5403.diff", "html_url": "https://github.com/huggingface/datasets/pull/5403", "merged_at": "2023-01-03T14:59:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/5403.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5403" }
PR_kwDODunzps5Gi3d9
[ "_The documentation is not available anymore as the PR was closed or merged._", "> Thanks for the docs fix for consistency.\r\n> \r\n> Again for consistency, it would be nice to make the same fix across all the docs, e.g.\r\n> \r\n> https://github.com/huggingface/datasets/blob/310cdddd1c43f9658de172b85b6509d07d5e...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/5403/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5403
https://github.com/huggingface/datasets/pull/5403
true
1,517,409,429
https://api.github.com/repos/huggingface/datasets/issues/5402/labels{/name}
### Describe the bug Using `load_dataset_builder` to create a builder, run `download_and_prepare` do upload it to S3. However when trying to load it, there are missing `state.json` files. Complete example: ```python from aiobotocore.session import AioSession as Session from datasets import load_from_disk, load_datase, load_dataset_builder import s3fs storage_options = {"session": Session()} fs = s3fs.S3FileSystem(**storage_options) output_dir = "s3://bucket/imdb" builder = load_dataset_builder("imdb") builder.download_and_prepare(output_dir, storage_options=storage_options) load_from_disk(output_dir, fs=fs) # ERROR # [Errno 2] No such file or directory: '/tmp/tmpy22yys8o/bucket/imdb/state.json' ``` As a comparison, if you use the non lazy `load_dataset`, it works and the S3 folder has different structure + state.json files. Example: ```python from aiobotocore.session import AioSession as Session from datasets import load_from_disk, load_dataset, load_dataset_builder import s3fs storage_options = {"session": Session()} fs = s3fs.S3FileSystem(**storage_options) output_dir = "s3://bucket/imdb" dataset = load_dataset("imdb",) dataset.save_to_disk(output_dir, fs=fs) load_from_disk(output_dir, fs=fs) # WORKS ``` You still want the 1st option for the laziness and the parquet conversion. Thanks! ### Steps to reproduce the bug ```python from aiobotocore.session import AioSession as Session from datasets import load_from_disk, load_datase, load_dataset_builder import s3fs storage_options = {"session": Session()} fs = s3fs.S3FileSystem(**storage_options) output_dir = "s3://bucket/imdb" builder = load_dataset_builder("imdb") builder.download_and_prepare(output_dir, storage_options=storage_options) load_from_disk(output_dir, fs=fs) # ERROR # [Errno 2] No such file or directory: '/tmp/tmpy22yys8o/bucket/imdb/state.json' ``` BTW, you need the AioSession as s3fs is now based on aiobotocore, see https://github.com/fsspec/s3fs/issues/385. ### Expected behavior Expected to be able to load the dataset from S3. ### Environment info ``` s3fs 2022.11.0 s3transfer 0.6.0 datasets 2.8.0 aiobotocore 2.4.2 boto3 1.24.59 botocore 1.27.59 ``` python 3.7.15.
2023-01-04T17:23:57Z
5,402
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-03T13:39:59Z
https://api.github.com/repos/huggingface/datasets/issues/5402/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5402/timeline
Missing state.json when creating a cloud dataset using a dataset_builder
https://api.github.com/repos/huggingface/datasets/issues/5402/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/22022514?v=4", "events_url": "https://api.github.com/users/danielfleischer/events{/privacy}", "followers_url": "https://api.github.com/users/danielfleischer/followers", "following_url": "https://api.github.com/users/danielfleischer/following{/other_user}", "gists_url": "https://api.github.com/users/danielfleischer/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/danielfleischer", "id": 22022514, "login": "danielfleischer", "node_id": "MDQ6VXNlcjIyMDIyNTE0", "organizations_url": "https://api.github.com/users/danielfleischer/orgs", "received_events_url": "https://api.github.com/users/danielfleischer/received_events", "repos_url": "https://api.github.com/users/danielfleischer/repos", "site_admin": false, "starred_url": "https://api.github.com/users/danielfleischer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/danielfleischer/subscriptions", "type": "User", "url": "https://api.github.com/users/danielfleischer" }
[]
null
null
NONE
null
null
I_kwDODunzps5acdSV
[ "`load_from_disk` must be used on datasets saved using `save_to_disk`: they correspond to fully serialized datasets including their state.\r\n\r\nOn the other hand, `download_and_prepare` just downloads the raw data and convert them to arrow (or parquet if you want). We are working on allowing you to reload a datas...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5402/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5402
https://github.com/huggingface/datasets/issues/5402
false
1,517,160,935
https://api.github.com/repos/huggingface/datasets/issues/5401/labels{/name}
This PR implements Spark integration by supporting `Dataset` conversion from/to Spark `DataFrame`.
2023-01-05T14:21:33Z
5,401
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-03T09:57:40Z
https://api.github.com/repos/huggingface/datasets/issues/5401/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5401/timeline
Support Dataset conversion from/to Spark
https://api.github.com/repos/huggingface/datasets/issues/5401/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
null
{ "diff_url": "https://github.com/huggingface/datasets/pull/5401.diff", "html_url": "https://github.com/huggingface/datasets/pull/5401", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5401.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5401" }
PR_kwDODunzps5Gh1XQ
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5401). All of your documentation changes will be reflected on that endpoint.", "Cool thanks !\r\n\r\nSpark DataFrame are usually quite big, and I believe here `from_spark` would load everything in the driver node's RAM, which i...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5401/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/5401
https://github.com/huggingface/datasets/pull/5401
true
1,517,032,972
https://api.github.com/repos/huggingface/datasets/issues/5400/labels{/name}
Support streaming datasets with `os.path.exists` and `pathlib.Path.exists`.
2023-01-06T10:42:44Z
5,400
null
https://api.github.com/repos/huggingface/datasets
false
[]
2023-01-03T07:42:37Z
https://api.github.com/repos/huggingface/datasets/issues/5400/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5400/timeline
Support streaming datasets with os.path.exists and Path.exists
https://api.github.com/repos/huggingface/datasets/issues/5400/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2023-01-06T10:35:44Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5400.diff", "html_url": "https://github.com/huggingface/datasets/pull/5400", "merged_at": "2023-01-06T10:35:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/5400.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5400" }
PR_kwDODunzps5GhaGI
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5400/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5400
https://github.com/huggingface/datasets/pull/5400
true
1,515,548,427
https://api.github.com/repos/huggingface/datasets/issues/5399/labels{/name}
### Describe the bug While trying to upload my image dataset of a CSV file type to huggingface by running the below code. The dataset consists of a little over 100k of image-caption pairs ### Steps to reproduce the bug ``` df = pd.read_csv('x.csv', encoding='utf-8-sig') features = Features({ 'link': Image(decode=True), 'caption': Value(dtype='string'), }) #make sure u r logged in to HF ds = Dataset.from_pandas(df, features=features) ds.features ds.push_to_hub("x/x") ``` I got the below error and It always stops at the same progress ``` 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [23:53<00:00, 358.48s/ba] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [24:37<00:00, 369.47s/ba]%|▍ | 1/22 [00:06<02:09, 6.16s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [25:00<00:00, 375.15s/ba]%|β–‰ | 2/22 [25:54<2:36:15, 468.80s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [24:53<00:00, 373.29s/ba]%|β–ˆβ–Ž | 3/22 [51:01<4:07:07, 780.39s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [24:01<00:00, 360.34s/ba]%|β–ˆβ–Š | 4/22 [1:17:00<5:04:07, 1013.74s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [23:59<00:00, 359.91s/ba]%|β–ˆβ–ˆβ–Ž | 5/22 [1:41:07<5:24:06, 1143.90s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [24:16<00:00, 364.06s/ba]%|β–ˆβ–ˆβ–‹ | 6/22 [2:05:14<5:29:15, 1234.74s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [25:24<00:00, 381.10s/ba]%|β–ˆβ–ˆβ–ˆβ– | 7/22 [2:29:38<5:25:52, 1303.52s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [25:24<00:00, 381.24s/ba]%|β–ˆβ–ˆβ–ˆβ–‹ | 8/22 [2:56:02<5:23:46, 1387.58s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [25:08<00:00, 377.23s/ba]%|β–ˆβ–ˆβ–ˆβ–ˆ | 9/22 [3:22:24<5:13:17, 1445.97s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [24:11<00:00, 362.87s/ba]%|β–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 10/22 [3:48:24<4:56:02, 1480.19s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [24:44<00:00, 371.11s/ba]%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 11/22 [4:12:42<4:30:10, 1473.66s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [24:35<00:00, 368.81s/ba]%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 12/22 [4:37:34<4:06:29, 1478.98s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [24:02<00:00, 360.67s/ba]%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰ | 13/22 [5:03:24<3:45:04, 1500.45s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [24:07<00:00, 361.78s/ba]%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 14/22 [5:27:33<3:17:59, 1484.97s/it] 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 4/4 [23:39<00:00, 354.85s/ba]%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 15/22 [5:51:48<2:52:10, 1475.82s/it] Pushing dataset shards to the dataset hub: 73%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 16/22 [6:16:58<2:28:37, 1486.31s/it]Got disconnected from remote data host. Retrying in 5sec [1/20] Got disconnected from remote data host. Retrying in 5sec [2/20] Got disconnected from remote data host. Retrying in 5sec [3/20] Got disconnected from remote data host. Retrying in 5sec [4/20] Got disconnected from remote data host. Retrying in 5sec [5/20] Got disconnected from remote data host. Retrying in 5sec [6/20] Got disconnected from remote data host. Retrying in 5sec [7/20] Got disconnected from remote data host. Retrying in 5sec [8/20] Got disconnected from remote data host. Retrying in 5sec [9/20] ... Got disconnected from remote data host. Retrying in 5sec [19/20] Got disconnected from remote data host. Retrying in 5sec [20/20] 75%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 3/4 [24:47<08:15, 495.86s/ba] Pushing dataset shards to the dataset hub: 73%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 16/22 [6:41:46<2:30:39, 1506.65s/it] Output exceeds the size limit. Open the full output data in a text editor --------------------------------------------------------------------------- ConnectionError Traceback (most recent call last) <ipython-input-1-dbf8530779e9> in <module> 16 ds.features ``` ### Expected behavior I was trying to upload an image dataset and expected it to be fully uploaded ### Environment info - `datasets` version: 2.8.0 - Platform: Windows-10-10.0.19041-SP0 - Python version: 3.7.9 - PyArrow version: 10.0.1 - Pandas version: 1.3.5
2023-01-02T07:21:52Z
5,399
null
https://api.github.com/repos/huggingface/datasets
null
[]
2023-01-01T13:00:11Z
https://api.github.com/repos/huggingface/datasets/issues/5399/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5399/timeline
Got disconnected from remote data host. Retrying in 5sec [2/20]
https://api.github.com/repos/huggingface/datasets/issues/5399/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/46427957?v=4", "events_url": "https://api.github.com/users/alhuri/events{/privacy}", "followers_url": "https://api.github.com/users/alhuri/followers", "following_url": "https://api.github.com/users/alhuri/following{/other_user}", "gists_url": "https://api.github.com/users/alhuri/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/alhuri", "id": 46427957, "login": "alhuri", "node_id": "MDQ6VXNlcjQ2NDI3OTU3", "organizations_url": "https://api.github.com/users/alhuri/orgs", "received_events_url": "https://api.github.com/users/alhuri/received_events", "repos_url": "https://api.github.com/users/alhuri/repos", "site_admin": false, "starred_url": "https://api.github.com/users/alhuri/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alhuri/subscriptions", "type": "User", "url": "https://api.github.com/users/alhuri" }
[]
null
completed
NONE
2023-01-02T07:21:52Z
null
I_kwDODunzps5aVW8L
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5399/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5399
https://github.com/huggingface/datasets/issues/5399
false
1,514,425,231
https://api.github.com/repos/huggingface/datasets/issues/5398/labels{/name}
Once `pydantic` fixes their issue in their 1.10.3 version, unpin it. See issue: - #5394 See temporary fix: - #5395
2022-12-30T10:43:41Z
5,398
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-12-30T10:37:31Z
https://api.github.com/repos/huggingface/datasets/issues/5398/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5398/timeline
Unpin pydantic
https://api.github.com/repos/huggingface/datasets/issues/5398/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
completed
MEMBER
2022-12-30T10:43:41Z
null
I_kwDODunzps5aREuP
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5398/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5398
https://github.com/huggingface/datasets/issues/5398
false
1,514,412,246
https://api.github.com/repos/huggingface/datasets/issues/5397/labels{/name}
Once pydantic-1.10.3 has been yanked, we can unpin it: https://pypi.org/project/pydantic/1.10.3/ See reply by pydantic team https://github.com/pydantic/pydantic/issues/4885#issuecomment-1367819807 ``` v1.10.3 has been yanked. ``` in response to spacy request: https://github.com/pydantic/pydantic/issues/4885#issuecomment-1367810049 ``` On behalf of spacy-related packages: would it be possible for you to temporarily yank v1.10.3? To address this and be compatible with v1.10.4, we'd have to release new versions of a whole series of packages and nearly everyone (including me) is currently on vacation. Even if v1.10.4 is released with a fix, pip would still back off to v1.10.3 for spacy, etc. because of its current pins for typing_extensions. If it could instead back off to v1.10.2, we'd have a bit more breathing room to make the updates on our end. ``` Close #5398.
2022-12-30T10:53:11Z
5,397
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-12-30T10:22:09Z
https://api.github.com/repos/huggingface/datasets/issues/5397/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5397/timeline
Unpin pydantic test dependency
https://api.github.com/repos/huggingface/datasets/issues/5397/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-12-30T10:43:40Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5397.diff", "html_url": "https://github.com/huggingface/datasets/pull/5397", "merged_at": "2022-12-30T10:43:40Z", "patch_url": "https://github.com/huggingface/datasets/pull/5397.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5397" }
PR_kwDODunzps5GYirs
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==6.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5397/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5397
https://github.com/huggingface/datasets/pull/5397
true
1,514,002,934
https://api.github.com/repos/huggingface/datasets/issues/5396/labels{/name}
Expected checksum was verified against checksum dict (not checksum).
2023-02-13T11:11:22Z
5,396
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-12-29T19:45:17Z
https://api.github.com/repos/huggingface/datasets/issues/5396/comments
null
https://api.github.com/repos/huggingface/datasets/issues/5396/timeline
Fix checksum verification
https://api.github.com/repos/huggingface/datasets/issues/5396/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/9336514?v=4", "events_url": "https://api.github.com/users/daskol/events{/privacy}", "followers_url": "https://api.github.com/users/daskol/followers", "following_url": "https://api.github.com/users/daskol/following{/other_user}", "gists_url": "https://api.github.com/users/daskol/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/daskol", "id": 9336514, "login": "daskol", "node_id": "MDQ6VXNlcjkzMzY1MTQ=", "organizations_url": "https://api.github.com/users/daskol/orgs", "received_events_url": "https://api.github.com/users/daskol/received_events", "repos_url": "https://api.github.com/users/daskol/repos", "site_admin": false, "starred_url": "https://api.github.com/users/daskol/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/daskol/subscriptions", "type": "User", "url": "https://api.github.com/users/daskol" }
[]
null
null
CONTRIBUTOR
2023-02-13T11:11:22Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/5396.diff", "html_url": "https://github.com/huggingface/datasets/pull/5396", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/5396.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/5396" }
PR_kwDODunzps5GXMhp
[ "Hi ! If I'm not mistaken both `expected_checksums[url]` and `recorded_checksums[url]` are dictionaries with keys \"checksum\" and \"num_bytes\". So we need to check whether `expected_checksums[url] != recorded_checksums[url]` (or simply `expected_checksums[url][\"checksum\"] != recorded_checksums[url][\"checksum\"...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/5396/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/5396
https://github.com/huggingface/datasets/pull/5396
true