id
int64
1.14B
2.23B
labels_url
stringlengths
75
75
body
stringlengths
2
33.9k
updated_at
stringlengths
20
20
number
int64
3.76k
6.79k
milestone
dict
repository_url
stringclasses
1 value
draft
bool
2 classes
labels
listlengths
0
4
created_at
stringlengths
20
20
comments_url
stringlengths
70
70
assignee
dict
timeline_url
stringlengths
70
70
title
stringlengths
1
290
events_url
stringlengths
68
68
active_lock_reason
null
user
dict
assignees
listlengths
0
3
performed_via_github_app
null
state_reason
stringclasses
3 values
author_association
stringclasses
3 values
closed_at
stringlengths
20
20
pull_request
dict
node_id
stringlengths
18
19
comments
listlengths
0
30
reactions
dict
state
stringclasses
2 values
locked
bool
1 class
url
stringlengths
61
61
html_url
stringlengths
49
51
is_pull_request
bool
2 classes
1,375,895,035
https://api.github.com/repos/huggingface/datasets/issues/4986/labels{/name}
Hello! ### Pull request overview * Fix broken snippet in https://huggingface.co/docs/datasets/main/en/process that has too many quotes ### Details The snippet in question can be found here: https://huggingface.co/docs/datasets/main/en/process#map This screenshot shows the issue, there is a quote too many, causing the snippet to be colored incorrectly: ![image](https://user-images.githubusercontent.com/37621491/190640627-f7587362-0e44-4464-a5d1-a0b98df6986f.png) The change speaks for itself. Thank you for the detailed documentation, by the way. - Tom Aarsen
2022-09-16T22:12:21Z
4,986
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-16T12:41:07Z
https://api.github.com/repos/huggingface/datasets/issues/4986/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4986/timeline
[doc] Fix broken snippet that had too many quotes
https://api.github.com/repos/huggingface/datasets/issues/4986/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/37621491?v=4", "events_url": "https://api.github.com/users/tomaarsen/events{/privacy}", "followers_url": "https://api.github.com/users/tomaarsen/followers", "following_url": "https://api.github.com/users/tomaarsen/following{/other_user}", "gists_url": "https://api.github.com/users/tomaarsen/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/tomaarsen", "id": 37621491, "login": "tomaarsen", "node_id": "MDQ6VXNlcjM3NjIxNDkx", "organizations_url": "https://api.github.com/users/tomaarsen/orgs", "received_events_url": "https://api.github.com/users/tomaarsen/received_events", "repos_url": "https://api.github.com/users/tomaarsen/repos", "site_admin": false, "starred_url": "https://api.github.com/users/tomaarsen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tomaarsen/subscriptions", "type": "User", "url": "https://api.github.com/users/tomaarsen" }
[]
null
null
MEMBER
2022-09-16T17:32:14Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4986.diff", "html_url": "https://github.com/huggingface/datasets/pull/4986", "merged_at": "2022-09-16T17:32:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/4986.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4986" }
PR_kwDODunzps4_GNSd
[ "_The documentation is not available anymore as the PR was closed or merged._", "Spent the day familiarising myself with the huggingface line of products, and happened to run into some small issues here and there. Magically, I've found exactly one small issue in `transformers`, one in `accelerate` and now one in ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4986/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4986
https://github.com/huggingface/datasets/pull/4986
true
1,375,807,768
https://api.github.com/repos/huggingface/datasets/issues/4985/labels{/name}
related to https://github.com/huggingface/datasets/issues/4895
2022-11-02T11:54:28Z
4,985
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-16T11:20:40Z
https://api.github.com/repos/huggingface/datasets/issues/4985/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4985/timeline
Prefer split patterns from directories over split patterns from filenames
https://api.github.com/repos/huggingface/datasets/issues/4985/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
null
null
CONTRIBUTOR
2022-09-29T08:07:49Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4985.diff", "html_url": "https://github.com/huggingface/datasets/pull/4985", "merged_at": "2022-09-29T08:07:49Z", "patch_url": "https://github.com/huggingface/datasets/pull/4985.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4985" }
PR_kwDODunzps4_F6kU
[ "_The documentation is not available anymore as the PR was closed or merged._", "Can we merge this one since the issue this PR fixes was reported for the second time? I also think we don't need a test for this simple change.", "@mariosasko sure! could you please approve it? ", "Hi there @polinaeterna @mariosa...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4985/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4985
https://github.com/huggingface/datasets/pull/4985
true
1,375,690,330
https://api.github.com/repos/huggingface/datasets/issues/4984/labels{/name}
I added some links to the Datasets API in the docs. See https://github.com/huggingface/datasets-server/pull/566 for a companion PR in the datasets-server. The idea is to improve the discovery of the API through the docs. I'm a bit shy about pasting a lot of links to the API in the docs, so it's minimal for now. I'm interested in ideas to integrate the API better in these docs without being too much. cc @lhoestq @julien-c @albertvillanova @stevhliu.
2022-09-16T13:10:14Z
4,984
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-16T09:34:12Z
https://api.github.com/repos/huggingface/datasets/issues/4984/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4984/timeline
docs: ✏️ add links to the Datasets API
https://api.github.com/repos/huggingface/datasets/issues/4984/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[]
null
null
CONTRIBUTOR
2022-09-16T13:07:33Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4984.diff", "html_url": "https://github.com/huggingface/datasets/pull/4984", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4984.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4984" }
PR_kwDODunzps4_FhTm
[ "_The documentation is not available anymore as the PR was closed or merged._", "OK, thanks @lhoestq. I'll close this PR, and come back to it with @stevhliu once we work on https://github.com/huggingface/datasets-server/issues/568" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4984/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4984
https://github.com/huggingface/datasets/pull/4984
true
1,375,667,654
https://api.github.com/repos/huggingface/datasets/issues/4983/labels{/name}
I look through the huggingface dataset docs, and it seems that there is no offical support function to convert `torch.utils.data.Dataset` to huggingface dataset. However, there is a way to convert huggingface dataset to `torch.utils.data.Dataset`, like below: ```python from datasets import Dataset data = [[1, 2],[3, 4]] ds = Dataset.from_dict({"data": data}) ds = ds.with_format("torch") ds[0] ds[:2] ``` So is there something I miss, or there IS no function to convert `torch.utils.data.Dataset` to huggingface dataset. If so, is there any way to do this convert? Thanks.
2023-12-14T20:54:15Z
4,983
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-16T09:15:10Z
https://api.github.com/repos/huggingface/datasets/issues/4983/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4983/timeline
How to convert torch.utils.data.Dataset to huggingface dataset?
https://api.github.com/repos/huggingface/datasets/issues/4983/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/77595952?v=4", "events_url": "https://api.github.com/users/DEROOCE/events{/privacy}", "followers_url": "https://api.github.com/users/DEROOCE/followers", "following_url": "https://api.github.com/users/DEROOCE/following{/other_user}", "gists_url": "https://api.github.com/users/DEROOCE/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/DEROOCE", "id": 77595952, "login": "DEROOCE", "node_id": "MDQ6VXNlcjc3NTk1OTUy", "organizations_url": "https://api.github.com/users/DEROOCE/orgs", "received_events_url": "https://api.github.com/users/DEROOCE/received_events", "repos_url": "https://api.github.com/users/DEROOCE/repos", "site_admin": false, "starred_url": "https://api.github.com/users/DEROOCE/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DEROOCE/subscriptions", "type": "User", "url": "https://api.github.com/users/DEROOCE" }
[]
null
completed
NONE
2022-09-20T11:23:43Z
null
I_kwDODunzps5R_wXG
[ "Hi! I think you can use the newly-added `from_generator` method for that:\r\n```python\r\nfrom datasets import Dataset\r\n\r\ndef gen():\r\n for idx in len(torch_dataset):\r\n yield torch_dataset[idx] # this has to be a dictionary\r\n ## or if it's an IterableDataset\r\n # for ex in torch_dataset:...
{ "+1": 6, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 6, "url": "https://api.github.com/repos/huggingface/datasets/issues/4983/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4983
https://github.com/huggingface/datasets/issues/4983
false
1,375,604,693
https://api.github.com/repos/huggingface/datasets/issues/4982/labels{/name}
The problem is described in that [issue](https://github.com/huggingface/datasets/issues/4895#issuecomment-1247975569). > When I try to create data_infos.json using datasets-cli test Peter.py --save_infos --all_configs I get an error: > ValueError: Unknown split "test". Should be one of ['train']. > > The data_infos.json is created perfectly fine when I use only one split - datasets.Split.TRAIN > > You can find the code here: https://huggingface.co/datasets/sberbank-ai/Peter/tree/add_splits (add_splits branch) I tried to clear the cache folder, than I got an another error. I run: ``` git clone https://huggingface.co/datasets/sberbank-ai/Peter cd Peter git checkout add_splits # switch to a add_splits branch rm dataset_infos.json # remove local dataset_infos.json rm -r ~/.cache/huggingface # remove cached dataset_infos.json datasets-cli test Peter.py --save_infos --all_configs # trying to create new dataset_infos.json ``` The error message: ``` Using custom data configuration default Testing builder 'default' (1/1) Downloading and preparing dataset peter/default to /Users/kalinin/.cache/huggingface/datasets/peter/default/0.0.0/ef579519e140d6a40df2555996f26165f04c47557d7373709c8d7e7b4fd7465d... Downloading data files: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████| 4/4 [00:00<00:00, 5160.63it/s] Extracting data files: 0%| | 0/4 [00:00<?, ?it/s]Traceback (most recent call last): File "/usr/local/bin/datasets-cli", line 8, in <module> sys.exit(main()) File "/usr/local/lib/python3.9/site-packages/datasets/commands/datasets_cli.py", line 39, in main service.run() File "/usr/local/lib/python3.9/site-packages/datasets/commands/test.py", line 137, in run builder.download_and_prepare( File "/usr/local/lib/python3.9/site-packages/datasets/builder.py", line 704, in download_and_prepare self._download_and_prepare( File "/usr/local/lib/python3.9/site-packages/datasets/builder.py", line 1227, in _download_and_prepare super()._download_and_prepare(dl_manager, verify_infos, check_duplicate_keys=verify_infos) File "/usr/local/lib/python3.9/site-packages/datasets/builder.py", line 771, in _download_and_prepare split_generators = self._split_generators(dl_manager, **split_generators_kwargs) File "/Users/kalinin/.cache/huggingface/modules/datasets_modules/datasets/Peter/ef579519e140d6a40df2555996f26165f04c47557d7373709c8d7e7b4fd7465d/Peter.py", line 23, in _split_generators data_files = dl_manager.download_and_extract(_URLS) File "/usr/local/lib/python3.9/site-packages/datasets/download/download_manager.py", line 431, in download_and_extract return self.extract(self.download(url_or_urls)) File "/usr/local/lib/python3.9/site-packages/datasets/download/download_manager.py", line 403, in extract extracted_paths = map_nested( File "/usr/local/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 393, in map_nested mapped = [ File "/usr/local/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 394, in <listcomp> _single_map_nested((function, obj, types, None, True, None)) File "/usr/local/lib/python3.9/site-packages/datasets/utils/py_utils.py", line 330, in _single_map_nested return function(data_struct) File "/usr/local/lib/python3.9/site-packages/datasets/utils/file_utils.py", line 213, in cached_path output_path = ExtractManager(cache_dir=download_config.cache_dir).extract( File "/usr/local/lib/python3.9/site-packages/datasets/utils/extract.py", line 46, in extract self.extractor.extract(input_path, output_path, extractor_format) File "/usr/local/lib/python3.9/site-packages/datasets/utils/extract.py", line 263, in extract with FileLock(lock_path): File "/usr/local/lib/python3.9/site-packages/datasets/utils/filelock.py", line 399, in __init__ max_filename_length = os.statvfs(os.path.dirname(lock_file)).f_namemax FileNotFoundError: [Errno 2] No such file or directory: '' Exception ignored in: <function BaseFileLock.__del__ at 0x11caeec10> Traceback (most recent call last): File "/usr/local/lib/python3.9/site-packages/datasets/utils/filelock.py", line 328, in __del__ self.release(force=True) File "/usr/local/lib/python3.9/site-packages/datasets/utils/filelock.py", line 303, in release with self._thread_lock: AttributeError: 'UnixFileLock' object has no attribute '_thread_lock' Extracting data files: 0%| | 0/4 [00:00<?, ?it/s] ``` Can you help me please? ## Environment info - `datasets` version: 2.4.0 - Platform: macOS-12.5.1-x86_64-i386-64bit - Python version: 3.9.5 - PyArrow version: 9.0.0 - Pandas version: 1.2.4
2022-09-28T07:59:39Z
4,982
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-16T08:21:19Z
https://api.github.com/repos/huggingface/datasets/issues/4982/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4982/timeline
Create dataset_infos.json with VALIDATION and TEST splits
https://api.github.com/repos/huggingface/datasets/issues/4982/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/26695348?v=4", "events_url": "https://api.github.com/users/skalinin/events{/privacy}", "followers_url": "https://api.github.com/users/skalinin/followers", "following_url": "https://api.github.com/users/skalinin/following{/other_user}", "gists_url": "https://api.github.com/users/skalinin/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/skalinin", "id": 26695348, "login": "skalinin", "node_id": "MDQ6VXNlcjI2Njk1MzQ4", "organizations_url": "https://api.github.com/users/skalinin/orgs", "received_events_url": "https://api.github.com/users/skalinin/received_events", "repos_url": "https://api.github.com/users/skalinin/repos", "site_admin": false, "starred_url": "https://api.github.com/users/skalinin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/skalinin/subscriptions", "type": "User", "url": "https://api.github.com/users/skalinin" }
[]
null
completed
NONE
2022-09-28T07:59:39Z
null
I_kwDODunzps5R_g_V
[ "@mariosasko could you help me with this issue? we've started the discussion from [here](https://github.com/huggingface/datasets/issues/4895#issuecomment-1248227130)", "Hi again! Can you please pass the directory name containing the dataset script instead of the script name to `datasets-cli test`?", "Yes, it wo...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4982/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4982
https://github.com/huggingface/datasets/issues/4982
false
1,375,086,773
https://api.github.com/repos/huggingface/datasets/issues/4981/labels{/name}
## Describe the bug I can't create a dataset with `float16` features. I understand from the traceback that this is a `pyarrow` error, but I don't see anywhere in the `datasets` documentation about how to successfully do this. Is it actually supported? I've tried older versions of `pyarrow` as well with the same exact error. The bug seems to arise from `datasets` casting the values to `double` and then `pyarrow` doesn't know how to convert those back to `float16`... does that sound right? Is there a way to bypass this since it's not necessary in the `numpy` and `torch` cases? Thanks! ## Steps to reproduce the bug All of the following raise the following error with the same exact (as far as I can tell) traceback: ```python ArrowNotImplementedError: Unsupported cast from double to halffloat using function cast_half_float ``` ```python from datasets import Dataset, Features, Value Dataset.from_dict({"x": [0.0, 1.0, 2.0]}, features=Features(x=Value("float16"))) import numpy as np Dataset.from_dict({"x": np.arange(3, dtype=np.float16)}, features=Features(x=Value("float16"))) import torch Dataset.from_dict({"x": torch.arange(3).to(torch.float16)}, features=Features(x=Value("float16"))) ``` ## Expected results A dataset with `float16` features is successfully created. ## Actual results ```python --------------------------------------------------------------------------- ArrowNotImplementedError Traceback (most recent call last) Cell In [14], line 1 ----> 1 Dataset.from_dict({"x": [1.0, 2.0, 3.0]}, features=Features(x=Value("float16"))) File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/datasets/arrow_dataset.py:870, in Dataset.from_dict(cls, mapping, features, info, split) 865 mapping = features.encode_batch(mapping) 866 mapping = { 867 col: OptimizedTypedSequence(data, type=features[col] if features is not None else None, col=col) 868 for col, data in mapping.items() 869 } --> 870 pa_table = InMemoryTable.from_pydict(mapping=mapping) 871 if info.features is None: 872 info.features = Features({col: ts.get_inferred_type() for col, ts in mapping.items()}) File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/datasets/table.py:750, in InMemoryTable.from_pydict(cls, *args, **kwargs) 734 @classmethod 735 def from_pydict(cls, *args, **kwargs): 736 """ 737 Construct a Table from Arrow arrays or columns 738 (...) 748 :class:`datasets.table.Table`: 749 """ --> 750 return cls(pa.Table.from_pydict(*args, **kwargs)) File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/table.pxi:3648, in pyarrow.lib.Table.from_pydict() File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/table.pxi:5174, in pyarrow.lib._from_pydict() File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/array.pxi:343, in pyarrow.lib.asarray() File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/array.pxi:231, in pyarrow.lib.array() File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/array.pxi:110, in pyarrow.lib._handle_arrow_array_protocol() File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py:197, in TypedSequence.__arrow_array__(self, type) 192 # otherwise we can finally use the user's type 193 elif type is not None: 194 # We use cast_array_to_feature to support casting to custom types like Audio and Image 195 # Also, when trying type "string", we don't want to convert integers or floats to "string". 196 # We only do it if trying_type is False - since this is what the user asks for. --> 197 out = cast_array_to_feature(out, type, allow_number_to_str=not self.trying_type) 198 return out 199 except (TypeError, pa.lib.ArrowInvalid) as e: # handle type errors and overflows File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/datasets/table.py:1683, in _wrap_for_chunked_arrays.<locals>.wrapper(array, *args, **kwargs) 1681 return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) 1682 else: -> 1683 return func(array, *args, **kwargs) File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/datasets/table.py:1853, in cast_array_to_feature(array, feature, allow_number_to_str) 1851 return array_cast(array, get_nested_type(feature), allow_number_to_str=allow_number_to_str) 1852 elif not isinstance(feature, (Sequence, dict, list, tuple)): -> 1853 return array_cast(array, feature(), allow_number_to_str=allow_number_to_str) 1854 raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{feature}") File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/datasets/table.py:1683, in _wrap_for_chunked_arrays.<locals>.wrapper(array, *args, **kwargs) 1681 return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks]) 1682 else: -> 1683 return func(array, *args, **kwargs) File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/datasets/table.py:1762, in array_cast(array, pa_type, allow_number_to_str) 1760 if pa.types.is_null(pa_type) and not pa.types.is_null(array.type): 1761 raise TypeError(f"Couldn't cast array of type {array.type} to {pa_type}") -> 1762 return array.cast(pa_type) 1763 raise TypeError(f"Couldn't cast array of type\n{array.type}\nto\n{pa_type}") File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/array.pxi:919, in pyarrow.lib.Array.cast() File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/compute.py:389, in cast(arr, target_type, safe, options) 387 else: 388 options = CastOptions.safe(target_type) --> 389 return call_function("cast", [arr], options) File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/_compute.pyx:560, in pyarrow._compute.call_function() File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/_compute.pyx:355, in pyarrow._compute.Function.call() File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/error.pxi:144, in pyarrow.lib.pyarrow_internal_check_status() File ~/scratch/scratch-env-39/.venv/lib/python3.9/site-packages/pyarrow/error.pxi:121, in pyarrow.lib.check_status() ArrowNotImplementedError: Unsupported cast from double to halffloat using function cast_half_float ``` ## Environment info - `datasets` version: 2.4.0 - Platform: macOS-12.5.1-arm64-arm-64bit - Python version: 3.9.13 - PyArrow version: 9.0.0 - Pandas version: 1.4.4
2023-03-22T21:40:09Z
4,981
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-15T21:03:24Z
https://api.github.com/repos/huggingface/datasets/issues/4981/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4981/timeline
Can't create a dataset with `float16` features
https://api.github.com/repos/huggingface/datasets/issues/4981/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/15098095?v=4", "events_url": "https://api.github.com/users/dconathan/events{/privacy}", "followers_url": "https://api.github.com/users/dconathan/followers", "following_url": "https://api.github.com/users/dconathan/following{/other_user}", "gists_url": "https://api.github.com/users/dconathan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/dconathan", "id": 15098095, "login": "dconathan", "node_id": "MDQ6VXNlcjE1MDk4MDk1", "organizations_url": "https://api.github.com/users/dconathan/orgs", "received_events_url": "https://api.github.com/users/dconathan/received_events", "repos_url": "https://api.github.com/users/dconathan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/dconathan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dconathan/subscriptions", "type": "User", "url": "https://api.github.com/users/dconathan" }
[]
null
null
CONTRIBUTOR
null
null
I_kwDODunzps5R9ii1
[ "Hi @dconathan, thanks for reporting.\r\n\r\nWe rely on Arrow as a backend, and as far as I know currently support for `float16` in Arrow is not fully implemented in Python (C++), hence the `ArrowNotImplementedError` you get.\r\n\r\nSee, e.g.: https://arrow.apache.org/docs/status.html?highlight=float16#data-types",...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4981/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4981
https://github.com/huggingface/datasets/issues/4981
false
1,374,868,083
https://api.github.com/repos/huggingface/datasets/issues/4980/labels{/name}
**Is your feature request related to a problem? Please describe.** Is `pyarrow` really needed for every dataset? **Describe the solution you'd like** It is made optional. **Describe alternatives you've considered** Likely, no.
2022-09-16T17:23:47Z
4,980
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-15T17:38:03Z
https://api.github.com/repos/huggingface/datasets/issues/4980/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4980/timeline
Make `pyarrow` optional
https://api.github.com/repos/huggingface/datasets/issues/4980/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/240344?v=4", "events_url": "https://api.github.com/users/KOLANICH/events{/privacy}", "followers_url": "https://api.github.com/users/KOLANICH/followers", "following_url": "https://api.github.com/users/KOLANICH/following{/other_user}", "gists_url": "https://api.github.com/users/KOLANICH/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/KOLANICH", "id": 240344, "login": "KOLANICH", "node_id": "MDQ6VXNlcjI0MDM0NA==", "organizations_url": "https://api.github.com/users/KOLANICH/orgs", "received_events_url": "https://api.github.com/users/KOLANICH/received_events", "repos_url": "https://api.github.com/users/KOLANICH/repos", "site_admin": false, "starred_url": "https://api.github.com/users/KOLANICH/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/KOLANICH/subscriptions", "type": "User", "url": "https://api.github.com/users/KOLANICH" }
[]
null
completed
NONE
2022-09-16T17:23:47Z
null
I_kwDODunzps5R8tJz
[ "The whole datasets library is pretty much a wrapper to pyarrow (just take a look at some of the source for a Dataset) https://github.com/huggingface/datasets/blob/51aef08ad7053c0bfe8f9a961207b26df15850d3/src/datasets/arrow_dataset.py#L639 \r\n\r\nI think removing the pyarrow dependency would involve a complete rew...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4980/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4980
https://github.com/huggingface/datasets/issues/4980
false
1,374,820,758
https://api.github.com/repos/huggingface/datasets/issues/4979/labels{/name}
Fix missing tags in dataset cards: - amazon_us_reviews - art - discofuse - indic_glue - ubuntu_dialogs_corpus This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833 - #4891 - #4896 - #4908 - #4921 - #4931
2022-09-22T12:37:55Z
4,979
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-15T16:51:03Z
https://api.github.com/repos/huggingface/datasets/issues/4979/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4979/timeline
Fix missing tags in dataset cards
https://api.github.com/repos/huggingface/datasets/issues/4979/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-15T17:12:09Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4979.diff", "html_url": "https://github.com/huggingface/datasets/pull/4979", "merged_at": "2022-09-15T17:12:09Z", "patch_url": "https://github.com/huggingface/datasets/pull/4979.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4979" }
PR_kwDODunzps4_CouM
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4979/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4979
https://github.com/huggingface/datasets/pull/4979
true
1,374,271,504
https://api.github.com/repos/huggingface/datasets/issues/4978/labels{/name}
null
2022-09-15T22:00:20Z
4,978
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-15T10:05:57Z
https://api.github.com/repos/huggingface/datasets/issues/4978/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4978/timeline
Update IndicGLUE download links
https://api.github.com/repos/huggingface/datasets/issues/4978/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/28291870?v=4", "events_url": "https://api.github.com/users/sumanthd17/events{/privacy}", "followers_url": "https://api.github.com/users/sumanthd17/followers", "following_url": "https://api.github.com/users/sumanthd17/following{/other_user}", "gists_url": "https://api.github.com/users/sumanthd17/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sumanthd17", "id": 28291870, "login": "sumanthd17", "node_id": "MDQ6VXNlcjI4MjkxODcw", "organizations_url": "https://api.github.com/users/sumanthd17/orgs", "received_events_url": "https://api.github.com/users/sumanthd17/received_events", "repos_url": "https://api.github.com/users/sumanthd17/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sumanthd17/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sumanthd17/subscriptions", "type": "User", "url": "https://api.github.com/users/sumanthd17" }
[]
null
null
CONTRIBUTOR
2022-09-15T21:57:34Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4978.diff", "html_url": "https://github.com/huggingface/datasets/pull/4978", "merged_at": "2022-09-15T21:57:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/4978.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4978" }
PR_kwDODunzps4_Axnh
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4978/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4978
https://github.com/huggingface/datasets/pull/4978
true
1,372,962,157
https://api.github.com/repos/huggingface/datasets/issues/4977/labels{/name}
**Is your feature request related to a problem? Please describe.** Especially for big datasets like [LAION](https://huggingface.co/datasets/laion/laion2B-en/), it's hard to know exactly the downloaded size (because there are many files and you don't have their exact size when downloaded). **Describe the solution you'd like** Auto-populating the downloaded dataset size on the dataset page would be really useful, including that of each split (when there are some). **Describe alternatives you've considered** People should be adding this to dataset cards, but I don't think that is systematically the case :slightly_smiling_face: **Additional context** Mentioned to @lhoestq
2022-09-15T16:03:58Z
4,977
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-14T13:09:27Z
https://api.github.com/repos/huggingface/datasets/issues/4977/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4977/timeline
Providing dataset size
https://api.github.com/repos/huggingface/datasets/issues/4977/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/14205986?v=4", "events_url": "https://api.github.com/users/sashavor/events{/privacy}", "followers_url": "https://api.github.com/users/sashavor/followers", "following_url": "https://api.github.com/users/sashavor/following{/other_user}", "gists_url": "https://api.github.com/users/sashavor/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sashavor", "id": 14205986, "login": "sashavor", "node_id": "MDQ6VXNlcjE0MjA1OTg2", "organizations_url": "https://api.github.com/users/sashavor/orgs", "received_events_url": "https://api.github.com/users/sashavor/received_events", "repos_url": "https://api.github.com/users/sashavor/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sashavor/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sashavor/subscriptions", "type": "User", "url": "https://api.github.com/users/sashavor" }
[]
null
null
NONE
null
null
I_kwDODunzps5R1b1t
[ "Hi @sashavor, thanks for your suggestion.\r\n\r\nUntil now we have the CLI command \r\n```\r\ndatasets-cli test datasets/<your-dataset-folder> --save_infos --all_configs\r\n```\r\nthat generates the `dataset_infos.json` with the size of the downloaded dataset, among other information.\r\n\r\nWe are currently in th...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4977/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4977
https://github.com/huggingface/datasets/issues/4977
false
1,372,322,382
https://api.github.com/repos/huggingface/datasets/issues/4976/labels{/name}
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. **Describe the solution you'd like** A clear and concise description of what you want to happen. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context about the feature request here.
2022-09-26T16:32:35Z
4,976
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-14T04:42:22Z
https://api.github.com/repos/huggingface/datasets/issues/4976/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4976/timeline
Hope to adapt Python3.9 as soon as possible
https://api.github.com/repos/huggingface/datasets/issues/4976/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/74012141?v=4", "events_url": "https://api.github.com/users/RedHeartSecretMan/events{/privacy}", "followers_url": "https://api.github.com/users/RedHeartSecretMan/followers", "following_url": "https://api.github.com/users/RedHeartSecretMan/following{/other_user}", "gists_url": "https://api.github.com/users/RedHeartSecretMan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/RedHeartSecretMan", "id": 74012141, "login": "RedHeartSecretMan", "node_id": "MDQ6VXNlcjc0MDEyMTQx", "organizations_url": "https://api.github.com/users/RedHeartSecretMan/orgs", "received_events_url": "https://api.github.com/users/RedHeartSecretMan/received_events", "repos_url": "https://api.github.com/users/RedHeartSecretMan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/RedHeartSecretMan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/RedHeartSecretMan/subscriptions", "type": "User", "url": "https://api.github.com/users/RedHeartSecretMan" }
[]
null
null
NONE
null
null
I_kwDODunzps5Ry_pO
[ "Hi! `datasets` should work in Python 3.9. What kind of issue have you encountered?", "There is this related issue already: https://github.com/huggingface/datasets/issues/4113\r\nAnd I guess we need a CI job for 3.9 ^^", "Perhaps we should report this issue in the `filelock` repo?" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4976/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4976
https://github.com/huggingface/datasets/issues/4976
false
1,371,703,691
https://api.github.com/repos/huggingface/datasets/issues/4975/labels{/name}
Add the `fn_kwargs` parameter to `IterableDataset.map`. ("Resolves" https://discuss.huggingface.co/t/how-to-use-large-image-text-datasets-in-hugging-face-hub-without-downloading-for-free/22780/3)
2023-05-05T16:53:43Z
4,975
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-13T16:19:05Z
https://api.github.com/repos/huggingface/datasets/issues/4975/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4975/timeline
Add `fn_kwargs` param to `IterableDataset.map`
https://api.github.com/repos/huggingface/datasets/issues/4975/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-09-13T16:45:34Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4975.diff", "html_url": "https://github.com/huggingface/datasets/pull/4975", "merged_at": "2022-09-13T16:45:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/4975.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4975" }
PR_kwDODunzps4-4NXX
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thank you for adding this fix! \r\n\r\nWould it be possible to get `fn_kwargs` added to `IterableDatasetDict.map` as well? It looks like a very similar problem, and hopefully shouldn't be a huge change. \r\n", "Hi @brianhill11! htt...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4975/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4975
https://github.com/huggingface/datasets/pull/4975
true
1,371,682,020
https://api.github.com/repos/huggingface/datasets/issues/4974/labels{/name}
Now that all the datasets live on the Hub we can remove the /datasets directory that contains all the dataset scripts of this repository - [x] Needs https://github.com/huggingface/datasets/pull/4973 to be merged first - [x] and PR to be enabled on the Hub for non-namespaced datasets
2022-10-03T17:09:39Z
4,974
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-13T16:01:12Z
https://api.github.com/repos/huggingface/datasets/issues/4974/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4974/timeline
[GH->HF] Part 2: Remove all dataset scripts from github
https://api.github.com/repos/huggingface/datasets/issues/4974/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-10-03T17:07:32Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4974.diff", "html_url": "https://github.com/huggingface/datasets/pull/4974", "merged_at": "2022-10-03T17:07:32Z", "patch_url": "https://github.com/huggingface/datasets/pull/4974.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4974" }
PR_kwDODunzps4-4Iri
[ "_The documentation is not available anymore as the PR was closed or merged._", "So this means metrics will be deleted from this repo in favor of the \"evaluate\" library? Maybe you guys could just redirect metrics to that library.", "We are deprecating the metrics in `datasets` indeed and suggest users to swit...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4974/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4974
https://github.com/huggingface/datasets/pull/4974
true
1,371,600,074
https://api.github.com/repos/huggingface/datasets/issues/4973/labels{/name}
Currently datasets with no namespace (e.g. squad, glue) are loaded from github. In this PR I changed this logic to use the Hugging Face Hub instead. This is the first step in removing all the dataset scripts in this repository related to discussions in https://github.com/huggingface/datasets/pull/4059 (I should have continued from this PR actually)
2023-09-24T10:06:02Z
4,973
null
https://api.github.com/repos/huggingface/datasets
true
[]
2022-09-13T15:01:41Z
https://api.github.com/repos/huggingface/datasets/issues/4973/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4973/timeline
[GH->HF] Load datasets from the Hub
https://api.github.com/repos/huggingface/datasets/issues/4973/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-15T15:24:26Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4973.diff", "html_url": "https://github.com/huggingface/datasets/pull/4973", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4973.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4973" }
PR_kwDODunzps4-33JW
[ "_The documentation is not available anymore as the PR was closed or merged._", "Duplicate of:\r\n- #4059" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 1, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4973/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4973
https://github.com/huggingface/datasets/pull/4973
true
1,371,443,306
https://api.github.com/repos/huggingface/datasets/issues/4972/labels{/name}
Reported in https://discuss.huggingface.co/t/typeerror-when-applying-map-after-set-format-type-torch/23067/2 Currently it fails if one uses batched `map` and the map function returns a torch tensor. I fixed it for torch, tf, jax and pandas series.
2022-09-20T09:42:02Z
4,972
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-13T13:16:34Z
https://api.github.com/repos/huggingface/datasets/issues/4972/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4972/timeline
Fix map batched with torch output
https://api.github.com/repos/huggingface/datasets/issues/4972/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-20T09:39:33Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4972.diff", "html_url": "https://github.com/huggingface/datasets/pull/4972", "merged_at": "2022-09-20T09:39:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/4972.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4972" }
PR_kwDODunzps4-3VVF
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4972/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4972
https://github.com/huggingface/datasets/pull/4972
true
1,370,319,516
https://api.github.com/repos/huggingface/datasets/issues/4971/labels{/name}
Currently, if the `input_columns` list in `Dataset.map` is specified, the columns not in that list are dropped after the `map` transform. This makes the behavior inconsistent with `IterableDataset.map`. (It seems this issue was introduced by mistake in https://github.com/huggingface/datasets/pull/2246) Fix https://github.com/huggingface/datasets/issues/4858
2022-09-13T13:51:08Z
4,971
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-12T18:08:24Z
https://api.github.com/repos/huggingface/datasets/issues/4971/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4971/timeline
Preserve non-`input_colums` in `Dataset.map` if `input_columns` are specified
https://api.github.com/repos/huggingface/datasets/issues/4971/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-09-13T13:48:45Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4971.diff", "html_url": "https://github.com/huggingface/datasets/pull/4971", "merged_at": "2022-09-13T13:48:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/4971.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4971" }
PR_kwDODunzps4-zk3g
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4971/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4971
https://github.com/huggingface/datasets/pull/4971
true
1,369,433,074
https://api.github.com/repos/huggingface/datasets/issues/4970/labels{/name}
Support streaming nli_tr dataset. This PR removes legacy `codecs.open` and replaces it with `open` that supports passing encoding. Fix #3186.
2022-09-12T08:45:04Z
4,970
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-12T07:48:45Z
https://api.github.com/repos/huggingface/datasets/issues/4970/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4970/timeline
Support streaming nli_tr dataset
https://api.github.com/repos/huggingface/datasets/issues/4970/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-12T08:43:08Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4970.diff", "html_url": "https://github.com/huggingface/datasets/pull/4970", "merged_at": "2022-09-12T08:43:08Z", "patch_url": "https://github.com/huggingface/datasets/pull/4970.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4970" }
PR_kwDODunzps4-wkY2
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4970/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4970
https://github.com/huggingface/datasets/pull/4970
true
1,369,334,740
https://api.github.com/repos/huggingface/datasets/issues/4969/labels{/name}
After contacting the authors of the VIVOS dataset to report that their data server is down, we have received a reply from Hieu-Thi Luong that their data is now hosted on Zenodo: https://doi.org/10.5281/zenodo.7068130 This PR updates their data URL and some metadata (homepage, citation and license). Fix #4936.
2022-09-12T07:16:15Z
4,969
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-12T06:12:34Z
https://api.github.com/repos/huggingface/datasets/issues/4969/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4969/timeline
Fix data URL and metadata of vivos dataset
https://api.github.com/repos/huggingface/datasets/issues/4969/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-12T07:14:19Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4969.diff", "html_url": "https://github.com/huggingface/datasets/pull/4969", "merged_at": "2022-09-12T07:14:19Z", "patch_url": "https://github.com/huggingface/datasets/pull/4969.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4969" }
PR_kwDODunzps4-wPOk
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4969/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4969
https://github.com/huggingface/datasets/pull/4969
true
1,369,312,877
https://api.github.com/repos/huggingface/datasets/issues/4968/labels{/name}
Support streaming `compguesswhat` dataset. Fix #3191.
2022-09-12T08:00:06Z
4,968
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-12T05:42:24Z
https://api.github.com/repos/huggingface/datasets/issues/4968/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4968/timeline
Support streaming compguesswhat dataset
https://api.github.com/repos/huggingface/datasets/issues/4968/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-12T07:58:06Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4968.diff", "html_url": "https://github.com/huggingface/datasets/pull/4968", "merged_at": "2022-09-12T07:58:06Z", "patch_url": "https://github.com/huggingface/datasets/pull/4968.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4968" }
PR_kwDODunzps4-wKkw
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4968/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4968
https://github.com/huggingface/datasets/pull/4968
true
1,369,092,452
https://api.github.com/repos/huggingface/datasets/issues/4967/labels{/name}
null
2022-09-29T10:46:21Z
4,967
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-11T23:09:16Z
https://api.github.com/repos/huggingface/datasets/issues/4967/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4967/timeline
Strip "/" in local dataset path to avoid empty dataset name error
https://api.github.com/repos/huggingface/datasets/issues/4967/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/40543?v=4", "events_url": "https://api.github.com/users/apohllo/events{/privacy}", "followers_url": "https://api.github.com/users/apohllo/followers", "following_url": "https://api.github.com/users/apohllo/following{/other_user}", "gists_url": "https://api.github.com/users/apohllo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/apohllo", "id": 40543, "login": "apohllo", "node_id": "MDQ6VXNlcjQwNTQz", "organizations_url": "https://api.github.com/users/apohllo/orgs", "received_events_url": "https://api.github.com/users/apohllo/received_events", "repos_url": "https://api.github.com/users/apohllo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/apohllo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/apohllo/subscriptions", "type": "User", "url": "https://api.github.com/users/apohllo" }
[]
null
null
CONTRIBUTOR
2022-09-12T15:30:38Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4967.diff", "html_url": "https://github.com/huggingface/datasets/pull/4967", "merged_at": "2022-09-12T15:30:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/4967.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4967" }
PR_kwDODunzps4-vbS-
[ "_The documentation is not available anymore as the PR was closed or merged._", "Cool :-)" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4967/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4967
https://github.com/huggingface/datasets/pull/4967
true
1,368,661,002
https://api.github.com/repos/huggingface/datasets/issues/4965/labels{/name}
## Describe the bug I'm trying to run `cast_column("audio", Audio())` on Apple M1 Pro, but it seems that it doesn't work. ## Steps to reproduce the bug ```python import datasets dataset = load_dataset("csv", data_files="./train.csv")["train"] dataset = dataset.map(lambda x: {"audio": str(DATA_DIR / "audio" / x["audio"])}) dataset = dataset.cast_column("audio", Audio()) dataset[0] ``` ## Expected results ``` {'audio': {'bytes': None, 'path': '/root/.cache/huggingface/datasets/downloads/extracted/f14948e0e84be638dd7943ac36518a4cf3324e8b7aa331c5ab11541518e9368c/en-US~JOINT_ACCOUNT/602ba55abb1e6d0fbce92065.wav'}, 'english_transcription': 'I would like to set up a joint account with my partner', 'intent_class': 11, 'lang_id': 4, 'path': '/root/.cache/huggingface/datasets/downloads/extracted/f14948e0e84be638dd7943ac36518a4cf3324e8b7aa331c5ab11541518e9368c/en-US~JOINT_ACCOUNT/602ba55abb1e6d0fbce92065.wav', 'transcription': 'I would like to set up a joint account with my partner'} ``` ## Actual results ````--------------------------------------------------------------------------- MemoryError Traceback (most recent call last) Input In [6], in <cell line: 1>() ----> 1 dataset[0] File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/arrow_dataset.py:2165, in Dataset.__getitem__(self, key) 2163 def __getitem__(self, key): # noqa: F811 2164 """Can be used to index columns (by string names) or rows (by integer index or iterable of indices or bools).""" -> 2165 return self._getitem( 2166 key, 2167 ) File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/arrow_dataset.py:2150, in Dataset._getitem(self, key, decoded, **kwargs) 2148 formatter = get_formatter(format_type, features=self.features, decoded=decoded, **format_kwargs) 2149 pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None) -> 2150 formatted_output = format_table( 2151 pa_subtable, key, formatter=formatter, format_columns=format_columns, output_all_columns=output_all_columns 2152 ) 2153 return formatted_output File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/formatting/formatting.py:532, in format_table(table, key, formatter, format_columns, output_all_columns) 530 python_formatter = PythonFormatter(features=None) 531 if format_columns is None: --> 532 return formatter(pa_table, query_type=query_type) 533 elif query_type == "column": 534 if key in format_columns: File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/formatting/formatting.py:281, in Formatter.__call__(self, pa_table, query_type) 279 def __call__(self, pa_table: pa.Table, query_type: str) -> Union[RowFormat, ColumnFormat, BatchFormat]: 280 if query_type == "row": --> 281 return self.format_row(pa_table) 282 elif query_type == "column": 283 return self.format_column(pa_table) File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/formatting/formatting.py:312, in PythonFormatter.format_row(self, pa_table) 310 row = self.python_arrow_extractor().extract_row(pa_table) 311 if self.decoded: --> 312 row = self.python_features_decoder.decode_row(row) 313 return row File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/formatting/formatting.py:221, in PythonFeaturesDecoder.decode_row(self, row) 220 def decode_row(self, row: dict) -> dict: --> 221 return self.features.decode_example(row) if self.features else row File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/features/features.py:1647, in Features.decode_example(self, example, token_per_repo_id) 1634 def decode_example(self, example: dict, token_per_repo_id: Optional[Dict[str, Union[str, bool, None]]] = None): 1635 """Decode example with custom feature decoding. 1636 1637 Args: (...) 1644 :obj:`dict[str, Any]` 1645 """ -> 1647 return { 1648 column_name: decode_nested_example(feature, value, token_per_repo_id=token_per_repo_id) 1649 if self._column_requires_decoding[column_name] 1650 else value 1651 for column_name, (feature, value) in zip_dict( 1652 {key: value for key, value in self.items() if key in example}, example 1653 ) 1654 } File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/features/features.py:1648, in <dictcomp>(.0) 1634 def decode_example(self, example: dict, token_per_repo_id: Optional[Dict[str, Union[str, bool, None]]] = None): 1635 """Decode example with custom feature decoding. 1636 1637 Args: (...) 1644 :obj:`dict[str, Any]` 1645 """ 1647 return { -> 1648 column_name: decode_nested_example(feature, value, token_per_repo_id=token_per_repo_id) 1649 if self._column_requires_decoding[column_name] 1650 else value 1651 for column_name, (feature, value) in zip_dict( 1652 {key: value for key, value in self.items() if key in example}, example 1653 ) 1654 } File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/features/features.py:1260, in decode_nested_example(schema, obj, token_per_repo_id) 1257 # Object with special decoding: 1258 elif isinstance(schema, (Audio, Image)): 1259 # we pass the token to read and decode files from private repositories in streaming mode -> 1260 return schema.decode_example(obj, token_per_repo_id=token_per_repo_id) if obj is not None else None 1261 return obj File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/features/audio.py:156, in Audio.decode_example(self, value, token_per_repo_id) 154 array, sampling_rate = self._decode_non_mp3_file_like(file) 155 else: --> 156 array, sampling_rate = self._decode_non_mp3_path_like(path, token_per_repo_id=token_per_repo_id) 157 return {"path": path, "array": array, "sampling_rate": sampling_rate} File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/datasets/features/audio.py:257, in Audio._decode_non_mp3_path_like(self, path, format, token_per_repo_id) 254 use_auth_token = None 256 with xopen(path, "rb", use_auth_token=use_auth_token) as f: --> 257 array, sampling_rate = librosa.load(f, sr=self.sampling_rate, mono=self.mono) 258 return array, sampling_rate File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/librosa/util/decorators.py:88, in deprecate_positional_args.<locals>._inner_deprecate_positional_args.<locals>.inner_f(*args, **kwargs) 86 extra_args = len(args) - len(all_args) 87 if extra_args <= 0: ---> 88 return f(*args, **kwargs) 90 # extra_args > 0 91 args_msg = [ 92 "{}={}".format(name, arg) 93 for name, arg in zip(kwonly_args[:extra_args], args[-extra_args:]) 94 ] File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/librosa/core/audio.py:164, in load(path, sr, mono, offset, duration, dtype, res_type) 161 else: 162 # Otherwise try soundfile first, and then fall back if necessary 163 try: --> 164 y, sr_native = __soundfile_load(path, offset, duration, dtype) 166 except RuntimeError as exc: 167 # If soundfile failed, try audioread instead 168 if isinstance(path, (str, pathlib.PurePath)): File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/librosa/core/audio.py:195, in __soundfile_load(path, offset, duration, dtype) 192 context = path 193 else: 194 # Otherwise, create the soundfile object --> 195 context = sf.SoundFile(path) 197 with context as sf_desc: 198 sr_native = sf_desc.samplerate File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/soundfile.py:629, in SoundFile.__init__(self, file, mode, samplerate, channels, subtype, endian, format, closefd) 626 self._mode = mode 627 self._info = _create_info_struct(file, mode, samplerate, channels, 628 format, subtype, endian) --> 629 self._file = self._open(file, mode_int, closefd) 630 if set(mode).issuperset('r+') and self.seekable(): 631 # Move write position to 0 (like in Python file objects) 632 self.seek(0) File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/soundfile.py:1179, in SoundFile._open(self, file, mode_int, closefd) 1177 file_ptr = _snd.sf_open_fd(file, mode_int, self._info, closefd) 1178 elif _has_virtual_io_attrs(file, mode_int): -> 1179 file_ptr = _snd.sf_open_virtual(self._init_virtual_io(file), 1180 mode_int, self._info, _ffi.NULL) 1181 else: 1182 raise TypeError("Invalid file: {0!r}".format(self.name)) File ~/miniconda3/envs/rodan/lib/python3.8/site-packages/soundfile.py:1197, in SoundFile._init_virtual_io(self, file) 1194 def _init_virtual_io(self, file): 1195 """Initialize callback functions for sf_open_virtual().""" 1196 @_ffi.callback("sf_vio_get_filelen") -> 1197 def vio_get_filelen(user_data): 1198 curr = file.tell() 1199 file.seek(0, SEEK_END) MemoryError: Cannot allocate write+execute memory for ffi.callback(). You might be running on a system that prevents this. For more information, see https://cffi.readthedocs.io/en/latest/using.html#callbacks ``` ## Environment info - `datasets` version: 2.4.0 - Platform: macOS-12.5.1-arm64-arm-64bit - Python version: 3.8.13 - PyArrow version: 9.0.0 - Pandas version: 1.4.4
2024-03-21T17:25:53Z
4,965
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-10T15:55:49Z
https://api.github.com/repos/huggingface/datasets/issues/4965/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4965/timeline
[Apple M1] MemoryError: Cannot allocate write+execute memory for ffi.callback()
https://api.github.com/repos/huggingface/datasets/issues/4965/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/35718590?v=4", "events_url": "https://api.github.com/users/hoangtnm/events{/privacy}", "followers_url": "https://api.github.com/users/hoangtnm/followers", "following_url": "https://api.github.com/users/hoangtnm/following{/other_user}", "gists_url": "https://api.github.com/users/hoangtnm/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hoangtnm", "id": 35718590, "login": "hoangtnm", "node_id": "MDQ6VXNlcjM1NzE4NTkw", "organizations_url": "https://api.github.com/users/hoangtnm/orgs", "received_events_url": "https://api.github.com/users/hoangtnm/received_events", "repos_url": "https://api.github.com/users/hoangtnm/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hoangtnm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoangtnm/subscriptions", "type": "User", "url": "https://api.github.com/users/hoangtnm" }
[]
null
completed
NONE
2023-07-21T14:45:50Z
null
I_kwDODunzps5RlBwK
[ "Hi! This seems like a bug in `soundfile`. Could you please open an issue in their repo? `soundfile` works without any issues on my M1, so I'm not sure we can help.", "Hi @mariosasko, can you share how you installed `soundfile` on your mac M1?", "Hi @hoangtnm - I upgraded to python 3.10 and it fixed the proble...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4965/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4965
https://github.com/huggingface/datasets/issues/4965
false
1,368,617,322
https://api.github.com/repos/huggingface/datasets/issues/4964/labels{/name}
## Describe the bug When trying to store `Array2D, Array3D, etc` as column values in a dataset, accessing that column (or creating depending on how you create it, see code below) will cause more than 10 fold of memory usage. ## Steps to reproduce the bug ```python from datasets import Dataset, Features, Array2D, Array3D import numpy as np column_name = "a" array_shape = (64, 64, 3) data = np.random.random((10000,) + array_shape) dataset = Dataset.from_dict({column_name: data}, features=Features({column_name: Array3D(shape=array_shape, dtype="float64")})) ``` the code above will use about 10Gb of RAM while constructing the `dataset` object. The code below will use roughly the same amount of memory (and time) when trying to actually access the data itself of that column. ```python from datasets import Dataset import numpy as np column_name = "a" array_shape = (64, 64, 3) data = np.random.random((10000,) + array_shape) dataset = Dataset.from_dict({column_name: data}) dataset[column_name] ``` ## Expected results Some memory overhead, but not like as it is now and certainly not an overhead of such runtime that is currently happening. ## Actual results Enormous memory- and runtime overhead. ## Environment info - `datasets` version: 2.3.2 - Platform: macOS-12.5.1-arm64-arm-64bit - Python version: 3.8.13 - PyArrow version: 9.0.0 - Pandas version: 1.4.4
2022-09-22T18:29:22Z
4,964
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-10T13:07:22Z
https://api.github.com/repos/huggingface/datasets/issues/4964/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4964/timeline
Column of arrays (2D+) are using unreasonably high memory
https://api.github.com/repos/huggingface/datasets/issues/4964/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/30353?v=4", "events_url": "https://api.github.com/users/vigsterkr/events{/privacy}", "followers_url": "https://api.github.com/users/vigsterkr/followers", "following_url": "https://api.github.com/users/vigsterkr/following{/other_user}", "gists_url": "https://api.github.com/users/vigsterkr/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/vigsterkr", "id": 30353, "login": "vigsterkr", "node_id": "MDQ6VXNlcjMwMzUz", "organizations_url": "https://api.github.com/users/vigsterkr/orgs", "received_events_url": "https://api.github.com/users/vigsterkr/received_events", "repos_url": "https://api.github.com/users/vigsterkr/repos", "site_admin": false, "starred_url": "https://api.github.com/users/vigsterkr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vigsterkr/subscriptions", "type": "User", "url": "https://api.github.com/users/vigsterkr" }
[]
null
null
CONTRIBUTOR
null
null
I_kwDODunzps5Rk3Fq
[ "note i have tried the same code with `datasets` version 2.4.0, the outcome is the very same as described above.", "Seems related to issues #4623 and #4802 so it would appear this issue has been around for a few months.", "Hi ! `Dataset.from_dict` keeps the data in memory. You can write on disk and reload them ...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4964/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4964
https://github.com/huggingface/datasets/issues/4964
false
1,368,201,188
https://api.github.com/repos/huggingface/datasets/issues/4963/labels{/name}
### Link https://huggingface.co/datasets/julien-c/label-studio-my-dogs ### Description <img width="1115" alt="image" src="https://user-images.githubusercontent.com/326577/189422048-7e9c390f-bea7-4521-a232-43f049ccbd1f.png"> ### Owner Yes
2022-09-20T15:40:07Z
4,963
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-09-09T18:45:33Z
https://api.github.com/repos/huggingface/datasets/issues/4963/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4963/timeline
Dataset without script does not support regular JSON data file
https://api.github.com/repos/huggingface/datasets/issues/4963/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/326577?v=4", "events_url": "https://api.github.com/users/julien-c/events{/privacy}", "followers_url": "https://api.github.com/users/julien-c/followers", "following_url": "https://api.github.com/users/julien-c/following{/other_user}", "gists_url": "https://api.github.com/users/julien-c/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/julien-c", "id": 326577, "login": "julien-c", "node_id": "MDQ6VXNlcjMyNjU3Nw==", "organizations_url": "https://api.github.com/users/julien-c/orgs", "received_events_url": "https://api.github.com/users/julien-c/received_events", "repos_url": "https://api.github.com/users/julien-c/repos", "site_admin": false, "starred_url": "https://api.github.com/users/julien-c/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/julien-c/subscriptions", "type": "User", "url": "https://api.github.com/users/julien-c" }
[]
null
completed
MEMBER
2022-09-20T15:40:07Z
null
I_kwDODunzps5RjRfk
[ "Hi @julien-c,\r\n\r\nOut of the box, we only support JSON lines (NDJSON) data files, but your data file is a regular JSON file. The reason is we use `pyarrow.json.read_json` and this only supports line-delimited JSON. " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4963/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4963
https://github.com/huggingface/datasets/issues/4963
false
1,368,155,365
https://api.github.com/repos/huggingface/datasets/issues/4962/labels{/name}
exclude broken version of fsspec. See the [related issue](https://github.com/huggingface/datasets/issues/4961)
2022-09-12T14:33:04Z
4,962
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-09T17:57:56Z
https://api.github.com/repos/huggingface/datasets/issues/4962/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4962/timeline
Update setup.py
https://api.github.com/repos/huggingface/datasets/issues/4962/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3616964?v=4", "events_url": "https://api.github.com/users/DCNemesis/events{/privacy}", "followers_url": "https://api.github.com/users/DCNemesis/followers", "following_url": "https://api.github.com/users/DCNemesis/following{/other_user}", "gists_url": "https://api.github.com/users/DCNemesis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/DCNemesis", "id": 3616964, "login": "DCNemesis", "node_id": "MDQ6VXNlcjM2MTY5NjQ=", "organizations_url": "https://api.github.com/users/DCNemesis/orgs", "received_events_url": "https://api.github.com/users/DCNemesis/received_events", "repos_url": "https://api.github.com/users/DCNemesis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/DCNemesis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DCNemesis/subscriptions", "type": "User", "url": "https://api.github.com/users/DCNemesis" }
[]
null
null
NONE
2022-09-12T14:33:04Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4962.diff", "html_url": "https://github.com/huggingface/datasets/pull/4962", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4962.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4962" }
PR_kwDODunzps4-sh-o
[ "Before addressing this PR, we should be sure about the issue. See my comment in:\r\n- https://github.com/huggingface/datasets/issues/4961#issuecomment-1243376247", "Once we know 2022.8.2 works, I'm closing this PR, as the corresponding issue." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4962/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4962
https://github.com/huggingface/datasets/pull/4962
true
1,368,124,033
https://api.github.com/repos/huggingface/datasets/issues/4961/labels{/name}
## Describe the bug When fsspec 2022.8.2 is installed in your environment, xopen will prematurely close files, making streaming mode inoperable. ## Steps to reproduce the bug ```python import datasets data = datasets.load_dataset('MLCommons/ml_spoken_words', 'id_wav', split='train', streaming=True) ``` ## Expected results Dataset should load as iterator. ## Actual results ``` [/usr/local/lib/python3.7/dist-packages/datasets/load.py](https://localhost:8080/#) in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, **config_kwargs) 1737 # Return iterable dataset in case of streaming 1738 if streaming: -> 1739 return builder_instance.as_streaming_dataset(split=split) 1740 1741 # Some datasets are already processed on the HF google storage [/usr/local/lib/python3.7/dist-packages/datasets/builder.py](https://localhost:8080/#) in as_streaming_dataset(self, split, base_path) 1023 ) 1024 self._check_manual_download(dl_manager) -> 1025 splits_generators = {sg.name: sg for sg in self._split_generators(dl_manager)} 1026 # By default, return all splits 1027 if split is None: [~/.cache/huggingface/modules/datasets_modules/datasets/MLCommons--ml_spoken_words/321ea853cf0a05abb7a2d7efea900692a3d8622af65a2f3ce98adb7800a5d57b/ml_spoken_words.py](https://localhost:8080/#) in _split_generators(self, dl_manager) 182 name=datasets.Split.TRAIN, 183 gen_kwargs={ --> 184 "audio_archives": [download_audio(split="train", lang=lang) for lang in self.config.languages], 185 "local_audio_archives_paths": [download_extract_audio(split="train", lang=lang) for lang in 186 self.config.languages] if not dl_manager.is_streaming else None, [~/.cache/huggingface/modules/datasets_modules/datasets/MLCommons--ml_spoken_words/321ea853cf0a05abb7a2d7efea900692a3d8622af65a2f3ce98adb7800a5d57b/ml_spoken_words.py](https://localhost:8080/#) in <listcomp>(.0) 182 name=datasets.Split.TRAIN, 183 gen_kwargs={ --> 184 "audio_archives": [download_audio(split="train", lang=lang) for lang in self.config.languages], 185 "local_audio_archives_paths": [download_extract_audio(split="train", lang=lang) for lang in 186 self.config.languages] if not dl_manager.is_streaming else None, [~/.cache/huggingface/modules/datasets_modules/datasets/MLCommons--ml_spoken_words/321ea853cf0a05abb7a2d7efea900692a3d8622af65a2f3ce98adb7800a5d57b/ml_spoken_words.py](https://localhost:8080/#) in _download_audio_archives(dl_manager, lang, format, split) 267 # for streaming case 268 def _download_audio_archives(dl_manager, lang, format, split): --> 269 archives_paths = _download_audio_archives_paths(dl_manager, lang, format, split) 270 return [dl_manager.iter_archive(archive_path) for archive_path in archives_paths] [~/.cache/huggingface/modules/datasets_modules/datasets/MLCommons--ml_spoken_words/321ea853cf0a05abb7a2d7efea900692a3d8622af65a2f3ce98adb7800a5d57b/ml_spoken_words.py](https://localhost:8080/#) in _download_audio_archives_paths(dl_manager, lang, format, split) 251 n_files_path = dl_manager.download(n_files_url) 252 --> 253 with open(n_files_path, "r", encoding="utf-8") as file: 254 n_files = int(file.read().strip()) # the file contains a number of archives 255 ValueError: I/O operation on closed file. ``` ## Environment info - `datasets` version: 2.4.0 - Platform: Linux-5.10.133+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.13 - PyArrow version: 6.0.1 - Pandas version: 1.3.5
2022-09-12T17:45:50Z
4,961
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-09T17:26:55Z
https://api.github.com/repos/huggingface/datasets/issues/4961/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4961/timeline
fsspec 2022.8.2 breaks xopen in streaming mode
https://api.github.com/repos/huggingface/datasets/issues/4961/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/3616964?v=4", "events_url": "https://api.github.com/users/DCNemesis/events{/privacy}", "followers_url": "https://api.github.com/users/DCNemesis/followers", "following_url": "https://api.github.com/users/DCNemesis/following{/other_user}", "gists_url": "https://api.github.com/users/DCNemesis/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/DCNemesis", "id": 3616964, "login": "DCNemesis", "node_id": "MDQ6VXNlcjM2MTY5NjQ=", "organizations_url": "https://api.github.com/users/DCNemesis/orgs", "received_events_url": "https://api.github.com/users/DCNemesis/received_events", "repos_url": "https://api.github.com/users/DCNemesis/repos", "site_admin": false, "starred_url": "https://api.github.com/users/DCNemesis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DCNemesis/subscriptions", "type": "User", "url": "https://api.github.com/users/DCNemesis" }
[]
null
completed
NONE
2022-09-12T14:32:05Z
null
I_kwDODunzps5Ri-qB
[ "loading `fsspec==2022.7.1` fixes this issue, setup.py would need to be changed to prevent users from using the latest version of fsspec.", "Opened [PR](https://github.com/huggingface/datasets/pull/4962) to address this.", "Hi @DCNemesis, thanks for reporting.\r\n\r\nThat was a temporary issue in `fsspec` relea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4961/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4961
https://github.com/huggingface/datasets/issues/4961
false
1,368,035,159
https://api.github.com/repos/huggingface/datasets/issues/4960/labels{/name}
## Describe the bug I am trying to load a dataset from drive and running into an error. ## Steps to reproduce the bug ```python data_dir = "/Users/dlituiev/repos/datasets/bioasq/BioASQ-training9b" bioasq_task_b = load_dataset("aps/bioasq_task_b", data_dir=data_dir) ``` ## Actual results `AttributeError: 'BuilderConfig' object has no attribute 'schema'` <details> ``` Using custom data configuration default-a1ca3e05be5abf2f --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) Input In [8], in <cell line: 2>() 1 data_dir = "/Users/dlituiev/repos/datasets/bioasq/BioASQ-training9b" ----> 2 bioasq_task_b = load_dataset("aps/bioasq_task_b", data_dir=data_dir) File ~/opt/anaconda3/envs/spacy3/lib/python3.10/site-packages/datasets/load.py:1723, in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, ignore_verifications, keep_in_memory, save_infos, revision, use_auth_token, task, streaming, **config_kwargs) 1720 ignore_verifications = ignore_verifications or save_infos 1722 # Create a dataset builder -> 1723 builder_instance = load_dataset_builder( 1724 path=path, 1725 name=name, 1726 data_dir=data_dir, 1727 data_files=data_files, 1728 cache_dir=cache_dir, 1729 features=features, 1730 download_config=download_config, 1731 download_mode=download_mode, 1732 revision=revision, 1733 use_auth_token=use_auth_token, 1734 **config_kwargs, 1735 ) 1737 # Return iterable dataset in case of streaming 1738 if streaming: File ~/opt/anaconda3/envs/spacy3/lib/python3.10/site-packages/datasets/load.py:1526, in load_dataset_builder(path, name, data_dir, data_files, cache_dir, features, download_config, download_mode, revision, use_auth_token, **config_kwargs) 1523 raise ValueError(error_msg) 1525 # Instantiate the dataset builder -> 1526 builder_instance: DatasetBuilder = builder_cls( 1527 cache_dir=cache_dir, 1528 config_name=config_name, 1529 data_dir=data_dir, 1530 data_files=data_files, 1531 hash=hash, 1532 features=features, 1533 use_auth_token=use_auth_token, 1534 **builder_kwargs, 1535 **config_kwargs, 1536 ) 1538 return builder_instance File ~/opt/anaconda3/envs/spacy3/lib/python3.10/site-packages/datasets/builder.py:1154, in GeneratorBasedBuilder.__init__(self, writer_batch_size, *args, **kwargs) 1153 def __init__(self, *args, writer_batch_size=None, **kwargs): -> 1154 super().__init__(*args, **kwargs) 1155 # Batch size used by the ArrowWriter 1156 # It defines the number of samples that are kept in memory before writing them 1157 # and also the length of the arrow chunks 1158 # None means that the ArrowWriter will use its default value 1159 self._writer_batch_size = writer_batch_size or self.DEFAULT_WRITER_BATCH_SIZE File ~/opt/anaconda3/envs/spacy3/lib/python3.10/site-packages/datasets/builder.py:307, in DatasetBuilder.__init__(self, cache_dir, config_name, hash, base_path, info, features, use_auth_token, repo_id, data_files, data_dir, name, **config_kwargs) 305 if info is None: 306 info = self.get_exported_dataset_info() --> 307 info.update(self._info()) 308 info.builder_name = self.name 309 info.config_name = self.config.name File ~/.cache/huggingface/modules/datasets_modules/datasets/aps--bioasq_task_b/3d54b1213f7e8001eef755af92877f9efa44161ee83c2a70d5d649defa95759e/bioasq_task_b.py:477, in BioasqTaskBDataset._info(self) 474 def _info(self): 475 476 # BioASQ Task B source schema --> 477 if self.config.schema == "source": 478 features = datasets.Features( 479 { 480 "id": datasets.Value("string"), (...) 504 } 505 ) 506 # simplified schema for QA tasks AttributeError: 'BuilderConfig' object has no attribute 'schema' ``` </details> ## Environment info - `datasets` version: 2.4.0 - Platform: macOS-10.16-x86_64-i386-64bit - Python version: 3.10.4 - PyArrow version: 9.0.0 - Pandas version: 1.4.3
2022-09-13T08:51:03Z
4,960
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
2022-09-09T16:06:43Z
https://api.github.com/repos/huggingface/datasets/issues/4960/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4960/timeline
BioASQ AttributeError: 'BuilderConfig' object has no attribute 'schema'
https://api.github.com/repos/huggingface/datasets/issues/4960/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8426290?v=4", "events_url": "https://api.github.com/users/DSLituiev/events{/privacy}", "followers_url": "https://api.github.com/users/DSLituiev/followers", "following_url": "https://api.github.com/users/DSLituiev/following{/other_user}", "gists_url": "https://api.github.com/users/DSLituiev/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/DSLituiev", "id": 8426290, "login": "DSLituiev", "node_id": "MDQ6VXNlcjg0MjYyOTA=", "organizations_url": "https://api.github.com/users/DSLituiev/orgs", "received_events_url": "https://api.github.com/users/DSLituiev/received_events", "repos_url": "https://api.github.com/users/DSLituiev/repos", "site_admin": false, "starred_url": "https://api.github.com/users/DSLituiev/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DSLituiev/subscriptions", "type": "User", "url": "https://api.github.com/users/DSLituiev" }
[]
null
null
NONE
null
null
I_kwDODunzps5Rio9X
[ "Following worked:\r\n\r\n```\r\ndata_dir = \"/Users/dlituiev/repos/datasets/bioasq/\"\r\nbioasq_task_b = load_dataset(\"aps/bioasq_task_b\", data_dir=data_dir, name=\"bioasq_9b_source\")\r\n```\r\n\r\nWould maintainers be open to one of the following:\r\n- automating this with a latest default config (e.g. `bioas...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4960/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4960
https://github.com/huggingface/datasets/issues/4960
false
1,367,924,429
https://api.github.com/repos/huggingface/datasets/issues/4959/labels{/name}
After we informed the `compguesswhat` dataset authors about an error with their data URLs, they have updated them: - https://github.com/CompGuessWhat/compguesswhat.github.io/issues/1 This PR updates their data URLs in our loading script. Related to: - #3191
2022-09-09T16:01:34Z
4,959
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-09T14:36:10Z
https://api.github.com/repos/huggingface/datasets/issues/4959/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4959/timeline
Fix data URLs of compguesswhat dataset
https://api.github.com/repos/huggingface/datasets/issues/4959/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-09T15:59:04Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4959.diff", "html_url": "https://github.com/huggingface/datasets/pull/4959", "merged_at": "2022-09-09T15:59:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/4959.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4959" }
PR_kwDODunzps4-rx6l
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4959/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4959
https://github.com/huggingface/datasets/pull/4959
true
1,367,695,376
https://api.github.com/repos/huggingface/datasets/issues/4958/labels{/name}
Hi, When I use load_dataset from local jsonl files, below error happens, and I type the link into the browser prompting me `404: Not Found`. I download the other `.py` files using the same method and it works. It seems that the server is missing the appropriate file, or it is a problem with the code version. ``` ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/2.3.0/datasets/jsonl/jsonl.py (ConnectionError(MaxRetryError("HTTPSConnectionPool(host='raw.githubusercontent.com', port=443): Max retries exceeded with url: /huggingface/datasets/2.3.0/datasets/jsonl/jsonl.py (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x2b08342004c0>: Failed to establish a new connection: [Errno 101] Network is unreachable'))"))) ```
2022-09-09T11:38:44Z
4,958
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-09T11:29:55Z
https://api.github.com/repos/huggingface/datasets/issues/4958/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4958/timeline
ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/2.4.0/datasets/jsonl/jsonl.py
https://api.github.com/repos/huggingface/datasets/issues/4958/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/66322047?v=4", "events_url": "https://api.github.com/users/hasakikiki/events{/privacy}", "followers_url": "https://api.github.com/users/hasakikiki/followers", "following_url": "https://api.github.com/users/hasakikiki/following{/other_user}", "gists_url": "https://api.github.com/users/hasakikiki/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hasakikiki", "id": 66322047, "login": "hasakikiki", "node_id": "MDQ6VXNlcjY2MzIyMDQ3", "organizations_url": "https://api.github.com/users/hasakikiki/orgs", "received_events_url": "https://api.github.com/users/hasakikiki/received_events", "repos_url": "https://api.github.com/users/hasakikiki/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hasakikiki/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hasakikiki/subscriptions", "type": "User", "url": "https://api.github.com/users/hasakikiki" }
[]
null
completed
NONE
2022-09-09T11:38:44Z
null
I_kwDODunzps5RhWAQ
[ "I have solved this problem... The extension of the file should be `.json` not `.jsonl`" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4958/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4958
https://github.com/huggingface/datasets/issues/4958
false
1,366,532,849
https://api.github.com/repos/huggingface/datasets/issues/4957/labels{/name}
Add `Dataset.from_generator` to the API to allow creating datasets from data larger than RAM. The implementation relies on a packaged module not exposed in `load_dataset` to tie this method with `datasets`' caching mechanism. Closes https://github.com/huggingface/datasets/issues/4417
2022-09-16T14:46:35Z
4,957
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-08T15:08:25Z
https://api.github.com/repos/huggingface/datasets/issues/4957/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4957/timeline
Add `Dataset.from_generator`
https://api.github.com/repos/huggingface/datasets/issues/4957/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-09-16T14:44:18Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4957.diff", "html_url": "https://github.com/huggingface/datasets/pull/4957", "merged_at": "2022-09-16T14:44:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/4957.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4957" }
PR_kwDODunzps4-nGIk
[ "I restarted the builder PR job just in case", "_The documentation is not available anymore as the PR was closed or merged._", "CI is now green. https://github.com/huggingface/doc-builder/pull/296 explains why it failed." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 2, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 2, "url": "https://api.github.com/repos/huggingface/datasets/issues/4957/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4957
https://github.com/huggingface/datasets/pull/4957
true
1,366,475,160
https://api.github.com/repos/huggingface/datasets/issues/4956/labels{/name}
Fixes #4953
2022-09-08T15:16:51Z
4,956
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-08T14:39:10Z
https://api.github.com/repos/huggingface/datasets/issues/4956/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4956/timeline
Fix TF tests for 2.10
https://api.github.com/repos/huggingface/datasets/issues/4956/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/12866554?v=4", "events_url": "https://api.github.com/users/Rocketknight1/events{/privacy}", "followers_url": "https://api.github.com/users/Rocketknight1/followers", "following_url": "https://api.github.com/users/Rocketknight1/following{/other_user}", "gists_url": "https://api.github.com/users/Rocketknight1/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Rocketknight1", "id": 12866554, "login": "Rocketknight1", "node_id": "MDQ6VXNlcjEyODY2NTU0", "organizations_url": "https://api.github.com/users/Rocketknight1/orgs", "received_events_url": "https://api.github.com/users/Rocketknight1/received_events", "repos_url": "https://api.github.com/users/Rocketknight1/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Rocketknight1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Rocketknight1/subscriptions", "type": "User", "url": "https://api.github.com/users/Rocketknight1" }
[]
null
null
MEMBER
2022-09-08T15:14:44Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4956.diff", "html_url": "https://github.com/huggingface/datasets/pull/4956", "merged_at": "2022-09-08T15:14:44Z", "patch_url": "https://github.com/huggingface/datasets/pull/4956.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4956" }
PR_kwDODunzps4-m5NU
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4956/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4956
https://github.com/huggingface/datasets/pull/4956
true
1,366,382,314
https://api.github.com/repos/huggingface/datasets/issues/4955/labels{/name}
See for example: - https://github.com/huggingface/datasets/issues/3191 - https://github.com/huggingface/datasets/issues/3186 It would help provide clearer information on the Hub and help the dataset maintainer solve the issue by themselves quicker. Currently: - https://huggingface.co/datasets/compguesswhat <img width="1029" alt="Capture d’écran 2022-09-08 à 15 51 37" src="https://user-images.githubusercontent.com/1676121/189139946-6deffb91-f21b-4281-8825-a98026c69740.png"> - https://huggingface.co/datasets/nli_tr <img width="1032" alt="Capture d’écran 2022-09-08 à 15 51 44" src="https://user-images.githubusercontent.com/1676121/189139963-d26490ed-ad23-48ea-9cfc-1ab9c4d08d0c.png"> cc @albertvillanova
2022-09-08T13:53:36Z
4,955
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-09-08T13:52:37Z
https://api.github.com/repos/huggingface/datasets/issues/4955/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4955/timeline
Raise a more precise error when the URL is unreachable in streaming mode
https://api.github.com/repos/huggingface/datasets/issues/4955/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[]
null
null
CONTRIBUTOR
null
null
I_kwDODunzps5RcVbq
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4955/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4955
https://github.com/huggingface/datasets/issues/4955
false
1,366,369,682
https://api.github.com/repos/huggingface/datasets/issues/4954/labels{/name}
Temporarily fix TensorFlow until a permanent solution is found. Related to: - #4953
2022-09-08T14:12:33Z
4,954
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-08T13:46:15Z
https://api.github.com/repos/huggingface/datasets/issues/4954/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4954/timeline
Pin TensorFlow temporarily
https://api.github.com/repos/huggingface/datasets/issues/4954/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-08T14:10:03Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4954.diff", "html_url": "https://github.com/huggingface/datasets/pull/4954", "merged_at": "2022-09-08T14:10:03Z", "patch_url": "https://github.com/huggingface/datasets/pull/4954.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4954" }
PR_kwDODunzps4-mhl5
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4954/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4954
https://github.com/huggingface/datasets/pull/4954
true
1,366,356,514
https://api.github.com/repos/huggingface/datasets/issues/4953/labels{/name}
## Describe the bug The following CI test fails: https://github.com/huggingface/datasets/runs/8246722693?check_suite_focus=true ``` FAILED tests/test_py_utils.py::TempSeedTest::test_tensorflow - AssertionError: ``` Details: ``` _________________________ TempSeedTest.test_tensorflow _________________________ [gw0] linux -- Python 3.7.13 /opt/hostedtoolcache/Python/3.7.13/x64/bin/python self = <tests.test_py_utils.TempSeedTest testMethod=test_tensorflow> @require_tf def test_tensorflow(self): import tensorflow as tf from tensorflow.keras import layers def gen_random_output(): model = layers.Dense(2) x = tf.random.uniform((1, 3)) return model(x).numpy() with temp_seed(42, set_tensorflow=True): out1 = gen_random_output() with temp_seed(42, set_tensorflow=True): out2 = gen_random_output() out3 = gen_random_output() > np.testing.assert_equal(out1, out2) E AssertionError: E Arrays are not equal E E Mismatched elements: 2 / 2 (100%) E Max absolute difference: 0.84619296 E Max relative difference: 16.083529 E x: array([[-0.793581, 0.333286]], dtype=float32) E y: array([[0.052612, 0.539708]], dtype=float32) tests/test_py_utils.py:149: AssertionError ```
2022-09-08T15:14:45Z
4,953
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-08T13:39:29Z
https://api.github.com/repos/huggingface/datasets/issues/4953/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4953/timeline
CI test of TensorFlow is failing
https://api.github.com/repos/huggingface/datasets/issues/4953/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
completed
MEMBER
2022-09-08T15:14:45Z
null
I_kwDODunzps5RcPIi
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4953/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4953
https://github.com/huggingface/datasets/issues/4953
false
1,366,354,604
https://api.github.com/repos/huggingface/datasets/issues/4952/labels{/name}
To avoid having too many conflicts in the datasets and metrics dependencies I split the CI into test and test-catalog test does the test of the core of the `datasets` lib, while test-catalog tests the datasets scripts and metrics scripts This also makes `pip install -e .[dev]` much smaller for developers WDYT @albertvillanova ?
2023-09-24T10:05:57Z
4,952
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-08T13:38:30Z
https://api.github.com/repos/huggingface/datasets/issues/4952/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4952/timeline
Add test-datasets CI job
https://api.github.com/repos/huggingface/datasets/issues/4952/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-16T13:25:48Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4952.diff", "html_url": "https://github.com/huggingface/datasets/pull/4952", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4952.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4952" }
PR_kwDODunzps4-meM0
[ "_The documentation is not available anymore as the PR was closed or merged._", "Closing this one since the dataset scripts will be removed in https://github.com/huggingface/datasets/pull/4974" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4952/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4952
https://github.com/huggingface/datasets/pull/4952
true
1,365,954,814
https://api.github.com/repos/huggingface/datasets/issues/4951/labels{/name}
This PR adds the license information to `qasc` dataset, once reported via GitHub by Tushar Khot, the dataset is licensed under CC BY 4.0: - https://github.com/allenai/qasc/issues/5
2022-09-08T14:54:47Z
4,951
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-08T10:04:39Z
https://api.github.com/repos/huggingface/datasets/issues/4951/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4951/timeline
Fix license information in qasc dataset card
https://api.github.com/repos/huggingface/datasets/issues/4951/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-08T14:52:05Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4951.diff", "html_url": "https://github.com/huggingface/datasets/pull/4951", "merged_at": "2022-09-08T14:52:05Z", "patch_url": "https://github.com/huggingface/datasets/pull/4951.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4951" }
PR_kwDODunzps4-lDqd
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4951/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4951
https://github.com/huggingface/datasets/pull/4951
true
1,365,458,633
https://api.github.com/repos/huggingface/datasets/issues/4950/labels{/name}
The current enwik8 dataset link give a 502 bad gateway error which can be view on https://huggingface.co/datasets/enwik8 (click the dropdown to see the dataset preview, it will show the error). This corrects the links, and json metadata as well as adds a little bit more information about enwik8.
2022-09-24T22:14:35Z
4,950
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-08T03:15:00Z
https://api.github.com/repos/huggingface/datasets/issues/4950/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4950/timeline
Update Enwik8 broken link and information
https://api.github.com/repos/huggingface/datasets/issues/4950/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/54819091?v=4", "events_url": "https://api.github.com/users/mtanghu/events{/privacy}", "followers_url": "https://api.github.com/users/mtanghu/followers", "following_url": "https://api.github.com/users/mtanghu/following{/other_user}", "gists_url": "https://api.github.com/users/mtanghu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mtanghu", "id": 54819091, "login": "mtanghu", "node_id": "MDQ6VXNlcjU0ODE5MDkx", "organizations_url": "https://api.github.com/users/mtanghu/orgs", "received_events_url": "https://api.github.com/users/mtanghu/received_events", "repos_url": "https://api.github.com/users/mtanghu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mtanghu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mtanghu/subscriptions", "type": "User", "url": "https://api.github.com/users/mtanghu" }
[]
null
null
CONTRIBUTOR
2022-09-08T14:51:00Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4950.diff", "html_url": "https://github.com/huggingface/datasets/pull/4950", "merged_at": "2022-09-08T14:51:00Z", "patch_url": "https://github.com/huggingface/datasets/pull/4950.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4950" }
PR_kwDODunzps4-jWZ1
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4950/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4950
https://github.com/huggingface/datasets/pull/4950
true
1,365,251,916
https://api.github.com/repos/huggingface/datasets/issues/4949/labels{/name}
The current enwik8 dataset link give a 502 bad gateway error which can be view on https://huggingface.co/datasets/enwik8 (click the dropdown to see the dataset preview, it will show the error). This corrects the links, and json metadata as well as adds a little bit more information about enwik8.
2022-09-08T03:14:04Z
4,949
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-07T22:17:14Z
https://api.github.com/repos/huggingface/datasets/issues/4949/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4949/timeline
Update enwik8 fixing the broken link
https://api.github.com/repos/huggingface/datasets/issues/4949/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/54819091?v=4", "events_url": "https://api.github.com/users/mtanghu/events{/privacy}", "followers_url": "https://api.github.com/users/mtanghu/followers", "following_url": "https://api.github.com/users/mtanghu/following{/other_user}", "gists_url": "https://api.github.com/users/mtanghu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mtanghu", "id": 54819091, "login": "mtanghu", "node_id": "MDQ6VXNlcjU0ODE5MDkx", "organizations_url": "https://api.github.com/users/mtanghu/orgs", "received_events_url": "https://api.github.com/users/mtanghu/received_events", "repos_url": "https://api.github.com/users/mtanghu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mtanghu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mtanghu/subscriptions", "type": "User", "url": "https://api.github.com/users/mtanghu" }
[]
null
null
CONTRIBUTOR
2022-09-08T03:14:04Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4949.diff", "html_url": "https://github.com/huggingface/datasets/pull/4949", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4949.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4949" }
PR_kwDODunzps4-iqzI
[ "Closing pull request to following contributing guidelines of making a new branch and will make a new pull request" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4949/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4949
https://github.com/huggingface/datasets/pull/4949
true
1,364,973,778
https://api.github.com/repos/huggingface/datasets/issues/4948/labels{/name}
null
2022-09-08T14:59:31Z
4,948
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-07T17:20:51Z
https://api.github.com/repos/huggingface/datasets/issues/4948/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4948/timeline
Fix minor typo in error message for missing imports
https://api.github.com/repos/huggingface/datasets/issues/4948/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "events_url": "https://api.github.com/users/mariosasko/events{/privacy}", "followers_url": "https://api.github.com/users/mariosasko/followers", "following_url": "https://api.github.com/users/mariosasko/following{/other_user}", "gists_url": "https://api.github.com/users/mariosasko/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mariosasko", "id": 47462742, "login": "mariosasko", "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "organizations_url": "https://api.github.com/users/mariosasko/orgs", "received_events_url": "https://api.github.com/users/mariosasko/received_events", "repos_url": "https://api.github.com/users/mariosasko/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mariosasko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mariosasko/subscriptions", "type": "User", "url": "https://api.github.com/users/mariosasko" }
[]
null
null
CONTRIBUTOR
2022-09-08T14:57:15Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4948.diff", "html_url": "https://github.com/huggingface/datasets/pull/4948", "merged_at": "2022-09-08T14:57:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/4948.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4948" }
PR_kwDODunzps4-hwsl
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4948/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4948
https://github.com/huggingface/datasets/pull/4948
true
1,364,967,957
https://api.github.com/repos/huggingface/datasets/issues/4947/labels{/name}
null
2023-09-24T10:05:38Z
4,947
null
https://api.github.com/repos/huggingface/datasets
true
[]
2022-09-07T17:14:49Z
https://api.github.com/repos/huggingface/datasets/issues/4947/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4947/timeline
Try to fix the Windows CI after TF update 2.10
https://api.github.com/repos/huggingface/datasets/issues/4947/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-08T09:13:10Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4947.diff", "html_url": "https://github.com/huggingface/datasets/pull/4947", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4947.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4947" }
PR_kwDODunzps4-hvbq
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4947). All of your documentation changes will be reflected on that endpoint." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4947/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4947
https://github.com/huggingface/datasets/pull/4947
true
1,364,692,069
https://api.github.com/repos/huggingface/datasets/issues/4946/labels{/name}
Closes https://github.com/huggingface/datasets/issues/4945 by adding a regex check when pushing to hub. Let me know if this is helpful and if it's the fix you would have in mind for the issue and I'm happy to contribute tests.
2022-09-13T10:19:01Z
4,946
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-07T13:45:58Z
https://api.github.com/repos/huggingface/datasets/issues/4946/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4946/timeline
Introduce regex check when pushing as well
https://api.github.com/repos/huggingface/datasets/issues/4946/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/30755778?v=4", "events_url": "https://api.github.com/users/LysandreJik/events{/privacy}", "followers_url": "https://api.github.com/users/LysandreJik/followers", "following_url": "https://api.github.com/users/LysandreJik/following{/other_user}", "gists_url": "https://api.github.com/users/LysandreJik/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/LysandreJik", "id": 30755778, "login": "LysandreJik", "node_id": "MDQ6VXNlcjMwNzU1Nzc4", "organizations_url": "https://api.github.com/users/LysandreJik/orgs", "received_events_url": "https://api.github.com/users/LysandreJik/received_events", "repos_url": "https://api.github.com/users/LysandreJik/repos", "site_admin": false, "starred_url": "https://api.github.com/users/LysandreJik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LysandreJik/subscriptions", "type": "User", "url": "https://api.github.com/users/LysandreJik" }
[]
null
null
MEMBER
2022-09-13T10:16:34Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4946.diff", "html_url": "https://github.com/huggingface/datasets/pull/4946", "merged_at": "2022-09-13T10:16:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/4946.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4946" }
PR_kwDODunzps4-g0Hz
[ "_The documentation is not available anymore as the PR was closed or merged._", "Let me take over this PR if you don't mind" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4946/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4946
https://github.com/huggingface/datasets/pull/4946
true
1,364,691,096
https://api.github.com/repos/huggingface/datasets/issues/4945/labels{/name}
## Describe the bug The `push_to_hub` method can push splits that do not respect the regex check that is used for downloads. Therefore, splits may be pushed but never re-used, which can be painful if the split was done after runtime preprocessing. ## Steps to reproduce the bug ```python >>> from datasets import Dataset, DatasetDict, load_dataset >>> d = Dataset.from_dict({'x': [1,2,3], 'y': [1,2,3]}) >>> di = DatasetDict() >>> di['identifier-with-column'] = d >>> di.push_to_hub('open-source-metrics/test') Pushing split identifier-with-column to the Hub. Pushing dataset shards to the dataset hub: 100%|██████████| 1/1 [00:04<00:00, 4.40s/it] ``` Loading it afterwards: ```python >>> load_dataset('open-source-metrics/test') Downloading: 100%|██████████| 610/610 [00:00<00:00, 432kB/s] Using custom data configuration open-source-metrics--test-28b63ec7cde80488 Downloading and preparing dataset None/None (download: 950 bytes, generated: 48 bytes, post-processed: Unknown size, total: 998 bytes) to /home/lysandre/.cache/huggingface/datasets/open-source-metrics___parquet/open-source-metrics--test-28b63ec7cde80488/0.0.0/2a3b91fbd88a2c90d1dbbb32b460cf621d31bd5b05b934492fdef7d8d6f236ec... Downloading data files: 0%| | 0/1 [00:00<?, ?it/s] Downloading data: 100%|██████████| 950/950 [00:00<00:00, 1.01MB/s] Downloading data files: 100%|██████████| 1/1 [00:01<00:00, 1.48s/it] Extracting data files: 100%|██████████| 1/1 [00:00<00:00, 2291.97it/s] Traceback (most recent call last): File "/home/lysandre/.pyenv/versions/3.10.6/lib/python3.10/code.py", line 90, in runcode exec(code, self.locals) File "<input>", line 1, in <module> File "/home/lysandre/Workspaces/python/Metrics/GitHub-Metrics/.env/lib/python3.10/site-packages/datasets/load.py", line 1746, in load_dataset builder_instance.download_and_prepare( File "/home/lysandre/Workspaces/python/Metrics/GitHub-Metrics/.env/lib/python3.10/site-packages/datasets/builder.py", line 704, in download_and_prepare self._download_and_prepare( File "/home/lysandre/Workspaces/python/Metrics/GitHub-Metrics/.env/lib/python3.10/site-packages/datasets/builder.py", line 771, in _download_and_prepare split_generators = self._split_generators(dl_manager, **split_generators_kwargs) File "/home/lysandre/Workspaces/python/Metrics/GitHub-Metrics/.env/lib/python3.10/site-packages/datasets/packaged_modules/parquet/parquet.py", line 48, in _split_generators splits.append(datasets.SplitGenerator(name=split_name, gen_kwargs={"files": files})) File "<string>", line 5, in __init__ File "/home/lysandre/Workspaces/python/Metrics/GitHub-Metrics/.env/lib/python3.10/site-packages/datasets/splits.py", line 599, in __post_init__ NamedSplit(self.name) # check that it's a valid split name File "/home/lysandre/Workspaces/python/Metrics/GitHub-Metrics/.env/lib/python3.10/site-packages/datasets/splits.py", line 346, in __init__ raise ValueError(f"Split name should match '{_split_re}' but got '{split_name}'.") ValueError: Split name should match '^\w+(\.\w+)*$' but got 'identifier-with-column'. ``` ## Expected results I would expect `push_to_hub` to stop me in my tracks if trying to upload a split that will not be working afterwards. ## Actual results See above ## Environment info - `datasets` version: 2.4.0 - Platform: Linux-5.15.64-1-lts-x86_64-with-glibc2.36 - Python version: 3.10.6 - PyArrow version: 9.0.0 - Pandas version: 1.4.4
2022-09-13T10:16:35Z
4,945
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-07T13:45:17Z
https://api.github.com/repos/huggingface/datasets/issues/4945/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4945/timeline
Push to hub can push splits that do not respect the regex
https://api.github.com/repos/huggingface/datasets/issues/4945/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/30755778?v=4", "events_url": "https://api.github.com/users/LysandreJik/events{/privacy}", "followers_url": "https://api.github.com/users/LysandreJik/followers", "following_url": "https://api.github.com/users/LysandreJik/following{/other_user}", "gists_url": "https://api.github.com/users/LysandreJik/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/LysandreJik", "id": 30755778, "login": "LysandreJik", "node_id": "MDQ6VXNlcjMwNzU1Nzc4", "organizations_url": "https://api.github.com/users/LysandreJik/orgs", "received_events_url": "https://api.github.com/users/LysandreJik/received_events", "repos_url": "https://api.github.com/users/LysandreJik/repos", "site_admin": false, "starred_url": "https://api.github.com/users/LysandreJik/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LysandreJik/subscriptions", "type": "User", "url": "https://api.github.com/users/LysandreJik" }
[]
null
completed
MEMBER
2022-09-13T10:16:35Z
null
I_kwDODunzps5RV4iY
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4945/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4945
https://github.com/huggingface/datasets/issues/4945
false
1,364,313,569
https://api.github.com/repos/huggingface/datasets/issues/4944/labels{/name}
from datasets import set_caching_enabled set_caching_enabled(False) for ds_name in ["squad","newsqa","nqopen","narrativeqa"]: train_ds = load_from_disk("../../../dall/downstream/processedproqa/{}-train.hf".format(ds_name)) break train_ds = concatenate_datasets([train_ds,train_ds,train_ds,train_ds]) #operation 1 trainer = QuestionAnsweringTrainer( #huggingface trainer model=model, args=training_args, train_dataset=train_ds, eval_dataset= None, eval_examples=None, answer_column_name=answer_column, dataset_name="squad", tokenizer=tokenizer, data_collator=data_collator, compute_metrics=compute_metrics if training_args.predict_with_generate else None, ) with operation 1, the GPU memory increases from 16G to 23G
2022-09-07T12:34:58Z
4,944
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-07T08:46:30Z
https://api.github.com/repos/huggingface/datasets/issues/4944/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4944/timeline
larger dataset, larger GPU memory in the training phase? Is that correct?
https://api.github.com/repos/huggingface/datasets/issues/4944/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/38886373?v=4", "events_url": "https://api.github.com/users/debby1103/events{/privacy}", "followers_url": "https://api.github.com/users/debby1103/followers", "following_url": "https://api.github.com/users/debby1103/following{/other_user}", "gists_url": "https://api.github.com/users/debby1103/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/debby1103", "id": 38886373, "login": "debby1103", "node_id": "MDQ6VXNlcjM4ODg2Mzcz", "organizations_url": "https://api.github.com/users/debby1103/orgs", "received_events_url": "https://api.github.com/users/debby1103/received_events", "repos_url": "https://api.github.com/users/debby1103/repos", "site_admin": false, "starred_url": "https://api.github.com/users/debby1103/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/debby1103/subscriptions", "type": "User", "url": "https://api.github.com/users/debby1103" }
[]
null
completed
NONE
2022-09-07T12:34:58Z
null
I_kwDODunzps5RUcXh
[ "does the trainer save it in GPU? sooo curious... how to fix it", "It's my bad. didn't limit the input length" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4944/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4944
https://github.com/huggingface/datasets/issues/4944
false
1,363,967,650
https://api.github.com/repos/huggingface/datasets/issues/4943/labels{/name}
This PR addresses https://github.com/huggingface/datasets/issues/4795
2022-09-13T12:29:19Z
4,943
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-07T01:18:31Z
https://api.github.com/repos/huggingface/datasets/issues/4943/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4943/timeline
Add splits to MBPP dataset
https://api.github.com/repos/huggingface/datasets/issues/4943/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/2788526?v=4", "events_url": "https://api.github.com/users/cwarny/events{/privacy}", "followers_url": "https://api.github.com/users/cwarny/followers", "following_url": "https://api.github.com/users/cwarny/following{/other_user}", "gists_url": "https://api.github.com/users/cwarny/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cwarny", "id": 2788526, "login": "cwarny", "node_id": "MDQ6VXNlcjI3ODg1MjY=", "organizations_url": "https://api.github.com/users/cwarny/orgs", "received_events_url": "https://api.github.com/users/cwarny/received_events", "repos_url": "https://api.github.com/users/cwarny/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cwarny/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cwarny/subscriptions", "type": "User", "url": "https://api.github.com/users/cwarny" }
[]
null
null
CONTRIBUTOR
2022-09-13T12:27:21Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4943.diff", "html_url": "https://github.com/huggingface/datasets/pull/4943", "merged_at": "2022-09-13T12:27:21Z", "patch_url": "https://github.com/huggingface/datasets/pull/4943.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4943" }
PR_kwDODunzps4-eZd_
[ "```\r\n(env) cwarny@Cedrics-Air datasets % RUN_SLOW=1 pytest tests/test_dataset_common.py::LocalDatasetTest::test_load_real_dataset_mbpp\r\n================================================================================================ test session starts ==========================================================...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4943/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4943
https://github.com/huggingface/datasets/pull/4943
true
1,363,869,421
https://api.github.com/repos/huggingface/datasets/issues/4942/labels{/name}
## Describe the bug Both coarse and fine labels seem to be out of line. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = "trec" raw_datasets = load_dataset(dataset) df = pd.DataFrame(raw_datasets["test"]) df.head() ``` ## Expected results text (string) | coarse_label (class label) | fine_label (class label) -- | -- | -- How far is it from Denver to Aspen ? | 5 (NUM) | 40 (NUM:dist) What county is Modesto , California in ? | 4 (LOC) | 32 (LOC:city) Who was Galileo ? | 3 (HUM) | 31 (HUM:desc) What is an atom ? | 2 (DESC) | 24 (DESC:def) When did Hawaii become a state ? | 5 (NUM) | 39 (NUM:date) ## Actual results index | label-coarse |label-fine | text -- |-- | -- | -- 0 | 4 | 40 | How far is it from Denver to Aspen ? 1 | 5 | 21 | What county is Modesto , California in ? 2 | 3 | 12 | Who was Galileo ? 3 | 0 | 7 | What is an atom ? 4 | 4 | 8 | When did Hawaii become a state ? ## Environment info - `datasets` version: 2.4.0 - Platform: Linux-5.4.0-1086-azure-x86_64-with-glibc2.27 - Python version: 3.9.13 - PyArrow version: 8.0.0 - Pandas version: 1.4.3
2022-09-08T11:12:03Z
4,942
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-06T22:13:40Z
https://api.github.com/repos/huggingface/datasets/issues/4942/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4942/timeline
Trec Dataset has incorrect labels
https://api.github.com/repos/huggingface/datasets/issues/4942/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/6539145?v=4", "events_url": "https://api.github.com/users/wmpauli/events{/privacy}", "followers_url": "https://api.github.com/users/wmpauli/followers", "following_url": "https://api.github.com/users/wmpauli/following{/other_user}", "gists_url": "https://api.github.com/users/wmpauli/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/wmpauli", "id": 6539145, "login": "wmpauli", "node_id": "MDQ6VXNlcjY1MzkxNDU=", "organizations_url": "https://api.github.com/users/wmpauli/orgs", "received_events_url": "https://api.github.com/users/wmpauli/received_events", "repos_url": "https://api.github.com/users/wmpauli/repos", "site_admin": false, "starred_url": "https://api.github.com/users/wmpauli/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wmpauli/subscriptions", "type": "User", "url": "https://api.github.com/users/wmpauli" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
NONE
2022-09-08T11:12:03Z
null
I_kwDODunzps5RSv7t
[ "Thanks for reporting, @wmpauli. \r\n\r\nIndeed we recently fixed this issue:\r\n- #4801 \r\n\r\nThe fix will be accessible after our next library release. In the meantime, you can have it by passing `revision=\"main\"` to `load_dataset`." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4942/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4942
https://github.com/huggingface/datasets/issues/4942
false
1,363,622,861
https://api.github.com/repos/huggingface/datasets/issues/4941/labels{/name}
This PR: - adds Papers with Code ID - forces sync between GitHub and Hub, which previously failed due to Hub validation error of the license tag: https://github.com/huggingface/datasets/runs/8200223631?check_suite_focus=true
2022-09-06T18:28:17Z
4,941
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-06T17:46:37Z
https://api.github.com/repos/huggingface/datasets/issues/4941/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4941/timeline
Add Papers with Code ID to scifact dataset
https://api.github.com/repos/huggingface/datasets/issues/4941/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-06T18:26:01Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4941.diff", "html_url": "https://github.com/huggingface/datasets/pull/4941", "merged_at": "2022-09-06T18:26:01Z", "patch_url": "https://github.com/huggingface/datasets/pull/4941.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4941" }
PR_kwDODunzps4-dQ9F
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4941/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4941
https://github.com/huggingface/datasets/pull/4941
true
1,363,513,058
https://api.github.com/repos/huggingface/datasets/issues/4940/labels{/name}
This PR fixes issue reported on the Hub: - Label as multilingual: https://huggingface.co/datasets/xquad_r/discussions/1
2022-09-12T10:11:07Z
4,940
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-06T16:05:35Z
https://api.github.com/repos/huggingface/datasets/issues/4940/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4940/timeline
Fix multilinguality tag and missing sections in xquad_r dataset card
https://api.github.com/repos/huggingface/datasets/issues/4940/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-12T10:08:48Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4940.diff", "html_url": "https://github.com/huggingface/datasets/pull/4940", "merged_at": "2022-09-12T10:08:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/4940.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4940" }
PR_kwDODunzps4-c6WY
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4940/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4940
https://github.com/huggingface/datasets/pull/4940
true
1,363,468,679
https://api.github.com/repos/huggingface/datasets/issues/4939/labels{/name}
Fix issue reported on the Hub: https://huggingface.co/datasets/adv_glue/discussions/1
2022-09-06T17:42:10Z
4,939
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-06T15:31:16Z
https://api.github.com/repos/huggingface/datasets/issues/4939/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4939/timeline
Fix NonMatchingChecksumError in adv_glue dataset
https://api.github.com/repos/huggingface/datasets/issues/4939/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-06T17:39:16Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4939.diff", "html_url": "https://github.com/huggingface/datasets/pull/4939", "merged_at": "2022-09-06T17:39:16Z", "patch_url": "https://github.com/huggingface/datasets/pull/4939.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4939" }
PR_kwDODunzps4-cw4A
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4939/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4939
https://github.com/huggingface/datasets/pull/4939
true
1,363,429,228
https://api.github.com/repos/huggingface/datasets/issues/4938/labels{/name}
We added a notice in README.md to show that we renamed the master branch to main, but we can remove it now (it's been 2 months) I also unpinned the github issue about the branch renaming
2022-09-06T16:46:11Z
4,938
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-06T15:03:05Z
https://api.github.com/repos/huggingface/datasets/issues/4938/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4938/timeline
Remove main branch rename notice
https://api.github.com/repos/huggingface/datasets/issues/4938/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-06T16:43:53Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4938.diff", "html_url": "https://github.com/huggingface/datasets/pull/4938", "merged_at": "2022-09-06T16:43:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/4938.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4938" }
PR_kwDODunzps4-coaB
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4938/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4938
https://github.com/huggingface/datasets/pull/4938
true
1,363,426,946
https://api.github.com/repos/huggingface/datasets/issues/4937/labels{/name}
`huggingface-hub` says that the `identical_ok` argument of `HfApi.upload_file` is now deprecated, and will be removed soon. It even has no effect at the moment when it's passed: ```python Args: ... identical_ok (`bool`, *optional*, defaults to `True`): Deprecated: will be removed in 0.11.0. Changing this value has no effect. ... ``` There was only one occurence of `identical_ok=False` but it's maybe not worth adding a check ti verify if the files were the same. cc @mariosasko
2022-09-06T22:24:09Z
4,937
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-06T15:01:24Z
https://api.github.com/repos/huggingface/datasets/issues/4937/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4937/timeline
Remove deprecated identical_ok
https://api.github.com/repos/huggingface/datasets/issues/4937/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-09-06T22:21:57Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4937.diff", "html_url": "https://github.com/huggingface/datasets/pull/4937", "merged_at": "2022-09-06T22:21:57Z", "patch_url": "https://github.com/huggingface/datasets/pull/4937.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4937" }
PR_kwDODunzps4-cn6W
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4937/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4937
https://github.com/huggingface/datasets/pull/4937
true
1,363,274,907
https://api.github.com/repos/huggingface/datasets/issues/4936/labels{/name}
## Describe the bug VIVOS data is not accessible anymore, neither of these links work (at least from France): * https://ailab.hcmus.edu.vn/assets/vivos.tar.gz (data) * https://ailab.hcmus.edu.vn/vivos (dataset page) Therefore `load_dataset` doesn't work. ## Steps to reproduce the bug ```python ds = load_dataset("vivos") ``` ## Expected results dataset loaded ## Actual results ``` ConnectionError: Couldn't reach https://ailab.hcmus.edu.vn/assets/vivos.tar.gz (ConnectionError(MaxRetryError("HTTPSConnectionPool(host='ailab.hcmus.edu.vn', port=443): Max retries exceeded with url: /assets/vivos.tar.gz (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7f9d8a27d190>: Failed to establish a new connection: [Errno -5] No address associated with hostname'))"))) ``` Will try to contact the authors, as we wanted to use Vivos as an example in documentation on how to create scripts for audio datasets (https://github.com/huggingface/datasets/pull/4872), because it's small and straightforward and uses tar archives.
2022-09-21T06:06:02Z
4,936
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library", "id": 2067388877, "name": "dataset bug", "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug" } ]
2022-09-06T13:17:55Z
https://api.github.com/repos/huggingface/datasets/issues/4936/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4936/timeline
vivos (Vietnamese speech corpus) dataset not accessible
https://api.github.com/repos/huggingface/datasets/issues/4936/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
CONTRIBUTOR
2022-09-12T07:14:20Z
null
I_kwDODunzps5RQeyb
[ "If you need an example of a small audio datasets, I just created few hours ago a speech dataset with only 300MB of compressed audio files https://huggingface.co/datasets/indonesian-nlp/librivox-indonesia. It works also with streaming (@albertvillanova helped me adding this functionality) :-)", "@cahya-wirawan om...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4936/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4936
https://github.com/huggingface/datasets/issues/4936
false
1,363,226,736
https://api.github.com/repos/huggingface/datasets/issues/4935/labels{/name}
### Link _No response_ ### Description _No response_ ### Owner _No response_
2022-09-06T12:51:25Z
4,935
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
2022-09-06T12:41:50Z
https://api.github.com/repos/huggingface/datasets/issues/4935/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
https://api.github.com/repos/huggingface/datasets/issues/4935/timeline
Dataset Viewer issue for ubuntu_dialogs_corpus
https://api.github.com/repos/huggingface/datasets/issues/4935/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/87330568?v=4", "events_url": "https://api.github.com/users/CibinQuadance/events{/privacy}", "followers_url": "https://api.github.com/users/CibinQuadance/followers", "following_url": "https://api.github.com/users/CibinQuadance/following{/other_user}", "gists_url": "https://api.github.com/users/CibinQuadance/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CibinQuadance", "id": 87330568, "login": "CibinQuadance", "node_id": "MDQ6VXNlcjg3MzMwNTY4", "organizations_url": "https://api.github.com/users/CibinQuadance/orgs", "received_events_url": "https://api.github.com/users/CibinQuadance/received_events", "repos_url": "https://api.github.com/users/CibinQuadance/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CibinQuadance/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CibinQuadance/subscriptions", "type": "User", "url": "https://api.github.com/users/CibinQuadance" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url...
null
completed
NONE
2022-09-06T12:51:25Z
null
I_kwDODunzps5RQTBw
[ "The dataset maintainers (https://huggingface.co/datasets/ubuntu_dialogs_corpus) decided to forbid the dataset from being downloaded automatically (https://huggingface.co/docs/datasets/v2.4.0/en/loading#manual-download), and the dataset viewer respects this.\r\nWe will try to improve the error display though. Thank...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4935/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4935
https://github.com/huggingface/datasets/issues/4935
false
1,363,034,253
https://api.github.com/repos/huggingface/datasets/issues/4934/labels{/name}
### Link https://huggingface.co/datasets/indonesian-nlp/librivox-indonesia ### Description I created a new speech dataset https://huggingface.co/datasets/indonesian-nlp/librivox-indonesia, but the dataset preview doesn't work with following error message: ``` Server error Status code: 400 Exception: TypeError Message: unsupported operand type(s) for +: 'NoneType' and 'str' ``` Please help, I am not sure what the problem here is. Thanks a lot. ### Owner Yes
2022-09-06T12:46:40Z
4,934
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-09-06T10:03:23Z
https://api.github.com/repos/huggingface/datasets/issues/4934/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4934/timeline
Dataset Viewer issue for indonesian-nlp/librivox-indonesia
https://api.github.com/repos/huggingface/datasets/issues/4934/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7669893?v=4", "events_url": "https://api.github.com/users/cahya-wirawan/events{/privacy}", "followers_url": "https://api.github.com/users/cahya-wirawan/followers", "following_url": "https://api.github.com/users/cahya-wirawan/following{/other_user}", "gists_url": "https://api.github.com/users/cahya-wirawan/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cahya-wirawan", "id": 7669893, "login": "cahya-wirawan", "node_id": "MDQ6VXNlcjc2Njk4OTM=", "organizations_url": "https://api.github.com/users/cahya-wirawan/orgs", "received_events_url": "https://api.github.com/users/cahya-wirawan/received_events", "repos_url": "https://api.github.com/users/cahya-wirawan/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cahya-wirawan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cahya-wirawan/subscriptions", "type": "User", "url": "https://api.github.com/users/cahya-wirawan" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
CONTRIBUTOR
2022-09-06T12:46:40Z
null
I_kwDODunzps5RPkCN
[ "The error is not related to the dataset viewer. I'm having a look...", "Thanks @albertvillanova for checking the issue. Actually, I can use the dataset like following:\r\n```\r\n>>> from datasets import load_dataset\r\n>>> ds=load_dataset(\"indonesian-nlp/librivox-indonesia\")\r\nNo config specified, defaulting ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4934/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4934
https://github.com/huggingface/datasets/issues/4934
false
1,363,013,023
https://api.github.com/repos/huggingface/datasets/issues/4933/labels{/name}
## Describe the bug `Dataset/DatasetDict.filter()` cannot have `batched=True` due to `mask` (numpy array?) being non-iterable. ## Steps to reproduce the bug (In a python 3.7.12 env, I've tried 2.4.0 and 2.3.2 with both `pyarraw==9.0.0` and `pyarrow==8.0.0`.) ```python from datasets import load_dataset ds_mc4_ja = load_dataset("mc4", "ja") # This will take 6+ hours... perhaps test it with a toy dataset instead? ds_mc4_ja_2020 = ds_mc4_ja.filter( lambda example: example["timestamp"][:4] == "2020", batched=True, ) ``` ## Expected results No error ## Actual results ```python --------------------------------------------------------------------------- RemoteTraceback Traceback (most recent call last) RemoteTraceback: """ Traceback (most recent call last): File "/opt/conda/lib/python3.7/site-packages/multiprocess/pool.py", line 121, in worker result = (True, func(*args, **kwds)) File "/opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 557, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 524, in wrapper out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) File "/opt/conda/lib/python3.7/site-packages/datasets/fingerprint.py", line 480, in wrapper out = func(self, *args, **kwargs) File "/opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 2779, in _map_single offset=offset, File "/opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 2655, in apply_function_on_filtered_inputs processed_inputs = function(*fn_args, *additional_args, **fn_kwargs) File "/opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 2347, in decorated result = f(decorated_item, *args, **kwargs) File "/opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py", line 4946, in get_indices_from_mask_function indices_array = [i for i, to_keep in zip(indices, mask) if to_keep] TypeError: zip argument #2 must support iteration """ The above exception was the direct cause of the following exception: TypeError Traceback (most recent call last) /tmp/ipykernel_51348/2345782281.py in <module> 7 batched=True, 8 # batch_size=10_000, ----> 9 num_proc=111, 10 ) 11 # ds_mc4_ja_clean_2020 = ds_mc4_ja.filter( /opt/conda/lib/python3.7/site-packages/datasets/dataset_dict.py in filter(self, function, with_indices, input_columns, batched, batch_size, keep_in_memory, load_from_cache_file, cache_file_names, writer_batch_size, fn_kwargs, num_proc, desc) 878 desc=desc, 879 ) --> 880 for k, dataset in self.items() 881 } 882 ) /opt/conda/lib/python3.7/site-packages/datasets/dataset_dict.py in <dictcomp>(.0) 878 desc=desc, 879 ) --> 880 for k, dataset in self.items() 881 } 882 ) /opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py in wrapper(*args, **kwargs) 522 } 523 # apply actual function --> 524 out: Union["Dataset", "DatasetDict"] = func(self, *args, **kwargs) 525 datasets: List["Dataset"] = list(out.values()) if isinstance(out, dict) else [out] 526 # re-apply format to the output /opt/conda/lib/python3.7/site-packages/datasets/fingerprint.py in wrapper(*args, **kwargs) 478 # Call actual function 479 --> 480 out = func(self, *args, **kwargs) 481 482 # Update fingerprint of in-place transforms + update in-place history of transforms /opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py in filter(self, function, with_indices, input_columns, batched, batch_size, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc) 2920 new_fingerprint=new_fingerprint, 2921 input_columns=input_columns, -> 2922 desc=desc, 2923 ) 2924 new_dataset = copy.deepcopy(self) /opt/conda/lib/python3.7/site-packages/datasets/arrow_dataset.py in map(self, function, with_indices, with_rank, input_columns, batched, batch_size, drop_last_batch, remove_columns, keep_in_memory, load_from_cache_file, cache_file_name, writer_batch_size, features, disable_nullable, fn_kwargs, num_proc, suffix_template, new_fingerprint, desc) 2498 2499 for index, async_result in results.items(): -> 2500 transformed_shards[index] = async_result.get() 2501 2502 assert ( /opt/conda/lib/python3.7/site-packages/multiprocess/pool.py in get(self, timeout) 655 return self._value 656 else: --> 657 raise self._value 658 659 def _set(self, i, obj): TypeError: zip argument #2 must support iteration ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.4.0 - Platform: Linux-4.19.0-21-cloud-amd64-x86_64-with-debian-10.12 - Python version: 3.7.12 - PyArrow version: 9.0.0 - Pandas version: 1.3.5 (I've tried 2.4.0 and 2.3.2 with both `pyarraw==9.0.0` and `pyarrow==8.0.0`.)
2022-09-06T11:44:27Z
4,933
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-06T09:47:48Z
https://api.github.com/repos/huggingface/datasets/issues/4933/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4933/timeline
Dataset/DatasetDict.filter() cannot have `batched=True` due to `mask` (numpy array?) being non-iterable.
https://api.github.com/repos/huggingface/datasets/issues/4933/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/4812544?v=4", "events_url": "https://api.github.com/users/tianjianjiang/events{/privacy}", "followers_url": "https://api.github.com/users/tianjianjiang/followers", "following_url": "https://api.github.com/users/tianjianjiang/following{/other_user}", "gists_url": "https://api.github.com/users/tianjianjiang/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/tianjianjiang", "id": 4812544, "login": "tianjianjiang", "node_id": "MDQ6VXNlcjQ4MTI1NDQ=", "organizations_url": "https://api.github.com/users/tianjianjiang/orgs", "received_events_url": "https://api.github.com/users/tianjianjiang/received_events", "repos_url": "https://api.github.com/users/tianjianjiang/repos", "site_admin": false, "starred_url": "https://api.github.com/users/tianjianjiang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tianjianjiang/subscriptions", "type": "User", "url": "https://api.github.com/users/tianjianjiang" }
[]
null
completed
CONTRIBUTOR
2022-09-06T11:44:27Z
null
I_kwDODunzps5RPe2f
[ "Hi ! When `batched=True`, you filter function must take a batch as input, and return a list of booleans.\r\n\r\nIn your case, something like\r\n```python\r\nfrom datasets import load_dataset\r\n\r\n\r\nds_mc4_ja = load_dataset(\"mc4\", \"ja\") # This will take 6+ hours... perhaps test it with a toy dataset instea...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4933/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4933
https://github.com/huggingface/datasets/issues/4933
false
1,362,522,423
https://api.github.com/repos/huggingface/datasets/issues/4932/labels{/name}
### Link https://huggingface.co/datasets/bigscience-biomedical/biosses ### Description I've just been working on adding the dataset loader script to this dataset and working with the relative imports. I'm not sure how to interpret the error below (show where the dataset preview used to be) . ``` Status code: 400 Exception: ModuleNotFoundError Message: No module named 'datasets_modules.datasets.bigscience-biomedical--biosses.ddbd5893bf6c2f4db06f407665eaeac619520ba41f69d94ead28f7cc5b674056.bigbiohub' ``` ### Owner Yes
2022-09-06T14:24:56Z
4,932
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-09-05T22:40:32Z
https://api.github.com/repos/huggingface/datasets/issues/4932/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4932/timeline
Dataset Viewer issue for bigscience-biomedical/biosses
https://api.github.com/repos/huggingface/datasets/issues/4932/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/663051?v=4", "events_url": "https://api.github.com/users/galtay/events{/privacy}", "followers_url": "https://api.github.com/users/galtay/followers", "following_url": "https://api.github.com/users/galtay/following{/other_user}", "gists_url": "https://api.github.com/users/galtay/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/galtay", "id": 663051, "login": "galtay", "node_id": "MDQ6VXNlcjY2MzA1MQ==", "organizations_url": "https://api.github.com/users/galtay/orgs", "received_events_url": "https://api.github.com/users/galtay/received_events", "repos_url": "https://api.github.com/users/galtay/repos", "site_admin": false, "starred_url": "https://api.github.com/users/galtay/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/galtay/subscriptions", "type": "User", "url": "https://api.github.com/users/galtay" }
[]
null
completed
NONE
2022-09-06T14:24:56Z
null
I_kwDODunzps5RNnE3
[ "Possibly not related to the dataset viewer in itself. cc @huggingface/datasets.\r\n\r\nIn particular, I think that the import of bigbiohub is not working here: https://huggingface.co/datasets/bigscience-biomedical/biosses/blob/main/biosses.py#L29 (requires a relative path?)\r\n\r\n```python\r\n>>> from datasets im...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4932/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4932
https://github.com/huggingface/datasets/issues/4932
false
1,362,298,764
https://api.github.com/repos/huggingface/datasets/issues/4931/labels{/name}
Fix missing tags in dataset cards: - coqa - hyperpartisan_news_detection - opinosis - scientific_papers - scifact - search_qa - wiki_qa - wiki_split - wikisql This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833 - #4891 - #4896 - #4908 - #4921
2022-09-22T12:40:15Z
4,931
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-05T17:03:04Z
https://api.github.com/repos/huggingface/datasets/issues/4931/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4931/timeline
Fix missing tags in dataset cards
https://api.github.com/repos/huggingface/datasets/issues/4931/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-06T05:39:29Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4931.diff", "html_url": "https://github.com/huggingface/datasets/pull/4931", "merged_at": "2022-09-06T05:39:29Z", "patch_url": "https://github.com/huggingface/datasets/pull/4931.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4931" }
PR_kwDODunzps4-Y3L6
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4931/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4931
https://github.com/huggingface/datasets/pull/4931
true
1,362,193,587
https://api.github.com/repos/huggingface/datasets/issues/4930/labels{/name}
This PR adds the `cc-by-nc-2.0` to the list of licenses because it is required by `scifact` dataset: https://github.com/allenai/scifact/blob/master/LICENSE.md
2022-09-06T16:43:32Z
4,930
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-05T15:37:32Z
https://api.github.com/repos/huggingface/datasets/issues/4930/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4930/timeline
Add cc-by-nc-2.0 to list of licenses
https://api.github.com/repos/huggingface/datasets/issues/4930/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-05T17:01:04Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4930.diff", "html_url": "https://github.com/huggingface/datasets/pull/4930", "merged_at": "2022-09-05T17:01:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/4930.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4930" }
PR_kwDODunzps4-Yflc
[ "_The documentation is not available anymore as the PR was closed or merged._", "this list needs to be kept in sync with the ones in moon-landing and hub-docs :)", "@julien-c don't you think it might be better to a have a single file (source of truth) in one of the repos and then use it in every other repo, ins...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4930/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4930
https://github.com/huggingface/datasets/pull/4930
true
1,361,508,366
https://api.github.com/repos/huggingface/datasets/issues/4929/labels{/name}
As show in the [documentation page](https://huggingface.co/docs/datasets/loading) here the `"tr"in` should be `"train`. ![image](https://user-images.githubusercontent.com/7144772/188390445-e1f04d54-e3e3-4762-8686-63ecbe4087e5.png)
2022-09-06T02:11:03Z
4,929
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-05T07:18:54Z
https://api.github.com/repos/huggingface/datasets/issues/4929/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4929/timeline
Fixes a typo in loading documentation
https://api.github.com/repos/huggingface/datasets/issues/4929/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7144772?v=4", "events_url": "https://api.github.com/users/sighingnow/events{/privacy}", "followers_url": "https://api.github.com/users/sighingnow/followers", "following_url": "https://api.github.com/users/sighingnow/following{/other_user}", "gists_url": "https://api.github.com/users/sighingnow/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sighingnow", "id": 7144772, "login": "sighingnow", "node_id": "MDQ6VXNlcjcxNDQ3NzI=", "organizations_url": "https://api.github.com/users/sighingnow/orgs", "received_events_url": "https://api.github.com/users/sighingnow/received_events", "repos_url": "https://api.github.com/users/sighingnow/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sighingnow/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sighingnow/subscriptions", "type": "User", "url": "https://api.github.com/users/sighingnow" }
[]
null
null
CONTRIBUTOR
2022-09-05T13:06:38Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4929.diff", "html_url": "https://github.com/huggingface/datasets/pull/4929", "merged_at": "2022-09-05T13:06:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/4929.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4929" }
PR_kwDODunzps4-WK2w
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4929/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4929
https://github.com/huggingface/datasets/pull/4929
true
1,360,941,172
https://api.github.com/repos/huggingface/datasets/issues/4928/labels{/name}
Fixes #3094 Add ability to read/write to SQLite files and also read from any SQL database supported by SQLAlchemy. I didn't add SQLAlchemy as a dependence as it is fairly big and it remains optional. I also recorded a Loom to showcase the feature. https://www.loom.com/share/f0e602c2de8a46f58bca4b43333d541f
2022-10-03T16:34:36Z
4,928
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-03T19:09:08Z
https://api.github.com/repos/huggingface/datasets/issues/4928/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4928/timeline
Add ability to read-write to SQL databases.
https://api.github.com/repos/huggingface/datasets/issues/4928/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8976546?v=4", "events_url": "https://api.github.com/users/Dref360/events{/privacy}", "followers_url": "https://api.github.com/users/Dref360/followers", "following_url": "https://api.github.com/users/Dref360/following{/other_user}", "gists_url": "https://api.github.com/users/Dref360/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Dref360", "id": 8976546, "login": "Dref360", "node_id": "MDQ6VXNlcjg5NzY1NDY=", "organizations_url": "https://api.github.com/users/Dref360/orgs", "received_events_url": "https://api.github.com/users/Dref360/received_events", "repos_url": "https://api.github.com/users/Dref360/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Dref360/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Dref360/subscriptions", "type": "User", "url": "https://api.github.com/users/Dref360" }
[]
null
null
CONTRIBUTOR
2022-10-03T16:32:28Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4928.diff", "html_url": "https://github.com/huggingface/datasets/pull/4928", "merged_at": "2022-10-03T16:32:28Z", "patch_url": "https://github.com/huggingface/datasets/pull/4928.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4928" }
PR_kwDODunzps4-Ubi4
[ "_The documentation is not available anymore as the PR was closed or merged._", "Ah CI runs with `pandas=1.3.5` which doesn't return the number of row inserted.", "wow this is super cool!", "@lhoestq I'm getting error in integration tests, not sure if it's related to my PR. Any help would be appreciated :) \r...
{ "+1": 2, "-1": 0, "confused": 0, "eyes": 2, "heart": 4, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 8, "url": "https://api.github.com/repos/huggingface/datasets/issues/4928/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4928
https://github.com/huggingface/datasets/pull/4928
true
1,360,428,139
https://api.github.com/repos/huggingface/datasets/issues/4927/labels{/name}
I've fixed some typos in BLEU metric card.
2022-09-09T16:28:15Z
4,927
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-02T17:00:56Z
https://api.github.com/repos/huggingface/datasets/issues/4927/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4927/timeline
fix BLEU metric card
https://api.github.com/repos/huggingface/datasets/issues/4927/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/40452030?v=4", "events_url": "https://api.github.com/users/antoniolanza1996/events{/privacy}", "followers_url": "https://api.github.com/users/antoniolanza1996/followers", "following_url": "https://api.github.com/users/antoniolanza1996/following{/other_user}", "gists_url": "https://api.github.com/users/antoniolanza1996/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/antoniolanza1996", "id": 40452030, "login": "antoniolanza1996", "node_id": "MDQ6VXNlcjQwNDUyMDMw", "organizations_url": "https://api.github.com/users/antoniolanza1996/orgs", "received_events_url": "https://api.github.com/users/antoniolanza1996/received_events", "repos_url": "https://api.github.com/users/antoniolanza1996/repos", "site_admin": false, "starred_url": "https://api.github.com/users/antoniolanza1996/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/antoniolanza1996/subscriptions", "type": "User", "url": "https://api.github.com/users/antoniolanza1996" }
[]
null
null
CONTRIBUTOR
2022-09-09T16:28:15Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4927.diff", "html_url": "https://github.com/huggingface/datasets/pull/4927", "merged_at": "2022-09-09T16:28:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/4927.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4927" }
PR_kwDODunzps4-S0we
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4927/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4927
https://github.com/huggingface/datasets/pull/4927
true
1,360,384,484
https://api.github.com/repos/huggingface/datasets/issues/4926/labels{/name}
To simplify the addition of new datasets, we'd like to have the dataset infos in the YAML and deprecate the dataset_infos.json file. YAML is readable and easy to edit, and the YAML metadata of the readme already contain dataset metadata so we would have everything in one place. To be more specific, I moved these fields from DatasetInfo to the YAML: - config_name (if there are several configs) - download_size - dataset_size - features - splits Here is what I ended up with for `squad`: ```yaml dataset_info: features: - name: id dtype: string - name: title dtype: string - name: context dtype: string - name: question dtype: string - name: answers sequence: - name: text dtype: string - name: answer_start dtype: int32 splits: - name: train num_bytes: 79346360 num_examples: 87599 - name: validation num_bytes: 10473040 num_examples: 10570 config_name: plain_text download_size: 35142551 dataset_size: 89819400 ``` and it can be a list if there are several configs I already did the change for `conll2000` and `crime_and_punish` as an example. ## Implementation details ### Load/Read This is done via `DatasetInfosDict.write_to_directory/from_directory` I had to implement a custom the YAML export logic for `SplitDict`, `Version` and `Features`. The first two are trivial, but the logic for `Features` is more complicated, because I added a simplification step (or the YAML would be too long and less readable): it's just a formatting step to remove unnecessary nesting of YAML data. ### Other changes I had to update the DatasetModule factories to also download the README.md alongside the dataset scripts/data files, and not just the dataset_infos.json ## YAML validation I removed the old validation code that was in metadata.py, now we can just use the Hub YAML validation ## Datasets-cli The `datasets-cli test --save_infos` command now creates a README.md file with the dataset_infos in it, instead of a datasets_infos.json file ## Backward compatibility `dataset_infos.json` files are still supported and loaded if they exist to have full backward compatibility. Though I removed the unnecessary keys when the value is the default (like all the `id: null` from the Value feature types) to make them easier to read. ## TODO - [x] add comments - [x] tests - [x] document the new YAML fields - [x] try to reload the new dataset_infos.json file content with an old version of `datasets` ## EDITS - removed "config_name" when there's only one config - removed "version" for now (?), because it's not useful in general - renamed the yaml field "dataset_info" instead of "dataset_infos", since it only has one by default (and because "infos" is not english) Fix https://github.com/huggingface/datasets/issues/4876
2022-10-03T09:13:07Z
4,926
null
https://api.github.com/repos/huggingface/datasets
false
[ { "color": "0e8a16", "default": false, "description": "Contribution to a dataset script", "id": 4564477500, "name": "dataset contribution", "node_id": "LA_kwDODunzps8AAAABEBBmPA", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20contribution" } ]
2022-09-02T16:10:05Z
https://api.github.com/repos/huggingface/datasets/issues/4926/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4926/timeline
Dataset infos in yaml
https://api.github.com/repos/huggingface/datasets/issues/4926/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
2022-10-03T09:11:12Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4926.diff", "html_url": "https://github.com/huggingface/datasets/pull/4926", "merged_at": "2022-10-03T09:11:12Z", "patch_url": "https://github.com/huggingface/datasets/pull/4926.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4926" }
PR_kwDODunzps4-Srm1
[ "_The documentation is not available anymore as the PR was closed or merged._", "Alright this is ready for review :)\r\nI mostly would like your opinion on the YAML structure and what we can do in the docs (IMO we can add the docs about those fields in the Hub docs). Other than that let me know if the changes in ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4926/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4926
https://github.com/huggingface/datasets/pull/4926
true
1,360,007,616
https://api.github.com/repos/huggingface/datasets/issues/4925/labels{/name}
This PR adds a small note about how to load image / audio datasets that have multiple splits in their dataset structure. Related forum thread: https://discuss.huggingface.co/t/loading-train-and-test-splits-with-audiofolder/22447 cc @NielsRogge
2022-09-26T12:21:30Z
4,925
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-09-02T10:31:58Z
https://api.github.com/repos/huggingface/datasets/issues/4925/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4925/timeline
Add note about loading image / audio files to docs
https://api.github.com/repos/huggingface/datasets/issues/4925/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
[]
null
null
MEMBER
2022-09-23T13:59:07Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4925.diff", "html_url": "https://github.com/huggingface/datasets/pull/4925", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/4925.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4925" }
PR_kwDODunzps4-RbP5
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4925). All of your documentation changes will be reflected on that endpoint.", "Thanks for the feedback @polinaeterna ! I've reworded the docs a bit to integrate your comments and this should be ready for another review :)", ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4925/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4925
https://github.com/huggingface/datasets/pull/4925
true
1,358,611,513
https://api.github.com/repos/huggingface/datasets/issues/4924/labels{/name}
## Describe the bug When loading the datasets seperately and saving them on disk, I want to concatenate them. But `concatenate_datasets` is filling up my RAM and the process gets killed. Is there a way to prevent this from happening or is this intended behaviour? Thanks in advance ## Steps to reproduce the bug ```python gcs = gcsfs.GCSFileSystem(project='project') datasets = [load_from_disk(f'path/to/slice/of/data/{i}', fs=gcs, keep_in_memory=False) for i in range(10)] dataset = concatenate_datasets(datasets) ``` ## Expected results A concatenated dataset which is stored on my disk. ## Actual results Concatenated dataset gets loaded into RAM and overflows it which gets the process killed. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.4.0 - Platform: Linux-4.19.0-21-cloud-amd64-x86_64-with-glibc2.10 - Python version: 3.8.13 - PyArrow version: 8.0.1 - Pandas version: 1.4.3
2022-09-01T11:50:54Z
4,924
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-09-01T10:25:17Z
https://api.github.com/repos/huggingface/datasets/issues/4924/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4924/timeline
Concatenate_datasets loads everything into RAM
https://api.github.com/repos/huggingface/datasets/issues/4924/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/39416047?v=4", "events_url": "https://api.github.com/users/louisdeneve/events{/privacy}", "followers_url": "https://api.github.com/users/louisdeneve/followers", "following_url": "https://api.github.com/users/louisdeneve/following{/other_user}", "gists_url": "https://api.github.com/users/louisdeneve/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/louisdeneve", "id": 39416047, "login": "louisdeneve", "node_id": "MDQ6VXNlcjM5NDE2MDQ3", "organizations_url": "https://api.github.com/users/louisdeneve/orgs", "received_events_url": "https://api.github.com/users/louisdeneve/received_events", "repos_url": "https://api.github.com/users/louisdeneve/repos", "site_admin": false, "starred_url": "https://api.github.com/users/louisdeneve/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/louisdeneve/subscriptions", "type": "User", "url": "https://api.github.com/users/louisdeneve" }
[]
null
completed
NONE
2022-09-01T11:50:54Z
null
I_kwDODunzps5Q-sQ5
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4924/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4924
https://github.com/huggingface/datasets/issues/4924
false
1,357,735,287
https://api.github.com/repos/huggingface/datasets/issues/4923/labels{/name}
`torchaudio>0.12` fails with decoding mp3 files if `ffmpeg<4`. currently we ask users to downgrade torchaudio, but sometimes it's not possible as torchaudio version is binded to torch version. as a temporary workaround we can decode mp3 with librosa (though it 60 times slower, at least it works) another option would be to ask users to install the required version of `ffmpeg`, but is non-trivial on colab: it's not in apt packages in ubuntu 18 and `conda` is not preinstalled (with `conda` it would be easily installable) - [x] decode with torchaudio anyway if the version of ffmpeg is correct? it's 60 times faster - [x] tests - [x] DO NOT FORGET to get back all the tests see https://github.com/huggingface/datasets/issues/4776 and https://github.com/huggingface/datasets/issues/3663#issuecomment-1225797165 (there is a Colab notebook to reproduce the error)
2022-11-02T11:54:33Z
4,923
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-31T18:57:59Z
https://api.github.com/repos/huggingface/datasets/issues/4923/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4923/timeline
decode mp3 with librosa if torchaudio is > 0.12 as a temporary workaround
https://api.github.com/repos/huggingface/datasets/issues/4923/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/16348744?v=4", "events_url": "https://api.github.com/users/polinaeterna/events{/privacy}", "followers_url": "https://api.github.com/users/polinaeterna/followers", "following_url": "https://api.github.com/users/polinaeterna/following{/other_user}", "gists_url": "https://api.github.com/users/polinaeterna/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/polinaeterna", "id": 16348744, "login": "polinaeterna", "node_id": "MDQ6VXNlcjE2MzQ4NzQ0", "organizations_url": "https://api.github.com/users/polinaeterna/orgs", "received_events_url": "https://api.github.com/users/polinaeterna/received_events", "repos_url": "https://api.github.com/users/polinaeterna/repos", "site_admin": false, "starred_url": "https://api.github.com/users/polinaeterna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/polinaeterna/subscriptions", "type": "User", "url": "https://api.github.com/users/polinaeterna" }
[]
null
null
CONTRIBUTOR
2022-09-20T13:12:52Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4923.diff", "html_url": "https://github.com/huggingface/datasets/pull/4923", "merged_at": "2022-09-20T13:12:52Z", "patch_url": "https://github.com/huggingface/datasets/pull/4923.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4923" }
PR_kwDODunzps4-Jv7C
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks ! Should we still support torchaudio>0.12 if it works ? And if it doesn't we can explain that downgrading is the right solution, or alternatively use librosa", "@lhoestq \r\n\r\n> Should we still support torchaudio>0.12 if i...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4923/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4923
https://github.com/huggingface/datasets/pull/4923
true
1,357,684,018
https://api.github.com/repos/huggingface/datasets/issues/4922/labels{/name}
## Describe the bug When trying to load a streaming dataset in Google Colab the loading fails with an I/O error ## Steps to reproduce the bug ```python import datasets from datasets import load_dataset hf_ds = load_dataset(path='wmt19', name='cs-en', streaming=True, split=datasets.Split.VALIDATION) list(hf_ds.take(5)) ``` ## Expected results It should load five data points ## Actual results ``` --------------------------------------------------------------------------- ValueError Traceback (most recent call last) [<ipython-input-13-7b5b8b1e7e58>](https://localhost:8080/#) in <module> 2 from datasets import load_dataset 3 hf_ds = load_dataset(path='wmt19', name='cs-en', streaming=True, split=datasets.Split.VALIDATION) ----> 4 list(hf_ds.take(5)) 6 frames [/usr/local/lib/python3.7/dist-packages/datasets/iterable_dataset.py](https://localhost:8080/#) in __iter__(self) 716 717 def __iter__(self): --> 718 for key, example in self._iter(): 719 if self.features: 720 # `IterableDataset` automatically fills missing columns with None. [/usr/local/lib/python3.7/dist-packages/datasets/iterable_dataset.py](https://localhost:8080/#) in _iter(self) 706 else: 707 ex_iterable = self._ex_iterable --> 708 yield from ex_iterable 709 710 def _iter_shard(self, shard_idx: int): [/usr/local/lib/python3.7/dist-packages/datasets/iterable_dataset.py](https://localhost:8080/#) in __iter__(self) 582 583 def __iter__(self): --> 584 yield from islice(self.ex_iterable, self.n) 585 586 def shuffle_data_sources(self, generator: np.random.Generator) -> "TakeExamplesIterable": [/usr/local/lib/python3.7/dist-packages/datasets/iterable_dataset.py](https://localhost:8080/#) in __iter__(self) 110 111 def __iter__(self): --> 112 yield from self.generate_examples_fn(**self.kwargs) 113 114 def shuffle_data_sources(self, generator: np.random.Generator) -> "ExamplesIterable": [~/.cache/huggingface/modules/datasets_modules/datasets/wmt19/aeadcbe9f1cbf9969e603239d33d3e43670cf250c1158edf74f5f6e74d4f21d0/wmt_utils.py](https://localhost:8080/#) in _generate_examples(self, split_subsets, extraction_map, with_translation) 845 raise ValueError("Invalid number of files: %d" % len(files)) 846 --> 847 for sub_key, ex in sub_generator(*sub_generator_args): 848 if not all(ex.values()): 849 continue [~/.cache/huggingface/modules/datasets_modules/datasets/wmt19/aeadcbe9f1cbf9969e603239d33d3e43670cf250c1158edf74f5f6e74d4f21d0/wmt_utils.py](https://localhost:8080/#) in _parse_parallel_sentences(f1, f2, filename1, filename2) 923 l2_sentences, l2 = parse_file(f2_i, filename2) 924 --> 925 for line_id, (s1, s2) in enumerate(zip(l1_sentences, l2_sentences)): 926 key = f"{f_id}/{line_id}" 927 yield key, {l1: s1, l2: s2} [~/.cache/huggingface/modules/datasets_modules/datasets/wmt19/aeadcbe9f1cbf9969e603239d33d3e43670cf250c1158edf74f5f6e74d4f21d0/wmt_utils.py](https://localhost:8080/#) in gen() 895 896 def gen(): --> 897 with open(path, encoding="utf-8") as f: 898 for line in f: 899 seg_match = re.match(seg_re, line) ValueError: I/O operation on closed file. ``` ## Environment info Copy-and-paste the text below in your GitHub issue. - `datasets` version: 2.4.0 - Platform: Linux-5.4.188+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.13 - PyArrow version: 9.0.0. (the same error happened with PyArrow version 6.0.0) - Pandas version: 1.3.5
2022-08-31T18:15:48Z
4,922
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-08-31T18:08:26Z
https://api.github.com/repos/huggingface/datasets/issues/4922/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4922/timeline
I/O error on Google Colab in streaming mode
https://api.github.com/repos/huggingface/datasets/issues/4922/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/5595043?v=4", "events_url": "https://api.github.com/users/jotterbach/events{/privacy}", "followers_url": "https://api.github.com/users/jotterbach/followers", "following_url": "https://api.github.com/users/jotterbach/following{/other_user}", "gists_url": "https://api.github.com/users/jotterbach/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jotterbach", "id": 5595043, "login": "jotterbach", "node_id": "MDQ6VXNlcjU1OTUwNDM=", "organizations_url": "https://api.github.com/users/jotterbach/orgs", "received_events_url": "https://api.github.com/users/jotterbach/received_events", "repos_url": "https://api.github.com/users/jotterbach/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jotterbach/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jotterbach/subscriptions", "type": "User", "url": "https://api.github.com/users/jotterbach" }
[]
null
not_planned
NONE
2022-08-31T18:15:48Z
null
I_kwDODunzps5Q7J0y
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4922/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4922
https://github.com/huggingface/datasets/issues/4922
false
1,357,609,003
https://api.github.com/repos/huggingface/datasets/issues/4921/labels{/name}
Fix missing tags in dataset cards: - eraser_multi_rc - hotpot_qa - metooma - movie_rationales - qanta - quora - quoref - race - ted_hrlr - ted_talks_iwslt This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833 - #4891 - #4896 - #4908
2022-09-22T14:34:11Z
4,921
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-31T16:52:27Z
https://api.github.com/repos/huggingface/datasets/issues/4921/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4921/timeline
Fix missing tags in dataset cards
https://api.github.com/repos/huggingface/datasets/issues/4921/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-09-01T05:04:53Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4921.diff", "html_url": "https://github.com/huggingface/datasets/pull/4921", "merged_at": "2022-09-01T05:04:53Z", "patch_url": "https://github.com/huggingface/datasets/pull/4921.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4921" }
PR_kwDODunzps4-JVFV
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4921/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4921
https://github.com/huggingface/datasets/pull/4921
true
1,357,564,589
https://api.github.com/repos/huggingface/datasets/issues/4920/labels{/name}
## Describe the bug Unable to load local tsv files through load_dataset method. ## Steps to reproduce the bug ```python # Sample code to reproduce the bug data_files = { 'train': 'train.tsv', 'test': 'test.tsv' } raw_datasets = load_dataset('tsv', data_files=data_files) ## Expected results I am pretty sure the data files exist in the current directory. The above code should load them as Datasets, but threw exceptions. ## Actual results --------------------------------------------------------------------------- FileNotFoundError Traceback (most recent call last) [<ipython-input-9-24207899c1af>](https://localhost:8080/#) in <module> ----> 1 raw_datasets = load_dataset('tsv', data_files='train.tsv') 2 frames [/usr/local/lib/python3.7/dist-packages/datasets/load.py](https://localhost:8080/#) in dataset_module_factory(path, revision, download_config, download_mode, dynamic_modules_path, data_dir, data_files, **download_kwargs) 1244 f"Couldn't find a dataset script at {relative_to_absolute_path(combined_path)} or any data file in the same directory. " 1245 f"Couldn't find '{path}' on the Hugging Face Hub either: {type(e1).__name__}: {e1}" -> 1246 ) from None 1247 raise e1 from None 1248 else: FileNotFoundError: Couldn't find a dataset script at /content/tsv/tsv.py or any data file in the same directory. Couldn't find 'tsv' on the Hugging Face Hub either: FileNotFoundError: Couldn't find file at https://raw.githubusercontent.com/huggingface/datasets/main/datasets/tsv/tsv.py ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.4.0 - Platform: Linux-5.4.188+-x86_64-with-Ubuntu-18.04-bionic - Python version: 3.7.13 - PyArrow version: 6.0.1 - Pandas version: 1.3.5
2022-09-01T05:31:30Z
4,920
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-08-31T16:13:39Z
https://api.github.com/repos/huggingface/datasets/issues/4920/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4920/timeline
Unable to load local tsv files through load_dataset method
https://api.github.com/repos/huggingface/datasets/issues/4920/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/44038517?v=4", "events_url": "https://api.github.com/users/DataNoob0723/events{/privacy}", "followers_url": "https://api.github.com/users/DataNoob0723/followers", "following_url": "https://api.github.com/users/DataNoob0723/following{/other_user}", "gists_url": "https://api.github.com/users/DataNoob0723/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/DataNoob0723", "id": 44038517, "login": "DataNoob0723", "node_id": "MDQ6VXNlcjQ0MDM4NTE3", "organizations_url": "https://api.github.com/users/DataNoob0723/orgs", "received_events_url": "https://api.github.com/users/DataNoob0723/received_events", "repos_url": "https://api.github.com/users/DataNoob0723/repos", "site_admin": false, "starred_url": "https://api.github.com/users/DataNoob0723/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DataNoob0723/subscriptions", "type": "User", "url": "https://api.github.com/users/DataNoob0723" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
NONE
2022-09-01T05:31:30Z
null
I_kwDODunzps5Q6sqt
[ "Hi @DataNoob0723,\r\n\r\nUnder the hood, we use `pandas` to load CSV/TSV files. Therefore, you should use \"csv\" and pass `sep=\"\\t\"`, as explained in our docs: https://huggingface.co/docs/datasets/v2.4.0/en/package_reference/loading_methods#from-files\r\n```python\r\nds = load_dataset('csv', sep=\"\\t\", data_...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4920/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4920
https://github.com/huggingface/datasets/issues/4920
false
1,357,441,599
https://api.github.com/repos/huggingface/datasets/issues/4919/labels{/name}
Hi @lhoestq what do you think? Let me give you a code sample: ```py >>> import datasets >>> foo = datasets.Dataset.from_dict({'foo':[0,1], 'bar':[2,3]}) >>> foo.save_to_disk('foo') # edit foo/dataset_info.json e.g. rename the 'foo' feature to 'baz' >>> datasets.load_from_disk('foo') --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-2-4863e606b330> in <module> ----> 1 datasets.load_from_disk('foo') ~/code/datasets/src/datasets/load.py in load_from_disk(dataset_path, fs, keep_in_memory) 1851 raise FileNotFoundError(f"Directory {dataset_path} not found") 1852 if fs.isfile(Path(dest_dataset_path, config.DATASET_INFO_FILENAME).as_posix()): -> 1853 return Dataset.load_from_disk(dataset_path, fs, keep_in_memory=keep_in_memory) 1854 elif fs.isfile(Path(dest_dataset_path, config.DATASETDICT_JSON_FILENAME).as_posix()): 1855 return DatasetDict.load_from_disk(dataset_path, fs, keep_in_memory=keep_in_memory) ~/code/datasets/src/datasets/arrow_dataset.py in load_from_disk(dataset_path, fs, keep_in_memory) 1230 info=dataset_info, 1231 split=split, -> 1232 fingerprint=state["_fingerprint"], 1233 ) 1234 ~/code/datasets/src/datasets/arrow_dataset.py in __init__(self, arrow_table, info, split, indices_table, fingerprint) 687 self.info.features = inferred_features 688 else: # make sure the nested columns are in the right order --> 689 self.info.features = self.info.features.reorder_fields_as(inferred_features) 690 691 # Infer fingerprint if None ~/code/datasets/src/datasets/features/features.py in reorder_fields_as(self, other) 1771 return source 1772 -> 1773 return Features(recursive_reorder(self, other)) 1774 1775 def flatten(self, max_depth=16) -> "Features": ~/code/datasets/src/datasets/features/features.py in recursive_reorder(source, target, stack) 1760 f"{source.keys()-target.keys()} are missing from dataset.arrow " 1761 f"and {target.keys()-source.keys()} are missing from dataset_info.json"+stack_position) -> 1762 raise ValueError(message) 1763 return {key: recursive_reorder(source[key], target[key], stack + f".{key}") for key in target} 1764 elif isinstance(source, list): ValueError: Keys mismatch: between {'baz': Value(dtype='int64', id=None), 'bar': Value(dtype='int64', id=None)} (dataset_info.json) and {'foo': Value(dtype='int64', id=None), 'bar': Value(dtype='int64', id=None)} (inferred from dataset.arrow). {'baz'} are missing from dataset.arrow and {'foo'} are missing from dataset_info.json ```
2022-09-05T08:46:01Z
4,919
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-31T14:41:36Z
https://api.github.com/repos/huggingface/datasets/issues/4919/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4919/timeline
feat: improve error message on Keys mismatch. closes #4917
https://api.github.com/repos/huggingface/datasets/issues/4919/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/25532159?v=4", "events_url": "https://api.github.com/users/PaulLerner/events{/privacy}", "followers_url": "https://api.github.com/users/PaulLerner/followers", "following_url": "https://api.github.com/users/PaulLerner/following{/other_user}", "gists_url": "https://api.github.com/users/PaulLerner/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/PaulLerner", "id": 25532159, "login": "PaulLerner", "node_id": "MDQ6VXNlcjI1NTMyMTU5", "organizations_url": "https://api.github.com/users/PaulLerner/orgs", "received_events_url": "https://api.github.com/users/PaulLerner/received_events", "repos_url": "https://api.github.com/users/PaulLerner/repos", "site_admin": false, "starred_url": "https://api.github.com/users/PaulLerner/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/PaulLerner/subscriptions", "type": "User", "url": "https://api.github.com/users/PaulLerner" }
[]
null
null
CONTRIBUTOR
2022-09-05T08:43:33Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4919.diff", "html_url": "https://github.com/huggingface/datasets/pull/4919", "merged_at": "2022-09-05T08:43:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/4919.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4919" }
PR_kwDODunzps4-IxDZ
[ "_The documentation is not available anymore as the PR was closed or merged._", "We are having an unrelated issue that makes several tests fail. We are working on that. Once fixed, you will be able to merge the main branch into this, so that you get the fix and the tests pass..." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4919/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4919
https://github.com/huggingface/datasets/pull/4919
true
1,357,242,757
https://api.github.com/repos/huggingface/datasets/issues/4918/labels{/name}
### Link https://huggingface.co/datasets/pysentimiento/spanish-targeted-sentiment-headlines ### Description After moving the dataset from my user (`finiteautomata`) to the `pysentimiento` organization, the dataset viewer says that it doesn't exist. ### Owner _No response_
2022-09-05T21:36:34Z
4,918
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
2022-08-31T12:09:07Z
https://api.github.com/repos/huggingface/datasets/issues/4918/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
https://api.github.com/repos/huggingface/datasets/issues/4918/timeline
Dataset Viewer issue for pysentimiento/spanish-targeted-sentiment-headlines
https://api.github.com/repos/huggingface/datasets/issues/4918/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/167943?v=4", "events_url": "https://api.github.com/users/finiteautomata/events{/privacy}", "followers_url": "https://api.github.com/users/finiteautomata/followers", "following_url": "https://api.github.com/users/finiteautomata/following{/other_user}", "gists_url": "https://api.github.com/users/finiteautomata/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/finiteautomata", "id": 167943, "login": "finiteautomata", "node_id": "MDQ6VXNlcjE2Nzk0Mw==", "organizations_url": "https://api.github.com/users/finiteautomata/orgs", "received_events_url": "https://api.github.com/users/finiteautomata/received_events", "repos_url": "https://api.github.com/users/finiteautomata/repos", "site_admin": false, "starred_url": "https://api.github.com/users/finiteautomata/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/finiteautomata/subscriptions", "type": "User", "url": "https://api.github.com/users/finiteautomata" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url...
null
completed
NONE
2022-09-05T16:32:44Z
null
I_kwDODunzps5Q5eGF
[ "Thanks for reporting, it's fixed now (I refreshed it manually). It's a known issue; we hope it will be fixed permanently in a few days.\r\n\r\n<img width=\"1508\" alt=\"Capture d’écran 2022-09-05 à 18 31 22\" src=\"https://user-images.githubusercontent.com/1676121/188489762-0ed86a7e-dfb3-46e8-a125-43b815a2c6f4.p...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4918/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4918
https://github.com/huggingface/datasets/issues/4918
false
1,357,193,841
https://api.github.com/repos/huggingface/datasets/issues/4917/labels{/name}
**Is your feature request related to a problem? Please describe.** When loading a dataset from disk with a defect in its `dataset_info.json` describing its features (I don’t know when/why/how this happens but it deserves its own issue), you will get an error message like: `ValueError: Keys mismatch: between {'bar': Value(dtype='int64', id=None)} and {'foo': Value(dtype='int64', id=None)}` Which is fine when you have only a few features like in the example but it gets very hard to read when you have a lot of features in your dataset. **Describe the solution you'd like** The error message should give the difference between the features (what keys are in A but missing in B and vice-versa). It should also tell which keys are inferred from `dataset.arrow` and which come from `dataset_info.json`. Willing to help :)
2022-09-05T08:43:38Z
4,917
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "7057ff", "default": true...
2022-08-31T11:24:34Z
https://api.github.com/repos/huggingface/datasets/issues/4917/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4917/timeline
Keys mismatch: make error message more informative
https://api.github.com/repos/huggingface/datasets/issues/4917/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/25532159?v=4", "events_url": "https://api.github.com/users/PaulLerner/events{/privacy}", "followers_url": "https://api.github.com/users/PaulLerner/followers", "following_url": "https://api.github.com/users/PaulLerner/following{/other_user}", "gists_url": "https://api.github.com/users/PaulLerner/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/PaulLerner", "id": 25532159, "login": "PaulLerner", "node_id": "MDQ6VXNlcjI1NTMyMTU5", "organizations_url": "https://api.github.com/users/PaulLerner/orgs", "received_events_url": "https://api.github.com/users/PaulLerner/received_events", "repos_url": "https://api.github.com/users/PaulLerner/repos", "site_admin": false, "starred_url": "https://api.github.com/users/PaulLerner/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/PaulLerner/subscriptions", "type": "User", "url": "https://api.github.com/users/PaulLerner" }
[]
null
completed
CONTRIBUTOR
2022-09-05T08:43:38Z
null
I_kwDODunzps5Q5SJx
[ "Good idea ! I think this can be improved in `Features.reorder_fields_as()` indeed at\r\n\r\nhttps://github.com/huggingface/datasets/blob/7feeb5648a63b6135a8259dedc3b1e19185ee4c7/src/datasets/features/features.py#L1739-L1740\r\n\r\nIs it something you would be interested in contributing ?", "Is this open to work ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4917/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4917
https://github.com/huggingface/datasets/issues/4917
false
1,357,076,940
https://api.github.com/repos/huggingface/datasets/issues/4916/labels{/name}
## Describe the bug Hi, I am currently trying to download wikipedia dataset using load_dataset("wikipedia", language="aa", date="20220401", split="train",beam_runner='DirectRunner'). However, I end up in getting filenotfound error. I get this error for any language I try to download. It downloads the file but while saving it in hugging face cache it fails to write. This happens for any available date of any language in wikipedia dump. I had raised another issue earlier #4915 but probably was not that clear and the solution provider misunderstood my problem. Hence raising one more issue. Any help is appreciated. ## Steps to reproduce the bug ```python from datasets import load_dataset load_dataset("wikipedia", language="aa", date="20220401", split="train",beam_runner='DirectRunner') ``` ## Expected results to load the dataset ## Actual results I am pasting the error trace here: Downloading builder script: 35.9kB [00:00, ?B/s] Downloading metadata: 30.4kB [00:00, 1.94MB/s] Using custom data configuration 20220401.aa-date=20220401,language=aa Downloading and preparing dataset wikipedia/20220401.aa to C:\Users\Shilpa.cache\huggingface\datasets\wikipedia\20220401.aa-date=20220401,language=aa\2.0.0\aa542ed919df55cc5d3347f42dd4521d05ca68751f50dbc32bae2a7f1e167559... Downloading data: 100%|████████████████████████████████████████████████████████████| 11.1k/11.1k [00:00<00:00, 712kB/s] Downloading data files: 100%|████████████████████████████████████████████████████████████| 1/1 [00:02<00:00, 2.82s/it] Extracting data files: 100%|█████████████████████████████████████████████████████████████████████| 1/1 [00:00<?, ?it/s] Downloading data: 100%|███████████████████████████████████████████████████████████| 35.6k/35.6k [00:00<00:00, 84.3kB/s] Downloading data files: 100%|████████████████████████████████████████████████████████████| 1/1 [00:02<00:00, 2.93s/it] Traceback (most recent call last): File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 837, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam\runners\common.py", line 981, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam\runners\common.py", line 1571, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "G:\Python3.7\lib\site-packages\apache_beam\io\iobase.py", line 1193, in process self.writer = self.sink.open_writer(init_result, str(uuid.uuid4())) File "G:\Python3.7\lib\site-packages\apache_beam\options\value_provider.py", line 193, in _f return fnc(self, *args, **kwargs) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 202, in open_writer return FileBasedSinkWriter(self, writer_path) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 419, in init self.temp_handle = self.sink.open(temp_shard_path) File "G:\Python3.7\lib\site-packages\apache_beam\io\parquetio.py", line 553, in open self._file_handle = super().open(temp_path) File "G:\Python3.7\lib\site-packages\apache_beam\options\value_provider.py", line 193, in _f return fnc(self, *args, **kwargs) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 139, in open temp_path, self.mime_type, self.compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\filesystems.py", line 224, in create return filesystem.create(path, mime_type, compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\localfilesystem.py", line 163, in create return self._path_open(path, 'wb', mime_type, compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\localfilesystem.py", line 140, in _path_open raw_file = io.open(path, mode) FileNotFoundError: [Errno 2] No such file or directory: 'C:\Users\Shilpa\.cache\huggingface\datasets\wikipedia\20220401.aa-date=20220401,language=aa\2.0.0\aa542ed919df55cc5d3347f42dd4521d05ca68751f50dbc32bae2a7f1e167559.incomplete\beam-temp-wikipedia-train-880233e8287e11edaf9d3ca067f2714e\20a05238-6106-4420-a713-4eca6dd5959a.wikipedia-train' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "G:/abc/temp.py", line 32, in beam_runner='DirectRunner') File "G:\Python3.7\lib\site-packages\datasets\load.py", line 1751, in load_dataset use_auth_token=use_auth_token, File "G:\Python3.7\lib\site-packages\datasets\builder.py", line 705, in download_and_prepare dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "G:\Python3.7\lib\site-packages\datasets\builder.py", line 1394, in _download_and_prepare pipeline_results = pipeline.run() File "G:\Python3.7\lib\site-packages\apache_beam\pipeline.py", line 574, in run return self.runner.run_pipeline(self, self._options) File "G:\Python3.7\lib\site-packages\apache_beam\runners\direct\direct_runner.py", line 131, in run_pipeline return runner.run_pipeline(pipeline, options) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 201, in run_pipeline options) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 212, in run_via_runner_api return self.run_stages(stage_context, stages) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 443, in run_stages runner_execution_context, bundle_context_manager, bundle_input) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 776, in _execute_bundle bundle_manager)) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 1000, in _run_bundle data_input, data_output, input_timers, expected_timer_output) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 1309, in process_bundle result_future = self._worker_handler.control_conn.push(process_bundle_req) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\worker_handlers.py", line 380, in push response = self.worker.do_instruction(request) File "G:\Python3.7\lib\site-packages\apache_beam\runners\worker\sdk_worker.py", line 598, in do_instruction getattr(request, request_type), request.instruction_id) File "G:\Python3.7\lib\site-packages\apache_beam\runners\worker\sdk_worker.py", line 635, in process_bundle bundle_processor.process_bundle(instruction_id)) File "G:\Python3.7\lib\site-packages\apache_beam\runners\worker\bundle_processor.py", line 1004, in process_bundle element.data) File "G:\Python3.7\lib\site-packages\apache_beam\runners\worker\bundle_processor.py", line 227, in process_encoded self.output(decoded_value) File "apache_beam\runners\worker\operations.py", line 526, in apache_beam.runners.worker.operations.Operation.output File "apache_beam\runners\worker\operations.py", line 528, in apache_beam.runners.worker.operations.Operation.output File "apache_beam\runners\worker\operations.py", line 237, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 623, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 623, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 837, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam\runners\common.py", line 981, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 623, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 324, in apache_beam.runners.worker.operations.GeneralPurposeConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 905, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 623, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 837, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam\runners\common.py", line 981, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1507, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 837, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam\runners\common.py", line 981, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam\runners\common.py", line 1571, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "G:\Python3.7\lib\site-packages\apache_beam\io\iobase.py", line 1193, in process self.writer = self.sink.open_writer(init_result, str(uuid.uuid4())) File "G:\Python3.7\lib\site-packages\apache_beam\options\value_provider.py", line 193, in _f return fnc(self, *args, **kwargs) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 202, in open_writer return FileBasedSinkWriter(self, writer_path) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 419, in init self.temp_handle = self.sink.open(temp_shard_path) File "G:\Python3.7\lib\site-packages\apache_beam\io\parquetio.py", line 553, in open self._file_handle = super().open(temp_path) File "G:\Python3.7\lib\site-packages\apache_beam\options\value_provider.py", line 193, in _f return fnc(self, *args, **kwargs) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 139, in open temp_path, self.mime_type, self.compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\filesystems.py", line 224, in create return filesystem.create(path, mime_type, compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\localfilesystem.py", line 163, in create return self._path_open(path, 'wb', mime_type, compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\localfilesystem.py", line 140, in _path_open raw_file = io.open(path, mode) RuntimeError: FileNotFoundError: [Errno 2] No such file or directory: 'C:\Users\Shilpa\.cache\huggingface\datasets\wikipedia\20220401.aa-date=20220401,language=aa\2.0.0\aa542ed919df55cc5d3347f42dd4521d05ca68751f50dbc32bae2a7f1e167559.incomplete\beam-temp-wikipedia-train-880233e8287e11edaf9d3ca067f2714e\20a05238-6106-4420-a713-4eca6dd5959a.wikipedia-train' [while running 'train/Save to parquet/Write/WriteImpl/WriteBundles'] ## Environment info Python: 3.7.6 Windows 10 Pro datasets :2.4.0 apache_beam: 2.41.0 mwparserfromhell: 0.6.4
2022-08-31T10:53:19Z
4,916
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-08-31T09:39:25Z
https://api.github.com/repos/huggingface/datasets/issues/4916/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4916/timeline
Apache Beam unable to write the downloaded wikipedia dataset
https://api.github.com/repos/huggingface/datasets/issues/4916/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/71849081?v=4", "events_url": "https://api.github.com/users/Shilpac20/events{/privacy}", "followers_url": "https://api.github.com/users/Shilpac20/followers", "following_url": "https://api.github.com/users/Shilpac20/following{/other_user}", "gists_url": "https://api.github.com/users/Shilpac20/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Shilpac20", "id": 71849081, "login": "Shilpac20", "node_id": "MDQ6VXNlcjcxODQ5MDgx", "organizations_url": "https://api.github.com/users/Shilpac20/orgs", "received_events_url": "https://api.github.com/users/Shilpac20/received_events", "repos_url": "https://api.github.com/users/Shilpac20/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Shilpac20/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Shilpac20/subscriptions", "type": "User", "url": "https://api.github.com/users/Shilpac20" }
[]
null
completed
NONE
2022-08-31T10:53:19Z
null
I_kwDODunzps5Q41nM
[ "See:\r\n- #4915" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4916/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4916
https://github.com/huggingface/datasets/issues/4916
false
1,356,009,042
https://api.github.com/repos/huggingface/datasets/issues/4915/labels{/name}
## Describe the bug Hi, I am currently trying to download wikipedia dataset using load_dataset("wikipedia", language="aa", date="20220401", split="train",beam_runner='DirectRunner'). However, I end up in getting filenotfound error. I get this error for any language I try to download. Environment: ## Steps to reproduce the bug ```python from datasets import load_dataset load_dataset("wikipedia", language="aa", date="20220401", split="train",beam_runner='DirectRunner') ``` ## Expected results to load the dataset ## Actual results I am pasting the error trace here: Downloading builder script: 35.9kB [00:00, ?B/s] Downloading metadata: 30.4kB [00:00, 1.94MB/s] Using custom data configuration 20220401.aa-date=20220401,language=aa Downloading and preparing dataset wikipedia/20220401.aa to C:\Users\Shilpa\.cache\huggingface\datasets\wikipedia\20220401.aa-date=20220401,language=aa\2.0.0\aa542ed919df55cc5d3347f42dd4521d05ca68751f50dbc32bae2a7f1e167559... Downloading data: 100%|████████████████████████████████████████████████████████████| 11.1k/11.1k [00:00<00:00, 712kB/s] Downloading data files: 100%|████████████████████████████████████████████████████████████| 1/1 [00:02<00:00, 2.82s/it] Extracting data files: 100%|█████████████████████████████████████████████████████████████████████| 1/1 [00:00<?, ?it/s] Downloading data: 100%|███████████████████████████████████████████████████████████| 35.6k/35.6k [00:00<00:00, 84.3kB/s] Downloading data files: 100%|████████████████████████████████████████████████████████████| 1/1 [00:02<00:00, 2.93s/it] Traceback (most recent call last): File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 837, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam\runners\common.py", line 981, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam\runners\common.py", line 1571, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "G:\Python3.7\lib\site-packages\apache_beam\io\iobase.py", line 1193, in process self.writer = self.sink.open_writer(init_result, str(uuid.uuid4())) File "G:\Python3.7\lib\site-packages\apache_beam\options\value_provider.py", line 193, in _f return fnc(self, *args, **kwargs) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 202, in open_writer return FileBasedSinkWriter(self, writer_path) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 419, in __init__ self.temp_handle = self.sink.open(temp_shard_path) File "G:\Python3.7\lib\site-packages\apache_beam\io\parquetio.py", line 553, in open self._file_handle = super().open(temp_path) File "G:\Python3.7\lib\site-packages\apache_beam\options\value_provider.py", line 193, in _f return fnc(self, *args, **kwargs) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 139, in open temp_path, self.mime_type, self.compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\filesystems.py", line 224, in create return filesystem.create(path, mime_type, compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\localfilesystem.py", line 163, in create return self._path_open(path, 'wb', mime_type, compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\localfilesystem.py", line 140, in _path_open raw_file = io.open(path, mode) FileNotFoundError: [Errno 2] No such file or directory: 'C:\\Users\\Shilpa\\.cache\\huggingface\\datasets\\wikipedia\\20220401.aa-date=20220401,language=aa\\2.0.0\\aa542ed919df55cc5d3347f42dd4521d05ca68751f50dbc32bae2a7f1e167559.incomplete\\beam-temp-wikipedia-train-880233e8287e11edaf9d3ca067f2714e\\20a05238-6106-4420-a713-4eca6dd5959a.wikipedia-train' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "G:/abc/temp.py", line 32, in <module> beam_runner='DirectRunner') File "G:\Python3.7\lib\site-packages\datasets\load.py", line 1751, in load_dataset use_auth_token=use_auth_token, File "G:\Python3.7\lib\site-packages\datasets\builder.py", line 705, in download_and_prepare dl_manager=dl_manager, verify_infos=verify_infos, **download_and_prepare_kwargs File "G:\Python3.7\lib\site-packages\datasets\builder.py", line 1394, in _download_and_prepare pipeline_results = pipeline.run() File "G:\Python3.7\lib\site-packages\apache_beam\pipeline.py", line 574, in run return self.runner.run_pipeline(self, self._options) File "G:\Python3.7\lib\site-packages\apache_beam\runners\direct\direct_runner.py", line 131, in run_pipeline return runner.run_pipeline(pipeline, options) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 201, in run_pipeline options) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 212, in run_via_runner_api return self.run_stages(stage_context, stages) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 443, in run_stages runner_execution_context, bundle_context_manager, bundle_input) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 776, in _execute_bundle bundle_manager)) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 1000, in _run_bundle data_input, data_output, input_timers, expected_timer_output) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\fn_runner.py", line 1309, in process_bundle result_future = self._worker_handler.control_conn.push(process_bundle_req) File "G:\Python3.7\lib\site-packages\apache_beam\runners\portability\fn_api_runner\worker_handlers.py", line 380, in push response = self.worker.do_instruction(request) File "G:\Python3.7\lib\site-packages\apache_beam\runners\worker\sdk_worker.py", line 598, in do_instruction getattr(request, request_type), request.instruction_id) File "G:\Python3.7\lib\site-packages\apache_beam\runners\worker\sdk_worker.py", line 635, in process_bundle bundle_processor.process_bundle(instruction_id)) File "G:\Python3.7\lib\site-packages\apache_beam\runners\worker\bundle_processor.py", line 1004, in process_bundle element.data) File "G:\Python3.7\lib\site-packages\apache_beam\runners\worker\bundle_processor.py", line 227, in process_encoded self.output(decoded_value) File "apache_beam\runners\worker\operations.py", line 526, in apache_beam.runners.worker.operations.Operation.output File "apache_beam\runners\worker\operations.py", line 528, in apache_beam.runners.worker.operations.Operation.output File "apache_beam\runners\worker\operations.py", line 237, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 623, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 623, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 837, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam\runners\common.py", line 981, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 623, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 324, in apache_beam.runners.worker.operations.GeneralPurposeConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 905, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 623, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1491, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 837, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam\runners\common.py", line 981, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam\runners\common.py", line 1581, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "apache_beam\runners\common.py", line 1694, in apache_beam.runners.common._OutputHandler._write_value_to_tag File "apache_beam\runners\worker\operations.py", line 240, in apache_beam.runners.worker.operations.SingletonElementConsumerSet.receive File "apache_beam\runners\worker\operations.py", line 907, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\worker\operations.py", line 908, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam\runners\common.py", line 1419, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 1507, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam\runners\common.py", line 1417, in apache_beam.runners.common.DoFnRunner.process File "apache_beam\runners\common.py", line 837, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam\runners\common.py", line 981, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "apache_beam\runners\common.py", line 1571, in apache_beam.runners.common._OutputHandler.handle_process_outputs File "G:\Python3.7\lib\site-packages\apache_beam\io\iobase.py", line 1193, in process self.writer = self.sink.open_writer(init_result, str(uuid.uuid4())) File "G:\Python3.7\lib\site-packages\apache_beam\options\value_provider.py", line 193, in _f return fnc(self, *args, **kwargs) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 202, in open_writer return FileBasedSinkWriter(self, writer_path) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 419, in __init__ self.temp_handle = self.sink.open(temp_shard_path) File "G:\Python3.7\lib\site-packages\apache_beam\io\parquetio.py", line 553, in open self._file_handle = super().open(temp_path) File "G:\Python3.7\lib\site-packages\apache_beam\options\value_provider.py", line 193, in _f return fnc(self, *args, **kwargs) File "G:\Python3.7\lib\site-packages\apache_beam\io\filebasedsink.py", line 139, in open temp_path, self.mime_type, self.compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\filesystems.py", line 224, in create return filesystem.create(path, mime_type, compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\localfilesystem.py", line 163, in create return self._path_open(path, 'wb', mime_type, compression_type) File "G:\Python3.7\lib\site-packages\apache_beam\io\localfilesystem.py", line 140, in _path_open raw_file = io.open(path, mode) RuntimeError: FileNotFoundError: [Errno 2] No such file or directory: 'C:\\Users\\Shilpa\\.cache\\huggingface\\datasets\\wikipedia\\20220401.aa-date=20220401,language=aa\\2.0.0\\aa542ed919df55cc5d3347f42dd4521d05ca68751f50dbc32bae2a7f1e167559.incomplete\\beam-temp-wikipedia-train-880233e8287e11edaf9d3ca067f2714e\\20a05238-6106-4420-a713-4eca6dd5959a.wikipedia-train' [while running 'train/Save to parquet/Write/WriteImpl/WriteBundles'] ## Environment info Python: 3.7.6 Windows 10 Pro datasets :2.4.0 apache_beam: 2.41.0 mwparserfromhell: 0.6.4
2022-12-04T22:20:33Z
4,915
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-08-30T16:15:46Z
https://api.github.com/repos/huggingface/datasets/issues/4915/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4915/timeline
FileNotFoundError while downloading wikipedia dataset for any language
https://api.github.com/repos/huggingface/datasets/issues/4915/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/71849081?v=4", "events_url": "https://api.github.com/users/Shilpac20/events{/privacy}", "followers_url": "https://api.github.com/users/Shilpac20/followers", "following_url": "https://api.github.com/users/Shilpac20/following{/other_user}", "gists_url": "https://api.github.com/users/Shilpac20/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Shilpac20", "id": 71849081, "login": "Shilpac20", "node_id": "MDQ6VXNlcjcxODQ5MDgx", "organizations_url": "https://api.github.com/users/Shilpac20/orgs", "received_events_url": "https://api.github.com/users/Shilpac20/received_events", "repos_url": "https://api.github.com/users/Shilpac20/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Shilpac20/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Shilpac20/subscriptions", "type": "User", "url": "https://api.github.com/users/Shilpac20" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
reopened
NONE
null
null
I_kwDODunzps5Q0w5S
[ "Hi @Shilpac20,\r\n\r\nAs explained in the Wikipedia dataset card: https://huggingface.co/datasets/wikipedia\r\n> You can find the full list of languages and dates [here](https://dumps.wikimedia.org/backup-index.html).\r\n\r\nThis means that, before passing a specific date, you should first make sure it is availabl...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4915/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4915
https://github.com/huggingface/datasets/issues/4915
false
1,355,482,624
https://api.github.com/repos/huggingface/datasets/issues/4914/labels{/name}
Support streaming swda dataset.
2022-08-30T11:16:33Z
4,914
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-30T09:46:28Z
https://api.github.com/repos/huggingface/datasets/issues/4914/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4914/timeline
Support streaming swda dataset
https://api.github.com/repos/huggingface/datasets/issues/4914/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-08-30T11:14:16Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4914.diff", "html_url": "https://github.com/huggingface/datasets/pull/4914", "merged_at": "2022-08-30T11:14:15Z", "patch_url": "https://github.com/huggingface/datasets/pull/4914.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4914" }
PR_kwDODunzps4-CFyN
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4914/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4914
https://github.com/huggingface/datasets/pull/4914
true
1,355,232,007
https://api.github.com/repos/huggingface/datasets/issues/4913/labels{/name}
This PR adds the license information to `cosmos_qa` dataset, once reported via email by Yejin Choi, the dataset is licensed under CC BY 4.0. This PR also updates the citation information.
2022-08-30T09:49:31Z
4,913
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-30T06:23:19Z
https://api.github.com/repos/huggingface/datasets/issues/4913/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4913/timeline
Add license and citation information to cosmos_qa dataset
https://api.github.com/repos/huggingface/datasets/issues/4913/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-08-30T09:47:35Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4913.diff", "html_url": "https://github.com/huggingface/datasets/pull/4913", "merged_at": "2022-08-30T09:47:35Z", "patch_url": "https://github.com/huggingface/datasets/pull/4913.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4913" }
PR_kwDODunzps4-BP00
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4913/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4913
https://github.com/huggingface/datasets/pull/4913
true
1,355,078,864
https://api.github.com/repos/huggingface/datasets/issues/4912/labels{/name}
**1. Background** Huggingface datasets package advises using `map()` to process data in batches. In the example code on pretraining masked language model, they use `map()` to tokenize all data at a stroke before the train loop. The corresponding code: ``` with accelerator.main_process_first(): tokenized_datasets = raw_datasets.map( tokenize_function, batched=True, num_proc=args.preprocessing_num_workers, remove_columns=column_names, load_from_cache_file=not args.overwrite_cache, desc="Running tokenizer on every text in dataset" ) ``` **2. The problem** Thus, when I try the same pertaining code with a much larger corpus, it takes quite a long time to tokenize. Also, we can choose to tokenize data in `data-collator`. In this way, the program only tokenizes one batch in the next training step and avoids getting stuck in tokenization. **3. My question** As described above, my questions are: * **Which is better? Process in `map()` or in `data-collator`** * **Why huggingface advises `map()` function?** There should be some advantages to using `map()` Thanks for your answers!
2023-04-06T09:43:58Z
4,912
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-08-30T02:25:56Z
https://api.github.com/repos/huggingface/datasets/issues/4912/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4912/timeline
datasets map() handles all data at a stroke and takes long time
https://api.github.com/repos/huggingface/datasets/issues/4912/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/40711748?v=4", "events_url": "https://api.github.com/users/BruceStayHungry/events{/privacy}", "followers_url": "https://api.github.com/users/BruceStayHungry/followers", "following_url": "https://api.github.com/users/BruceStayHungry/following{/other_user}", "gists_url": "https://api.github.com/users/BruceStayHungry/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/BruceStayHungry", "id": 40711748, "login": "BruceStayHungry", "node_id": "MDQ6VXNlcjQwNzExNzQ4", "organizations_url": "https://api.github.com/users/BruceStayHungry/orgs", "received_events_url": "https://api.github.com/users/BruceStayHungry/received_events", "repos_url": "https://api.github.com/users/BruceStayHungry/repos", "site_admin": false, "starred_url": "https://api.github.com/users/BruceStayHungry/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceStayHungry/subscriptions", "type": "User", "url": "https://api.github.com/users/BruceStayHungry" }
[]
null
completed
NONE
2022-09-06T09:23:35Z
null
I_kwDODunzps5QxNzQ
[ "Hi ! Interesting question ;)\r\n\r\n> Which is better? Process in map() or in data-collator\r\n\r\nAs you said, both can be used in practice: map() if you want to preprocess before training, or a data-collator (or the equivalent `dataset.set_transform`) if you want to preprocess on-the-fly during training. Both op...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4912/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4912
https://github.com/huggingface/datasets/issues/4912
false
1,354,426,978
https://api.github.com/repos/huggingface/datasets/issues/4911/labels{/name}
On https://hf.co/datasets you can rename a dataset (or sometimes move it to another user/org). The website handles redirections correctly and AFAIK `datasets` does as well. However it would be nice to have an integration test to make sure we don't break support for renamed datasets. To implement this we can use the /api/repos/move endpoint on hub-ci to rename/move a repo (it is documented at https://huggingface.co/docs/hub/api)
2022-08-29T15:31:03Z
4,911
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "BDE59C", "default": false, "description": "Issues a bit more difficult than \"Good First\" issues", "id": 3761482852, "name": "good second issue", "node_id": "LA_kwDODunzps7gM6xk", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20second%20issue" } ]
2022-08-29T14:46:14Z
https://api.github.com/repos/huggingface/datasets/issues/4911/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4911/timeline
[Tests] Ensure `datasets` supports renamed repositories
https://api.github.com/repos/huggingface/datasets/issues/4911/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[]
null
null
MEMBER
null
null
I_kwDODunzps5Quupi
[ "You could also switch to using `huggingface_hub` more directly, where such a guarantee is already tested =)\r\n\r\ncc @Wauplin " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4911/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4911
https://github.com/huggingface/datasets/issues/4911
false
1,354,374,328
https://api.github.com/repos/huggingface/datasets/issues/4910/labels{/name}
## Describe the bug In `load_dataset_builder()`, `build_kwargs` and `config_kwargs` can contain the same keywords leading to a TypeError("type object got multiple values for keyword argument "xyz"). I ran into this problem with the keyword: `base_path`. It might happen with other kwargs as well. I think a quickfix would be ```python builder_cls = import_main_class(dataset_module.module_path) builder_kwargs = dataset_module.builder_kwargs data_files = builder_kwargs.pop("data_files", data_files) config_name = builder_kwargs.pop("config_name", name) hash = builder_kwargs.pop("hash") base_path = builder_kwargs.pop("base_path") ``` and then pass base_path into `builder_cls`. ## Steps to reproduce the bug ```python from datasets import load_dataset load_dataset("rotten_tomatoes", base_path="./sample_data") ``` ## Expected results The docs state: `**config_kwargs` — Keyword arguments to be passed to the [BuilderConfig](https://huggingface.co/docs/datasets/v2.4.0/en/package_reference/builder_classes#datasets.BuilderConfig) and used in the [DatasetBuilder](https://huggingface.co/docs/datasets/v2.4.0/en/package_reference/builder_classes#datasets.DatasetBuilder). So I would expect to be able to pass the base_path into `load_dataset()`. ## Actual results TypeError("type object got multiple values for keyword argument "base_path"). ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.4.0 - Platform: macOS-12.5-arm64-arm-64bit - Python version: 3.8.9 - PyArrow version: 9.0.0
2022-09-13T11:58:46Z
4,910
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" }, { "color": "7057ff", "default": true, "descript...
2022-08-29T14:11:48Z
https://api.github.com/repos/huggingface/datasets/issues/4910/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/21123710?v=4", "events_url": "https://api.github.com/users/thepurpleowl/events{/privacy}", "followers_url": "https://api.github.com/users/thepurpleowl/followers", "following_url": "https://api.github.com/users/thepurpleowl/following{/other_user}", "gists_url": "https://api.github.com/users/thepurpleowl/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/thepurpleowl", "id": 21123710, "login": "thepurpleowl", "node_id": "MDQ6VXNlcjIxMTIzNzEw", "organizations_url": "https://api.github.com/users/thepurpleowl/orgs", "received_events_url": "https://api.github.com/users/thepurpleowl/received_events", "repos_url": "https://api.github.com/users/thepurpleowl/repos", "site_admin": false, "starred_url": "https://api.github.com/users/thepurpleowl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thepurpleowl/subscriptions", "type": "User", "url": "https://api.github.com/users/thepurpleowl" }
https://api.github.com/repos/huggingface/datasets/issues/4910/timeline
Identical keywords in build_kwargs and config_kwargs lead to TypeError in load_dataset_builder()
https://api.github.com/repos/huggingface/datasets/issues/4910/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/57184353?v=4", "events_url": "https://api.github.com/users/bablf/events{/privacy}", "followers_url": "https://api.github.com/users/bablf/followers", "following_url": "https://api.github.com/users/bablf/following{/other_user}", "gists_url": "https://api.github.com/users/bablf/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/bablf", "id": 57184353, "login": "bablf", "node_id": "MDQ6VXNlcjU3MTg0MzUz", "organizations_url": "https://api.github.com/users/bablf/orgs", "received_events_url": "https://api.github.com/users/bablf/received_events", "repos_url": "https://api.github.com/users/bablf/repos", "site_admin": false, "starred_url": "https://api.github.com/users/bablf/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bablf/subscriptions", "type": "User", "url": "https://api.github.com/users/bablf" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/21123710?v=4", "events_url": "https://api.github.com/users/thepurpleowl/events{/privacy}", "followers_url": "https://api.github.com/users/thepurpleowl/followers", "following_url": "https://api.github.com/users/thepurpleowl/following{/other_use...
null
null
NONE
null
null
I_kwDODunzps5Quhy4
[ "I am getting similar error - `TypeError: type object got multiple values for keyword argument 'name'` while following this [tutorial](https://huggingface.co/docs/datasets/dataset_script#create-a-dataset-loading-script). I am getting this error with the `dataset-cli test` command.\r\n\r\n`datasets` version: 2.4.0",...
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4910/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4910
https://github.com/huggingface/datasets/issues/4910
false
1,353,997,788
https://api.github.com/repos/huggingface/datasets/issues/4909/labels{/name}
This PR updates the evaluation metadata for GLUE to: * Include defaults for all configs except `ax` (which only has a `test` split with no known labels) * Fix the default split from `test` to `validation` since `test` splits in GLUE have no labels (they're private) * Fix the `task_id` for some existing defaults cc @sashavor @douwekiela
2022-08-29T14:53:29Z
4,909
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-29T09:43:44Z
https://api.github.com/repos/huggingface/datasets/issues/4909/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4909/timeline
Update GLUE evaluation metadata
https://api.github.com/repos/huggingface/datasets/issues/4909/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
[]
null
null
MEMBER
2022-08-29T14:51:18Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4909.diff", "html_url": "https://github.com/huggingface/datasets/pull/4909", "merged_at": "2022-08-29T14:51:18Z", "patch_url": "https://github.com/huggingface/datasets/pull/4909.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4909" }
PR_kwDODunzps499Fhe
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4909/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4909
https://github.com/huggingface/datasets/pull/4909
true
1,353,995,574
https://api.github.com/repos/huggingface/datasets/issues/4908/labels{/name}
Fix missing tags in dataset cards: - asnq - clue - common_gen - cosmos_qa - guardian_authorship - hindi_discourse - py_ast - x_stance This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833 - #4891 - #4896
2022-09-22T14:35:56Z
4,908
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-29T09:41:53Z
https://api.github.com/repos/huggingface/datasets/issues/4908/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4908/timeline
Fix missing tags in dataset cards
https://api.github.com/repos/huggingface/datasets/issues/4908/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-08-29T16:13:07Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4908.diff", "html_url": "https://github.com/huggingface/datasets/pull/4908", "merged_at": "2022-08-29T16:13:07Z", "patch_url": "https://github.com/huggingface/datasets/pull/4908.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4908" }
PR_kwDODunzps499FDS
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4908/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4908
https://github.com/huggingface/datasets/pull/4908
true
1,353,808,348
https://api.github.com/repos/huggingface/datasets/issues/4907/labels{/name}
## Describe the bug I got `'NoneType' object is not callable` error while calling the swda datasets. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("swda") ``` ## Expected results Run without error ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.4.0 - Python version: 3.8.10
2022-08-30T14:43:41Z
4,907
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-08-29T07:05:20Z
https://api.github.com/repos/huggingface/datasets/issues/4907/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4907/timeline
None Type error for swda datasets
https://api.github.com/repos/huggingface/datasets/issues/4907/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8229163?v=4", "events_url": "https://api.github.com/users/hannan72/events{/privacy}", "followers_url": "https://api.github.com/users/hannan72/followers", "following_url": "https://api.github.com/users/hannan72/following{/other_user}", "gists_url": "https://api.github.com/users/hannan72/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/hannan72", "id": 8229163, "login": "hannan72", "node_id": "MDQ6VXNlcjgyMjkxNjM=", "organizations_url": "https://api.github.com/users/hannan72/orgs", "received_events_url": "https://api.github.com/users/hannan72/received_events", "repos_url": "https://api.github.com/users/hannan72/repos", "site_admin": false, "starred_url": "https://api.github.com/users/hannan72/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hannan72/subscriptions", "type": "User", "url": "https://api.github.com/users/hannan72" }
[]
null
completed
NONE
2022-08-30T14:43:41Z
null
I_kwDODunzps5QsXnc
[ "Thanks for reporting @hannan72 ! I couldn't reproduce the error on my side, can you share the full stack trace please ?", "Thanks a lot for your response @lhoestq \r\nThe problem is solved accidentally today and I don't know exactly why it was happened yesterday.\r\nThe issue can be closed.", "Ok, let us know ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4907/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4907
https://github.com/huggingface/datasets/issues/4907
false
1,353,223,925
https://api.github.com/repos/huggingface/datasets/issues/4906/labels{/name}
## Describe the bug A clear and concise description of what the bug is. Not able to import datasets ## Steps to reproduce the bug ```python # Sample code to reproduce the bug import os os.environ["WANDB_API_KEY"] = "0" ## to silence warning import numpy as np import random import sklearn import matplotlib.pyplot as plt import pandas as pd import sys import tensorflow as tf import plotly.express as px import transformers import tokenizers import nlp as nlp import utils import datasets ``` ## Expected results A clear and concise description of the expected results. import should work normal ## Actual results Specify the actual results or traceback. --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) <ipython-input-21-b3b5b0b62103> in <module> 13 import nlp as nlp 14 import utils ---> 15 import datasets ~\anaconda3\lib\site-packages\datasets\__init__.py in <module> 44 from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled 45 from .info import DatasetInfo, MetricInfo ---> 46 from .inspect import ( 47 get_dataset_config_info, 48 get_dataset_config_names, ~\anaconda3\lib\site-packages\datasets\inspect.py in <module> 28 from .download.streaming_download_manager import StreamingDownloadManager 29 from .info import DatasetInfo ---> 30 from .load import dataset_module_factory, import_main_class, load_dataset_builder, metric_module_factory 31 from .utils.file_utils import relative_to_absolute_path 32 from .utils.logging import get_logger ~\anaconda3\lib\site-packages\datasets\load.py in <module> 53 from .iterable_dataset import IterableDataset 54 from .metric import Metric ---> 55 from .packaged_modules import ( 56 _EXTENSION_TO_MODULE, 57 _MODULE_SUPPORTS_METADATA, ~\anaconda3\lib\site-packages\datasets\packaged_modules\__init__.py in <module> 4 from typing import List 5 ----> 6 from .csv import csv 7 from .imagefolder import imagefolder 8 from .json import json ~\anaconda3\lib\site-packages\datasets\packaged_modules\csv\csv.py in <module> 13 14 ---> 15 logger = datasets.utils.logging.get_logger(__name__) 16 17 _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS = ["names", "prefix"] AttributeError: partially initialized module 'datasets' has no attribute 'utils' (most likely due to a circular import) ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> Copy-and-paste the text below in your GitHub issue. - `datasets` version: 2.4.0 - Platform: Windows-10-10.0.22000-SP0 - Python version: 3.8.8 - PyArrow version: 9.0.0 - Pandas version: 1.2.4
2023-10-27T20:08:28Z
4,906
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-08-28T02:23:24Z
https://api.github.com/repos/huggingface/datasets/issues/4906/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4906/timeline
Can't import datasets AttributeError: partially initialized module 'datasets' has no attribute 'utils' (most likely due to a circular import)
https://api.github.com/repos/huggingface/datasets/issues/4906/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/63536981?v=4", "events_url": "https://api.github.com/users/OPterminator/events{/privacy}", "followers_url": "https://api.github.com/users/OPterminator/followers", "following_url": "https://api.github.com/users/OPterminator/following{/other_user}", "gists_url": "https://api.github.com/users/OPterminator/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/OPterminator", "id": 63536981, "login": "OPterminator", "node_id": "MDQ6VXNlcjYzNTM2OTgx", "organizations_url": "https://api.github.com/users/OPterminator/orgs", "received_events_url": "https://api.github.com/users/OPterminator/received_events", "repos_url": "https://api.github.com/users/OPterminator/repos", "site_admin": false, "starred_url": "https://api.github.com/users/OPterminator/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/OPterminator/subscriptions", "type": "User", "url": "https://api.github.com/users/OPterminator" }
[]
null
completed
NONE
2022-10-03T12:22:50Z
null
I_kwDODunzps5QqI71
[ "Thanks for reporting, @OPterminator.\r\n\r\nHowever, we are not able to reproduce this issue.\r\n\r\nThere might be 2 reasons why you get this exception:\r\n- Either the name of your local Python file: if it is called `datasets.py` this could generate a circular import when trying to import the Hugging Face `data...
{ "+1": 5, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 5, "url": "https://api.github.com/repos/huggingface/datasets/issues/4906/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4906
https://github.com/huggingface/datasets/issues/4906
false
1,353,002,837
https://api.github.com/repos/huggingface/datasets/issues/4904/labels{/name}
We define the keys for the `_DL_URLS` of the dev split as `dev.clean` and `dev.other`: https://github.com/huggingface/datasets/blob/2e7142a3c6500b560da45e8d5128e320a09fcbd4/datasets/librispeech_asr/librispeech_asr.py#L60-L61 These keys get forwarded to the `dl_manager` and thus the `local_extracted_archive`. However, when calling `SplitGenerator` for the dev sets, we query the `local_extracted_archive` keys `validation.clean` and `validation.other`: https://github.com/huggingface/datasets/blob/2e7142a3c6500b560da45e8d5128e320a09fcbd4/datasets/librispeech_asr/librispeech_asr.py#L212 https://github.com/huggingface/datasets/blob/2e7142a3c6500b560da45e8d5128e320a09fcbd4/datasets/librispeech_asr/librispeech_asr.py#L219 The consequence of this is that the `local_extracted_archive` arg passed to `_generate_examples` is always `None`, as the keys `validation.clean` and `validation.other` do not exists in the `local_extracted_archive`. When defining the `audio_file` in `_generate_examples`, since `local_extracted_archive` is always `None`, we always omit the `local_extracted_archive` path from the `audio_file` path, **even** if in non-streaming mode: https://github.com/huggingface/datasets/blob/2e7142a3c6500b560da45e8d5128e320a09fcbd4/datasets/librispeech_asr/librispeech_asr.py#L259-L263 Thus, `audio_file` will only ever be the streaming path (`audio_file`, not `os.path.join(local_extracted_archive, audio_file)`). This PR fixes the `.get()` keys for the `local_extracted_archive` for the dev splits.
2022-08-30T10:06:21Z
4,904
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-27T10:04:57Z
https://api.github.com/repos/huggingface/datasets/issues/4904/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4904/timeline
[LibriSpeech] Fix dev split local_extracted_archive for 'all' config
https://api.github.com/repos/huggingface/datasets/issues/4904/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/93869735?v=4", "events_url": "https://api.github.com/users/sanchit-gandhi/events{/privacy}", "followers_url": "https://api.github.com/users/sanchit-gandhi/followers", "following_url": "https://api.github.com/users/sanchit-gandhi/following{/other_user}", "gists_url": "https://api.github.com/users/sanchit-gandhi/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sanchit-gandhi", "id": 93869735, "login": "sanchit-gandhi", "node_id": "U_kgDOBZhWpw", "organizations_url": "https://api.github.com/users/sanchit-gandhi/orgs", "received_events_url": "https://api.github.com/users/sanchit-gandhi/received_events", "repos_url": "https://api.github.com/users/sanchit-gandhi/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sanchit-gandhi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sanchit-gandhi/subscriptions", "type": "User", "url": "https://api.github.com/users/sanchit-gandhi" }
[]
null
null
CONTRIBUTOR
2022-08-30T10:03:25Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4904.diff", "html_url": "https://github.com/huggingface/datasets/pull/4904", "merged_at": "2022-08-30T10:03:25Z", "patch_url": "https://github.com/huggingface/datasets/pull/4904.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4904" }
PR_kwDODunzps4959Ad
[ "_The documentation is not available anymore as the PR was closed or merged._", "This PR fixes a bug introduced in:\r\n- #4184" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4904/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4904
https://github.com/huggingface/datasets/pull/4904
true
1,352,539,075
https://api.github.com/repos/huggingface/datasets/issues/4903/labels{/name}
Fix CI so that it reports defaults (failed and error) besides the custom (xfailed and xpassed) in the test summary. This PR fixes a regression introduced by: - #4845 This introduced the reporting of xfailed and xpassed, but wrongly removed the reporting of the defaults failed and error.
2022-08-26T17:49:33Z
4,903
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-26T17:16:30Z
https://api.github.com/repos/huggingface/datasets/issues/4903/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4903/timeline
Fix CI reporting
https://api.github.com/repos/huggingface/datasets/issues/4903/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-08-26T17:46:59Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4903.diff", "html_url": "https://github.com/huggingface/datasets/pull/4903", "merged_at": "2022-08-26T17:46:59Z", "patch_url": "https://github.com/huggingface/datasets/pull/4903.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4903" }
PR_kwDODunzps494aud
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4903/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4903
https://github.com/huggingface/datasets/pull/4903
true
1,352,469,196
https://api.github.com/repos/huggingface/datasets/issues/4902/labels{/name}
Currently, if a dataset has no configuration, a default configuration is created from the dataset name. For example, for a dataset loaded from the hub repository, such as https://huggingface.co/datasets/user/dataset (repo id is `user/dataset`), the default configuration will be `user--dataset`. It might be easier to handle to set it to `default`, or another reserved word.
2023-07-24T21:15:31Z
4,902
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" }, { "color": "d876e3", "default": true...
2022-08-26T16:16:22Z
https://api.github.com/repos/huggingface/datasets/issues/4902/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4902/timeline
Name the default config `default`
https://api.github.com/repos/huggingface/datasets/issues/4902/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
[]
null
completed
CONTRIBUTOR
2023-07-24T21:15:31Z
null
I_kwDODunzps5QnQrM
[ "Addressed in #5331." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 1, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/4902/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4902
https://github.com/huggingface/datasets/issues/4902
false
1,352,438,915
https://api.github.com/repos/huggingface/datasets/issues/4901/labels{/name}
This PRs raises a specific `ManualDownloadError` when `get_dataset_config_info` is called for a dataset that requires manual download. Related to: - #4898 CC: @severo
2022-08-30T10:42:21Z
4,901
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-26T15:45:56Z
https://api.github.com/repos/huggingface/datasets/issues/4901/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4901/timeline
Raise ManualDownloadError from get_dataset_config_info
https://api.github.com/repos/huggingface/datasets/issues/4901/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-08-30T10:40:04Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4901.diff", "html_url": "https://github.com/huggingface/datasets/pull/4901", "merged_at": "2022-08-30T10:40:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/4901.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4901" }
PR_kwDODunzps494FNX
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4901/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4901
https://github.com/huggingface/datasets/pull/4901
true
1,352,405,855
https://api.github.com/repos/huggingface/datasets/issues/4900/labels{/name}
### Link _No response_ ### Description _No response_ ### Owner _No response_
2023-07-24T15:42:09Z
4,900
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-08-26T15:15:44Z
https://api.github.com/repos/huggingface/datasets/issues/4900/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4900/timeline
Dataset Viewer issue for asaxena1990/Dummy_dataset
https://api.github.com/repos/huggingface/datasets/issues/4900/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/56627657?v=4", "events_url": "https://api.github.com/users/ankurcl/events{/privacy}", "followers_url": "https://api.github.com/users/ankurcl/followers", "following_url": "https://api.github.com/users/ankurcl/following{/other_user}", "gists_url": "https://api.github.com/users/ankurcl/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ankurcl", "id": 56627657, "login": "ankurcl", "node_id": "MDQ6VXNlcjU2NjI3NjU3", "organizations_url": "https://api.github.com/users/ankurcl/orgs", "received_events_url": "https://api.github.com/users/ankurcl/received_events", "repos_url": "https://api.github.com/users/ankurcl/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ankurcl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ankurcl/subscriptions", "type": "User", "url": "https://api.github.com/users/ankurcl" }
[]
null
completed
NONE
2023-07-24T15:42:09Z
null
I_kwDODunzps5QnBNf
[ "Seems to be linked to the use of the undocumented `_resolve_features` method in the dataset viewer backend:\r\n\r\n```\r\n>>> from datasets import load_dataset\r\n>>> dataset = load_dataset(\"asaxena1990/Dummy_dataset\", name=\"asaxena1990--Dummy_dataset\", split=\"train\", streaming=True)\r\nUsing custom data con...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4900/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4900
https://github.com/huggingface/datasets/issues/4900
false
1,352,031,286
https://api.github.com/repos/huggingface/datasets/issues/4899/labels{/name}
This PR fixes the removal of 2 language tags done by: - #4882 The tags are: - "code": this is not a IANA tag but needed - "und": this is one of the special scoped tags removed by 0d53202b9abce6fd0358cb00d06fcfd904b875af - used in "mc4" and "udhr" datasets
2022-08-26T10:27:18Z
4,899
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-26T09:48:57Z
https://api.github.com/repos/huggingface/datasets/issues/4899/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4899/timeline
Re-add code and und language tags
https://api.github.com/repos/huggingface/datasets/issues/4899/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-08-26T10:24:20Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4899.diff", "html_url": "https://github.com/huggingface/datasets/pull/4899", "merged_at": "2022-08-26T10:24:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/4899.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4899" }
PR_kwDODunzps492uTO
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4899/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4899
https://github.com/huggingface/datasets/pull/4899
true
1,351,851,254
https://api.github.com/repos/huggingface/datasets/issues/4898/labels{/name}
### Link _No response_ ### Description _No response_ ### Owner _No response_
2022-10-03T12:40:28Z
4,898
null
https://api.github.com/repos/huggingface/datasets
null
[]
2022-08-26T07:12:05Z
https://api.github.com/repos/huggingface/datasets/issues/4898/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
https://api.github.com/repos/huggingface/datasets/issues/4898/timeline
Dataset Viewer issue for timit_asr
https://api.github.com/repos/huggingface/datasets/issues/4898/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/91126978?v=4", "events_url": "https://api.github.com/users/InayatUllah932/events{/privacy}", "followers_url": "https://api.github.com/users/InayatUllah932/followers", "following_url": "https://api.github.com/users/InayatUllah932/following{/other_user}", "gists_url": "https://api.github.com/users/InayatUllah932/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/InayatUllah932", "id": 91126978, "login": "InayatUllah932", "node_id": "MDQ6VXNlcjkxMTI2OTc4", "organizations_url": "https://api.github.com/users/InayatUllah932/orgs", "received_events_url": "https://api.github.com/users/InayatUllah932/received_events", "repos_url": "https://api.github.com/users/InayatUllah932/repos", "site_admin": false, "starred_url": "https://api.github.com/users/InayatUllah932/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/InayatUllah932/subscriptions", "type": "User", "url": "https://api.github.com/users/InayatUllah932" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/o...
null
completed
NONE
2022-10-03T12:40:27Z
null
I_kwDODunzps5Qk5z2
[ "Yes, the dataset viewer is based on `datasets`, and the following does not work:\r\n\r\n```\r\n>>> from datasets import get_dataset_split_names\r\n>>> get_dataset_split_names('timit_asr')\r\nDownloading builder script: 7.48kB [00:00, 6.69MB/s]\r\nTraceback (most recent call last):\r\n File \"/home/slesage/hf/data...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4898/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4898
https://github.com/huggingface/datasets/issues/4898
false
1,351,784,727
https://api.github.com/repos/huggingface/datasets/issues/4897/labels{/name}
Checking the large file in disk, and found the large cache file in the cifar10 data directory: ![image](https://user-images.githubusercontent.com/18533904/186830449-ba96cdeb-0fe8-4543-994d-2abe7145933f.png) As we know, the size of cifar10 dataset is ~130MB, but the cache file has almost 30GB size, there may be some problems here.
2022-09-18T05:07:52Z
4,897
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-08-26T05:51:16Z
https://api.github.com/repos/huggingface/datasets/issues/4897/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4897/timeline
datasets generate large arrow file
https://api.github.com/repos/huggingface/datasets/issues/4897/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/18533904?v=4", "events_url": "https://api.github.com/users/jax11235/events{/privacy}", "followers_url": "https://api.github.com/users/jax11235/followers", "following_url": "https://api.github.com/users/jax11235/following{/other_user}", "gists_url": "https://api.github.com/users/jax11235/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jax11235", "id": 18533904, "login": "jax11235", "node_id": "MDQ6VXNlcjE4NTMzOTA0", "organizations_url": "https://api.github.com/users/jax11235/orgs", "received_events_url": "https://api.github.com/users/jax11235/received_events", "repos_url": "https://api.github.com/users/jax11235/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jax11235/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jax11235/subscriptions", "type": "User", "url": "https://api.github.com/users/jax11235" }
[]
null
completed
NONE
2022-09-18T05:07:52Z
null
I_kwDODunzps5QkpkX
[ "Hi ! The cache files are the results of all the transforms you applied to the dataset using `map` for example.\r\nDid you run a transform that could potentially blow up the size of the dataset ?", "@lhoestq,\r\nI don't remember, but I can't imagine what kind of transform may generate data that grow over 200 time...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4897/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4897
https://github.com/huggingface/datasets/issues/4897
false
1,351,180,409
https://api.github.com/repos/huggingface/datasets/issues/4896/labels{/name}
Fix missing tags in dataset cards: - anli - coarse_discourse - commonsense_qa - cos_e - ilist - lc_quad - web_questions - xsum This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833 - #4891
2022-09-22T14:37:16Z
4,896
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-25T16:41:43Z
https://api.github.com/repos/huggingface/datasets/issues/4896/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4896/timeline
Fix missing tags in dataset cards
https://api.github.com/repos/huggingface/datasets/issues/4896/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-08-26T04:41:48Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4896.diff", "html_url": "https://github.com/huggingface/datasets/pull/4896", "merged_at": "2022-08-26T04:41:48Z", "patch_url": "https://github.com/huggingface/datasets/pull/4896.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4896" }
PR_kwDODunzps49z4fU
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4896/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4896
https://github.com/huggingface/datasets/pull/4896
true
1,350,798,527
https://api.github.com/repos/huggingface/datasets/issues/4895/labels{/name}
## Describe the bug The `datasets.load_dataset` returns a `ValueError: Unknown split "validation". Should be one of ['train', 'test'].` when running `load_dataset(local_data_dir_path, split="validation")` even if the `validation` sub-directory exists in the local data path. The data directories are as follows and attached to this issue: ``` test_data1 |_ train |_ 1012.png |_ metadata.jsonl ... |_ test ... |_ validation |_ 234.png |_ metadata.jsonl ... test_data2 |_ train |_ train_1012.png |_ metadata.jsonl ... |_ test ... |_ validation |_ val_234.png |_ metadata.jsonl ... ``` They contain the same image files and `metadata.jsonl` but the images in `test_data2` have the split names prepended i.e. `train_1012.png, val_234.png` and the images in `test_data1` do not have the split names prepended to the image names i.e. `1012.png, 234.png` I actually saw in another issue `val` was not recognized as a split name but here I would expect the files to take the split from the parent directory name i.e. val should become part of the validation split? ## Steps to reproduce the bug ```python import datasets datasets.logging.set_verbosity_error() from datasets import load_dataset, get_dataset_split_names # the following only finds train, validation and test splits correctly path = "./test_data1" print("######################", get_dataset_split_names(path), "######################") dataset_list = [] for spt in ["train", "test", "validation"]: dataset = load_dataset(path, split=spt) dataset_list.append(dataset) # the following only finds train and test splits path = "./test_data2" print("######################", get_dataset_split_names(path), "######################") dataset_list = [] for spt in ["train", "test", "validation"]: dataset = load_dataset(path, split=spt) dataset_list.append(dataset) ``` ## Expected results ``` ###################### ['train', 'test', 'validation'] ###################### ###################### ['train', 'test', 'validation'] ###################### ``` ## Actual results ``` Traceback (most recent call last): File "test_data_loader.py", line 11, in <module> dataset = load_dataset(path, split=spt) File "/home/venv/lib/python3.8/site-packages/datasets/load.py", line 1758, in load_dataset ds = builder_instance.as_dataset(split=split, ignore_verifications=ignore_verifications, in_memory=keep_in_memory) File "/home/venv/lib/python3.8/site-packages/datasets/builder.py", line 893, in as_dataset datasets = map_nested( File "/home/venv/lib/python3.8/site-packages/datasets/utils/py_utils.py", line 385, in map_nested return function(data_struct) File "/home/venv/lib/python3.8/site-packages/datasets/builder.py", line 924, in _build_single_dataset ds = self._as_dataset( File "/home/venv/lib/python3.8/site-packages/datasets/builder.py", line 993, in _as_dataset dataset_kwargs = ArrowReader(self._cache_dir, self.info).read( File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 211, in read files = self.get_file_instructions(name, instructions, split_infos) File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 184, in get_file_instructions file_instructions = make_file_instructions( File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 107, in make_file_instructions absolute_instructions = instruction.to_absolute(name2len) File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 616, in to_absolute return [_rel_to_abs_instr(rel_instr, name2len) for rel_instr in self._relative_instructions] File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 616, in <listcomp> return [_rel_to_abs_instr(rel_instr, name2len) for rel_instr in self._relative_instructions] File "/home/venv/lib/python3.8/site-packages/datasets/arrow_reader.py", line 433, in _rel_to_abs_instr raise ValueError(f'Unknown split "{split}". Should be one of {list(name2len)}.') ValueError: Unknown split "validation". Should be one of ['train', 'test']. ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: - Platform: Linux Ubuntu 18.04 - Python version: 3.8.12 - PyArrow version: 9.0.0 Data files [test_data1.zip](https://github.com/huggingface/datasets/files/9424463/test_data1.zip) [test_data2.zip](https://github.com/huggingface/datasets/files/9424468/test_data2.zip)
2024-03-26T16:47:48Z
4,895
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-08-25T12:11:00Z
https://api.github.com/repos/huggingface/datasets/issues/4895/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4895/timeline
load_dataset method returns Unknown split "validation" even if this dir exists
https://api.github.com/repos/huggingface/datasets/issues/4895/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/13418507?v=4", "events_url": "https://api.github.com/users/SamSamhuns/events{/privacy}", "followers_url": "https://api.github.com/users/SamSamhuns/followers", "following_url": "https://api.github.com/users/SamSamhuns/following{/other_user}", "gists_url": "https://api.github.com/users/SamSamhuns/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/SamSamhuns", "id": 13418507, "login": "SamSamhuns", "node_id": "MDQ6VXNlcjEzNDE4NTA3", "organizations_url": "https://api.github.com/users/SamSamhuns/orgs", "received_events_url": "https://api.github.com/users/SamSamhuns/received_events", "repos_url": "https://api.github.com/users/SamSamhuns/repos", "site_admin": false, "starred_url": "https://api.github.com/users/SamSamhuns/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SamSamhuns/subscriptions", "type": "User", "url": "https://api.github.com/users/SamSamhuns" }
[]
null
completed
NONE
2022-09-29T08:07:50Z
null
I_kwDODunzps5Qg4y_
[ "I don't know the main problem but it looks like, it is ignoring the last directory in your case. So, create a directory called 'zzz' in the same folder as train, validation and test. if it doesn't work, create a directory called \"aaa\". It worked for me.\r\n", "@SamSamhuns could you please try to load it with t...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4895/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4895
https://github.com/huggingface/datasets/issues/4895
false
1,350,667,270
https://api.github.com/repos/huggingface/datasets/issues/4894/labels{/name}
This PR adds the citation information to `makhzan` dataset, once they have replied to our request for that information: - https://github.com/zeerakahmed/makhzan/issues/43
2022-08-30T06:21:54Z
4,894
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-25T10:16:40Z
https://api.github.com/repos/huggingface/datasets/issues/4894/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4894/timeline
Add citation information to makhzan dataset
https://api.github.com/repos/huggingface/datasets/issues/4894/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-08-25T13:19:41Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4894.diff", "html_url": "https://github.com/huggingface/datasets/pull/4894", "merged_at": "2022-08-25T13:19:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/4894.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4894" }
PR_kwDODunzps49yIvr
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4894/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4894
https://github.com/huggingface/datasets/pull/4894
true
1,350,655,674
https://api.github.com/repos/huggingface/datasets/issues/4893/labels{/name}
In https://github.com/huggingface/datasets/pull/4831 @ylacombe added an oversampling strategy for `interleave_datasets`. However right now it doesn't work for datasets loaded using `load_dataset(..., streaming=True)`, which are `IterableDataset` objects. It would be nice to expand `interleave_datasets` for iterable datasets as well to support this oversampling strategy ```python >>> from datasets.iterable_dataset import IterableDataset, ExamplesIterable >>> d1 = IterableDataset(ExamplesIterable(lambda: [(yield i, {"a": i}) for i in [0, 1, 2]], {})) >>> d2 = IterableDataset(ExamplesIterable(lambda: [(yield i, {"a": i}) for i in [10, 11, 12, 13]], {})) >>> d3 = IterableDataset(ExamplesIterable(lambda: [(yield i, {"a": i}) for i in [20, 21, 22, 23, 24]], {})) >>> dataset = interleave_datasets([d1, d2, d3]) # is supported >>> [x["a"] for x in dataset] [0, 10, 20, 1, 11, 21, 2, 12, 22] >>> dataset = interleave_datasets([d1, d2, d3], stopping_strategy="all_exhausted") # is not supported yet >>> [x["a"] for x in dataset] [0, 10, 20, 1, 11, 21, 2, 12, 22, 0, 13, 23, 1, 0, 24] ``` This can be implemented by adding the strategy to both `CyclingMultiSourcesExamplesIterable` and `RandomlyCyclingMultiSourcesExamplesIterable` used in `_interleave_iterable_datasets` in `iterable_dataset.py` I would be happy to share some guidance if anyone would like to give it a shot :)
2022-10-03T12:37:46Z
4,893
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "BDE59C", "default": false, "description": "Issues a bit more difficult than \"Good First\" issues", "id": 3761482852, "name": "good second issue", "node_id": "LA_kwDODunzps7gM6xk", "url": "https://api.github.com/repos/huggingface/datasets/labels/good%20second%20issue" } ]
2022-08-25T10:06:55Z
https://api.github.com/repos/huggingface/datasets/issues/4893/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/52246514?v=4", "events_url": "https://api.github.com/users/ylacombe/events{/privacy}", "followers_url": "https://api.github.com/users/ylacombe/followers", "following_url": "https://api.github.com/users/ylacombe/following{/other_user}", "gists_url": "https://api.github.com/users/ylacombe/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ylacombe", "id": 52246514, "login": "ylacombe", "node_id": "MDQ6VXNlcjUyMjQ2NTE0", "organizations_url": "https://api.github.com/users/ylacombe/orgs", "received_events_url": "https://api.github.com/users/ylacombe/received_events", "repos_url": "https://api.github.com/users/ylacombe/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ylacombe/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ylacombe/subscriptions", "type": "User", "url": "https://api.github.com/users/ylacombe" }
https://api.github.com/repos/huggingface/datasets/issues/4893/timeline
Oversampling strategy for iterable datasets in `interleave_datasets`
https://api.github.com/repos/huggingface/datasets/issues/4893/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/52246514?v=4", "events_url": "https://api.github.com/users/ylacombe/events{/privacy}", "followers_url": "https://api.github.com/users/ylacombe/followers", "following_url": "https://api.github.com/users/ylacombe/following{/other_user}", "gi...
null
completed
MEMBER
2022-10-03T12:37:46Z
null
I_kwDODunzps5QgV66
[ "Hi @lhoestq,\r\nI plunged into the code and it should be manageable for me to work on it!\r\n#take\r\n\r\nAlso, setting `d1`, `d2` and `d3` as you did raised a `SyntaxError: 'yield' inside list comprehension` for me, on Python 3.8.10.\r\nThe following snippet works for me though:\r\n```\r\nd1 = IterableDataset(Exa...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4893/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4893
https://github.com/huggingface/datasets/issues/4893
false
1,350,636,499
https://api.github.com/repos/huggingface/datasets/issues/4892/labels{/name}
This PR adds the citation information to `ro_sts_parallel` and `ro_sts_parallel` datasets, once they have replied our request for that information: - https://github.com/dumitrescustefan/RO-STS/issues/4
2022-08-25T10:49:56Z
4,892
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-25T09:51:06Z
https://api.github.com/repos/huggingface/datasets/issues/4892/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4892/timeline
Add citation to ro_sts and ro_sts_parallel datasets
https://api.github.com/repos/huggingface/datasets/issues/4892/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-08-25T10:49:56Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4892.diff", "html_url": "https://github.com/huggingface/datasets/pull/4892", "merged_at": "2022-08-25T10:49:56Z", "patch_url": "https://github.com/huggingface/datasets/pull/4892.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4892" }
PR_kwDODunzps49yCD3
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4892). All of your documentation changes will be reflected on that endpoint." ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4892/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4892
https://github.com/huggingface/datasets/pull/4892
true
1,350,589,813
https://api.github.com/repos/huggingface/datasets/issues/4891/labels{/name}
Fix missing tags in dataset cards: - aslg_pc12 - librispeech_lm - mwsc - opus100 - qasc - quail - squadshifts - winograd_wsc This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task. Related to: - #4833
2022-09-22T14:39:02Z
4,891
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-25T09:14:17Z
https://api.github.com/repos/huggingface/datasets/issues/4891/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4891/timeline
Fix missing tags in dataset cards
https://api.github.com/repos/huggingface/datasets/issues/4891/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "events_url": "https://api.github.com/users/albertvillanova/events{/privacy}", "followers_url": "https://api.github.com/users/albertvillanova/followers", "following_url": "https://api.github.com/users/albertvillanova/following{/other_user}", "gists_url": "https://api.github.com/users/albertvillanova/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/albertvillanova", "id": 8515462, "login": "albertvillanova", "node_id": "MDQ6VXNlcjg1MTU0NjI=", "organizations_url": "https://api.github.com/users/albertvillanova/orgs", "received_events_url": "https://api.github.com/users/albertvillanova/received_events", "repos_url": "https://api.github.com/users/albertvillanova/repos", "site_admin": false, "starred_url": "https://api.github.com/users/albertvillanova/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/albertvillanova/subscriptions", "type": "User", "url": "https://api.github.com/users/albertvillanova" }
[]
null
null
MEMBER
2022-08-25T13:43:34Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4891.diff", "html_url": "https://github.com/huggingface/datasets/pull/4891", "merged_at": "2022-08-25T13:43:34Z", "patch_url": "https://github.com/huggingface/datasets/pull/4891.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4891" }
PR_kwDODunzps49x382
[]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4891/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4891
https://github.com/huggingface/datasets/pull/4891
true
1,350,578,029
https://api.github.com/repos/huggingface/datasets/issues/4890/labels{/name}
As discussed in #4885 I initially added this bit at the end, thinking filling this field was necessary as it is done in from_dict. However, it seems the constructor takes care of filling info when it is empty. ``` if info.features is None: info.features = Features( { col: generate_from_arrow_type(coldata.type) for col, coldata in zip(pa_table.column_names, pa_table.columns) } ) ```
2022-09-02T10:22:59Z
4,890
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-25T09:05:58Z
https://api.github.com/repos/huggingface/datasets/issues/4890/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4890/timeline
add Dataset.from_list
https://api.github.com/repos/huggingface/datasets/issues/4890/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/48946947?v=4", "events_url": "https://api.github.com/users/sanderland/events{/privacy}", "followers_url": "https://api.github.com/users/sanderland/followers", "following_url": "https://api.github.com/users/sanderland/following{/other_user}", "gists_url": "https://api.github.com/users/sanderland/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sanderland", "id": 48946947, "login": "sanderland", "node_id": "MDQ6VXNlcjQ4OTQ2OTQ3", "organizations_url": "https://api.github.com/users/sanderland/orgs", "received_events_url": "https://api.github.com/users/sanderland/received_events", "repos_url": "https://api.github.com/users/sanderland/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sanderland/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sanderland/subscriptions", "type": "User", "url": "https://api.github.com/users/sanderland" }
[]
null
null
CONTRIBUTOR
2022-09-02T10:20:33Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4890.diff", "html_url": "https://github.com/huggingface/datasets/pull/4890", "merged_at": "2022-09-02T10:20:33Z", "patch_url": "https://github.com/huggingface/datasets/pull/4890.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4890" }
PR_kwDODunzps49x1YC
[ "_The documentation is not available anymore as the PR was closed or merged._", "@albertvillanova it seems tests fail on pyarrow 6, perhaps from_pylist is a v7 method? How do you usually handle these version differences?\r\nAdded something that at least works" ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4890/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4890
https://github.com/huggingface/datasets/pull/4890
true
1,349,758,525
https://api.github.com/repos/huggingface/datasets/issues/4889/labels{/name}
## Describe the bug When loading Common Voice with torchaudio 0.11.0 the results are different to 0.12.1 which leads to problems in transformers see: https://github.com/huggingface/transformers/pull/18749 ## Steps to reproduce the bug If you run the following code once with `torchaudio==0.11.0+cu102` and `torchaudio==0.12.1+cu102` you can see that the tensors differ. This is a pretty big breaking change and makes some integration tests fail in Transformers. ```python #!/usr/bin/env python3 from datasets import load_dataset import datasets import numpy as np import torch import torchaudio print("torch vesion", torch.__version__) print("torchaudio vesion", torchaudio.__version__) save_audio = True load_audios = False if save_audio: ds = load_dataset("common_voice", "en", split="train", streaming=True) ds = ds.cast_column("audio", datasets.Audio(sampling_rate=16_000)) ds_iter = iter(ds) sample = next(ds_iter) np.save(f"audio_sample_{torch.__version__}", sample["audio"]["array"]) print(sample["audio"]["array"]) if load_audios: array_torch_11 = np.load("/home/patrick/audio_sample_1.11.0+cu102.npy") print("Array 11 Shape", array_torch_11.shape) print("Array 11 abs sum", np.sum(np.abs(array_torch_11))) array_torch_12 = np.load("/home/patrick/audio_sample_1.12.1+cu102.npy") print("Array 12 Shape", array_torch_12.shape) print("Array 12 abs sum", np.sum(np.abs(array_torch_12))) ``` Having saved the tensors the print output yields: ``` torch vesion 1.12.1+cu102 torchaudio vesion 0.12.1+cu102 Array 11 Shape (122880,) Array 11 abs sum 1396.4988 Array 12 Shape (123264,) Array 12 abs sum 1396.5193 ``` ## Expected results torchaudio 11.0 and 12.1 should yield same results. ## Actual results See above. ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: 2.1.1.dev0 - Platform: Linux-5.18.10-76051810-generic-x86_64-with-glibc2.34 - Python version: 3.9.7 - PyArrow version: 6.0.1 - Pandas version: 1.4.2
2023-03-02T15:33:05Z
4,889
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-08-24T16:54:43Z
https://api.github.com/repos/huggingface/datasets/issues/4889/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4889/timeline
torchaudio 11.0 yields different results than torchaudio 12.1 when loading MP3
https://api.github.com/repos/huggingface/datasets/issues/4889/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "events_url": "https://api.github.com/users/patrickvonplaten/events{/privacy}", "followers_url": "https://api.github.com/users/patrickvonplaten/followers", "following_url": "https://api.github.com/users/patrickvonplaten/following{/other_user}", "gists_url": "https://api.github.com/users/patrickvonplaten/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/patrickvonplaten", "id": 23423619, "login": "patrickvonplaten", "node_id": "MDQ6VXNlcjIzNDIzNjE5", "organizations_url": "https://api.github.com/users/patrickvonplaten/orgs", "received_events_url": "https://api.github.com/users/patrickvonplaten/received_events", "repos_url": "https://api.github.com/users/patrickvonplaten/repos", "site_admin": false, "starred_url": "https://api.github.com/users/patrickvonplaten/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/patrickvonplaten/subscriptions", "type": "User", "url": "https://api.github.com/users/patrickvonplaten" }
[]
null
completed
CONTRIBUTOR
2023-03-02T15:33:04Z
null
I_kwDODunzps5Qc649
[ "Maybe we can just pass this along to torchaudio @lhoestq @albertvillanova ? It be great if you could investigate if the errors lies in datasets or in torchaudio.", "torchaudio did a change in [0.12](https://github.com/pytorch/audio/releases/tag/v0.12.0) on MP3 decoding (which affects common voice):\r\n> MP3 deco...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4889/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4889
https://github.com/huggingface/datasets/issues/4889
false
1,349,447,521
https://api.github.com/repos/huggingface/datasets/issues/4888/labels{/name}
### Link https://huggingface.co/datasets/subjqa ### Description Getting the following error for this dataset: ``` Status code: 500 Exception: Status500Error Message: 2 or more items returned, instead of 1 ``` Not sure what's causing it though 🤔 ### Owner Yes
2022-09-08T08:23:42Z
4,888
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "E5583E", "default": false, "description": "Related to the dataset viewer on huggingface.co", "id": 3470211881, "name": "dataset-viewer", "node_id": "LA_kwDODunzps7O1zsp", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset-viewer" } ]
2022-08-24T13:26:20Z
https://api.github.com/repos/huggingface/datasets/issues/4888/comments
{ "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url": "https://api.github.com/users/severo/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/severo", "id": 1676121, "login": "severo", "node_id": "MDQ6VXNlcjE2NzYxMjE=", "organizations_url": "https://api.github.com/users/severo/orgs", "received_events_url": "https://api.github.com/users/severo/received_events", "repos_url": "https://api.github.com/users/severo/repos", "site_admin": false, "starred_url": "https://api.github.com/users/severo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/severo/subscriptions", "type": "User", "url": "https://api.github.com/users/severo" }
https://api.github.com/repos/huggingface/datasets/issues/4888/timeline
Dataset Viewer issue for subjqa
https://api.github.com/repos/huggingface/datasets/issues/4888/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "events_url": "https://api.github.com/users/lewtun/events{/privacy}", "followers_url": "https://api.github.com/users/lewtun/followers", "following_url": "https://api.github.com/users/lewtun/following{/other_user}", "gists_url": "https://api.github.com/users/lewtun/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lewtun", "id": 26859204, "login": "lewtun", "node_id": "MDQ6VXNlcjI2ODU5MjA0", "organizations_url": "https://api.github.com/users/lewtun/orgs", "received_events_url": "https://api.github.com/users/lewtun/received_events", "repos_url": "https://api.github.com/users/lewtun/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lewtun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lewtun/subscriptions", "type": "User", "url": "https://api.github.com/users/lewtun" }
[ { "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "events_url": "https://api.github.com/users/severo/events{/privacy}", "followers_url": "https://api.github.com/users/severo/followers", "following_url": "https://api.github.com/users/severo/following{/other_user}", "gists_url...
null
completed
MEMBER
2022-09-08T08:23:42Z
null
I_kwDODunzps5Qbu9h
[ "It's a bug in the viewer, thanks for reporting it. We're hoping to update to a new version in the next few days which should fix it.", "Fixed \r\n\r\nhttps://huggingface.co/datasets/subjqa\r\n\r\n<img width=\"1040\" alt=\"Capture d’écran 2022-09-08 à 10 23 26\" src=\"https://user-images.githubusercontent.com/1...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4888/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4888
https://github.com/huggingface/datasets/issues/4888
false
1,349,426,693
https://api.github.com/repos/huggingface/datasets/issues/4887/labels{/name}
Datasets side of https://github.com/huggingface/hub-docs/pull/285
2022-08-26T10:31:32Z
4,887
null
https://api.github.com/repos/huggingface/datasets
false
[]
2022-08-24T13:11:49Z
https://api.github.com/repos/huggingface/datasets/issues/4887/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4887/timeline
Add "cc-by-nc-sa-2.0" to list of licenses
https://api.github.com/repos/huggingface/datasets/issues/4887/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/7246357?v=4", "events_url": "https://api.github.com/users/osanseviero/events{/privacy}", "followers_url": "https://api.github.com/users/osanseviero/followers", "following_url": "https://api.github.com/users/osanseviero/following{/other_user}", "gists_url": "https://api.github.com/users/osanseviero/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/osanseviero", "id": 7246357, "login": "osanseviero", "node_id": "MDQ6VXNlcjcyNDYzNTc=", "organizations_url": "https://api.github.com/users/osanseviero/orgs", "received_events_url": "https://api.github.com/users/osanseviero/received_events", "repos_url": "https://api.github.com/users/osanseviero/repos", "site_admin": false, "starred_url": "https://api.github.com/users/osanseviero/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/osanseviero/subscriptions", "type": "User", "url": "https://api.github.com/users/osanseviero" }
[]
null
null
MEMBER
2022-08-26T10:29:20Z
{ "diff_url": "https://github.com/huggingface/datasets/pull/4887.diff", "html_url": "https://github.com/huggingface/datasets/pull/4887", "merged_at": "2022-08-26T10:29:20Z", "patch_url": "https://github.com/huggingface/datasets/pull/4887.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/4887" }
PR_kwDODunzps49t_PM
[ "_The documentation is not available anymore as the PR was closed or merged._", "Sorry for the issue @albertvillanova! I think it's now fixed! :heart: " ]
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4887/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4887
https://github.com/huggingface/datasets/pull/4887
true
1,349,285,569
https://api.github.com/repos/huggingface/datasets/issues/4886/labels{/name}
## Describe the bug Loading huggan/CelebA-HQ throws pyarrow.lib.ArrowInvalid ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset('huggan/CelebA-HQ') ``` ## Expected results See https://colab.research.google.com/drive/141LJCcM2XyqprPY83nIQ-Zk3BbxWeahq?usp=sharing#scrollTo=N3ml_7f8kzDd ## Actual results ``` File "/home/jean/projects/cold_diffusion/celebA.py", line 4, in <module> dataset = load_dataset('huggan/CelebA-HQ') File "/home/jean/miniconda3/envs/seq/lib/python3.10/site-packages/datasets/load.py", line 1793, in load_dataset builder_instance.download_and_prepare( File "/home/jean/miniconda3/envs/seq/lib/python3.10/site-packages/datasets/builder.py", line 704, in download_and_prepare self._download_and_prepare( File "/home/jean/miniconda3/envs/seq/lib/python3.10/site-packages/datasets/builder.py", line 793, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/home/jean/miniconda3/envs/seq/lib/python3.10/site-packages/datasets/builder.py", line 1274, in _prepare_split for key, table in logging.tqdm( File "/home/jean/miniconda3/envs/seq/lib/python3.10/site-packages/tqdm/std.py", line 1195, in __iter__ for obj in iterable: File "/home/jean/miniconda3/envs/seq/lib/python3.10/site-packages/datasets/packaged_modules/parquet/parquet.py", line 67, in _generate_tables parquet_file = pq.ParquetFile(f) File "/home/jean/miniconda3/envs/seq/lib/python3.10/site-packages/pyarrow/parquet/__init__.py", line 286, in __init__ self.reader.open( File "pyarrow/_parquet.pyx", line 1227, in pyarrow._parquet.ParquetReader.open File "pyarrow/error.pxi", line 100, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: Parquet magic bytes not found in footer. Either the file is corrupted or this is not a parquet file. ``` ## Environment info <!-- You can run the command `datasets-cli env` and copy-and-paste its output below. --> - `datasets` version: datasets-2.4.1.dev0 - Platform: Ubuntu 18.04 - Python version: 3.10 - PyArrow version: pyarrow 9.0.0
2023-02-02T02:40:53Z
4,886
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "d73a4a", "default": true, "description": "Something isn't working", "id": 1935892857, "name": "bug", "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug" } ]
2022-08-24T11:24:21Z
https://api.github.com/repos/huggingface/datasets/issues/4886/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4886/timeline
Loading huggan/CelebA-HQ throws pyarrow.lib.ArrowInvalid
https://api.github.com/repos/huggingface/datasets/issues/4886/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/11850255?v=4", "events_url": "https://api.github.com/users/JeanKaddour/events{/privacy}", "followers_url": "https://api.github.com/users/JeanKaddour/followers", "following_url": "https://api.github.com/users/JeanKaddour/following{/other_user}", "gists_url": "https://api.github.com/users/JeanKaddour/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JeanKaddour", "id": 11850255, "login": "JeanKaddour", "node_id": "MDQ6VXNlcjExODUwMjU1", "organizations_url": "https://api.github.com/users/JeanKaddour/orgs", "received_events_url": "https://api.github.com/users/JeanKaddour/received_events", "repos_url": "https://api.github.com/users/JeanKaddour/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JeanKaddour/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JeanKaddour/subscriptions", "type": "User", "url": "https://api.github.com/users/JeanKaddour" }
[]
null
null
NONE
null
null
I_kwDODunzps5QbHbB
[ "Hi! IIRC one of the files in this dataset is corrupted due to https://github.com/huggingface/datasets/pull/4081 (fixed now).\r\n\r\n@NielsRogge Could you please re-generate and re-push this dataset (or I can do it if you share the generation script)?", "Could you put something in place to catch these problems? ...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4886/reactions" }
open
false
https://api.github.com/repos/huggingface/datasets/issues/4886
https://github.com/huggingface/datasets/issues/4886
false
1,349,181,448
https://api.github.com/repos/huggingface/datasets/issues/4885/labels{/name}
I often find myself with data from a variety of sources, and a list of dicts is very common among these. However, converting this to a Dataset is a little awkward, requiring either ```Dataset.from_pandas(pd.DataFrame(formatted_training_data))``` Which can error out on some more exotic values as 2-d arrays for reasons that are not entirely clear > ArrowInvalid: ('Can only convert 1-dimensional array values', 'Conversion failed for column labels with type object') Alternatively: ```Dataset.from_dict({k: [s[k] for s in formatted_training_data] for k in formatted_training_data[0].keys()})``` Which works, but is a little ugly. **Describe the solution you'd like** Either `.from_dict` accepting a list of dicts, or a `.from_records` function accepting such. I am happy to PR this, just wanted to check you are happy to accept this I haven't missed something obvious, and which of the solutions would be prefered.
2022-09-08T16:02:52Z
4,885
null
https://api.github.com/repos/huggingface/datasets
null
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
2022-08-24T10:01:24Z
https://api.github.com/repos/huggingface/datasets/issues/4885/comments
null
https://api.github.com/repos/huggingface/datasets/issues/4885/timeline
Create dataset from list of dicts
https://api.github.com/repos/huggingface/datasets/issues/4885/events
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/48946947?v=4", "events_url": "https://api.github.com/users/sanderland/events{/privacy}", "followers_url": "https://api.github.com/users/sanderland/followers", "following_url": "https://api.github.com/users/sanderland/following{/other_user}", "gists_url": "https://api.github.com/users/sanderland/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sanderland", "id": 48946947, "login": "sanderland", "node_id": "MDQ6VXNlcjQ4OTQ2OTQ3", "organizations_url": "https://api.github.com/users/sanderland/orgs", "received_events_url": "https://api.github.com/users/sanderland/received_events", "repos_url": "https://api.github.com/users/sanderland/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sanderland/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sanderland/subscriptions", "type": "User", "url": "https://api.github.com/users/sanderland" }
[]
null
completed
CONTRIBUTOR
2022-09-08T16:02:52Z
null
I_kwDODunzps5QauAI
[ "Hi @sanderland, thanks for your enhancement proposal.\r\n\r\nI agree with you that this would be useful.\r\n\r\nPlease note that under the hood, we use PyArrow tables as backend:\r\n- The implementation of `Dataset.from_dict` uses the PyArrow `Table.from_pydict`\r\n\r\nTherefore, I would suggest:\r\n- Implementin...
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/4885/reactions" }
closed
false
https://api.github.com/repos/huggingface/datasets/issues/4885
https://github.com/huggingface/datasets/issues/4885
false