id
int64
599M
3.48B
number
int64
1
7.8k
title
stringlengths
1
290
state
stringclasses
2 values
comments
listlengths
0
30
created_at
timestamp[s]date
2020-04-14 10:18:02
2025-10-05 06:37:50
updated_at
timestamp[s]date
2020-04-27 16:04:17
2025-10-05 10:32:43
closed_at
timestamp[s]date
2020-04-14 12:01:40
2025-10-01 13:56:03
body
stringlengths
0
228k
user
stringlengths
3
26
html_url
stringlengths
46
51
pull_request
dict
is_pull_request
bool
2 classes
1,805,087,687
6,035
Dataset representation
open
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6035). All of your documentation changes will be reflected on that endpoint." ]
2023-07-14T15:42:37
2023-07-19T19:41:35
null
__repr__ and _repr_html_ now both are similar to that of Polars
Ganryuu
https://github.com/huggingface/datasets/pull/6035
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6035", "html_url": "https://github.com/huggingface/datasets/pull/6035", "diff_url": "https://github.com/huggingface/datasets/pull/6035.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6035.patch", "merged_at": null }
true
1,804,501,361
6,034
load_dataset hangs on WSL
closed
[ "Even if a dataset is cached, we still make requests to check whether the cache is up-to-date. [This](https://huggingface.co/docs/datasets/v2.13.1/en/loading#offline) section in the docs explains how to avoid them and directly load the cached version.", "Thanks - that works! However it doesn't resolve the origina...
2023-07-14T09:03:10
2023-07-14T14:48:29
2023-07-14T14:48:29
### Describe the bug load_dataset simply hangs. It happens once every ~5 times, and interestingly hangs for a multiple of 5 minutes (hangs for 5/10/15 minutes). Using the profiler in PyCharm shows that it spends the time at <method 'connect' of '_socket.socket' objects>. However, a local cache is available so I am not...
Andy-Zhou2
https://github.com/huggingface/datasets/issues/6034
null
false
1,804,482,051
6,033
`map` function doesn't fully utilize `input_columns`.
closed
[]
2023-07-14T08:49:28
2023-07-14T09:16:04
2023-07-14T09:16:04
### Describe the bug I wanted to select only some columns of data. And I thought that's why the argument `input_columns` exists. What I expected is like this: If there are ["a", "b", "c", "d"] columns, and if I set `input_columns=["a", "d"]`, the data will have only ["a", "d"] columns. But it doesn't select co...
kwonmha
https://github.com/huggingface/datasets/issues/6033
null
false
1,804,358,679
6,032
DownloadConfig.proxies not work when load_dataset_builder calling HfApi.dataset_info
open
[ "`HfApi` comes from the `huggingface_hub` package. You can use [this](https://huggingface.co/docs/huggingface_hub/v0.16.3/en/package_reference/utilities#huggingface_hub.configure_http_backend) utility to change the `huggingface_hub`'s `Session` proxies (see the example).\r\n\r\nWe plan to implement https://github.c...
2023-07-14T07:22:55
2023-09-11T13:50:41
null
### Describe the bug ```python download_config = DownloadConfig(proxies={'https': '<my proxy>'}) builder = load_dataset_builder(..., download_config=download_config) ``` But, when getting the dataset_info from HfApi, the http requests not using the proxies. ### Steps to reproduce the bug 1. Setup proxies i...
codingl2k1
https://github.com/huggingface/datasets/issues/6032
null
false
1,804,183,858
6,031
Argument type for map function changes when using `input_columns` for `IterableDataset`
closed
[ "Yes, this is intended." ]
2023-07-14T05:11:14
2023-07-14T14:44:15
2023-07-14T14:44:15
### Describe the bug I wrote `tokenize(examples)` function as an argument for `map` function for `IterableDataset`. It process dictionary type `examples` as a parameter. It is used in `train_dataset = train_dataset.map(tokenize, batched=True)` No error is raised. And then, I found some unnecessary keys and val...
kwonmha
https://github.com/huggingface/datasets/issues/6031
null
false
1,803,864,744
6,030
fixed typo in comment
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-07-13T22:49:57
2023-07-14T14:21:58
2023-07-14T14:13:38
This mistake was a bit confusing, so I thought it was worth sending a PR over.
NightMachinery
https://github.com/huggingface/datasets/pull/6030
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6030", "html_url": "https://github.com/huggingface/datasets/pull/6030", "diff_url": "https://github.com/huggingface/datasets/pull/6030.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6030.patch", "merged_at": "2023-07-14T14:13...
true
1,803,460,046
6,029
[docs] Fix link
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-07-13T17:24:12
2023-07-13T17:47:41
2023-07-13T17:38:59
Fixes link to the builder classes :)
stevhliu
https://github.com/huggingface/datasets/pull/6029
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6029", "html_url": "https://github.com/huggingface/datasets/pull/6029", "diff_url": "https://github.com/huggingface/datasets/pull/6029.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6029.patch", "merged_at": "2023-07-13T17:38...
true
1,803,294,981
6,028
Use new hffs
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-07-13T15:41:44
2023-07-17T17:09:39
2023-07-17T17:01:00
Thanks to @janineguo 's work in https://github.com/huggingface/datasets/pull/5919 which was needed to support HfFileSystem. Switching to `HfFileSystem` will help implementing optimization in data files resolution ## Implementation details I replaced all the from_hf_repo and from_local_or_remote in data_files.p...
lhoestq
https://github.com/huggingface/datasets/pull/6028
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6028", "html_url": "https://github.com/huggingface/datasets/pull/6028", "diff_url": "https://github.com/huggingface/datasets/pull/6028.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6028.patch", "merged_at": "2023-07-17T17:01...
true
1,803,008,486
6,027
Delete `task_templates` in `IterableDataset` when they are no longer valid
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-07-13T13:16:17
2023-07-13T14:06:20
2023-07-13T13:57:35
Fix #6025
mariosasko
https://github.com/huggingface/datasets/pull/6027
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6027", "html_url": "https://github.com/huggingface/datasets/pull/6027", "diff_url": "https://github.com/huggingface/datasets/pull/6027.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6027.patch", "merged_at": "2023-07-13T13:57...
true
1,802,929,222
6,026
Fix style with ruff 0.0.278
closed
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6026). All of your documentation changes will be reflected on that endpoint.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma...
2023-07-13T12:34:24
2023-07-13T12:46:26
2023-07-13T12:37:01
null
lhoestq
https://github.com/huggingface/datasets/pull/6026
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6026", "html_url": "https://github.com/huggingface/datasets/pull/6026", "diff_url": "https://github.com/huggingface/datasets/pull/6026.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6026.patch", "merged_at": "2023-07-13T12:37...
true
1,801,852,601
6,025
Using a dataset for a use other than it was intended for.
closed
[ "I've opened a PR with a fix. In the meantime, you can avoid the error by deleting `task_templates` with `dataset.info.task_templates = None` before the `interleave_datasets` call.\r\n` " ]
2023-07-12T22:33:17
2023-07-13T13:57:36
2023-07-13T13:57:36
### Describe the bug Hi, I want to use the rotten tomatoes dataset but for a task other than classification, but when I interleave the dataset, it throws ```'ValueError: Column label is not present in features.'```. It seems that the label_col must be there in the dataset for some reason? Here is the full stacktra...
surya-narayanan
https://github.com/huggingface/datasets/issues/6025
null
false
1,801,708,808
6,024
Don't reference self in Spark._validate_cache_dir
closed
[ "Ptal @lhoestq :) I tested this manually on a multi-node Databricks cluster", "Hm looks like the check_code_quality failures are unrelated to me change... https://github.com/huggingface/datasets/actions/runs/5536162850/jobs/10103451883?pr=6024", "_The documentation is not available anymore as the PR was closed ...
2023-07-12T20:31:16
2023-07-13T16:58:32
2023-07-13T12:37:09
Fix for https://github.com/huggingface/datasets/issues/5963
maddiedawson
https://github.com/huggingface/datasets/pull/6024
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6024", "html_url": "https://github.com/huggingface/datasets/pull/6024", "diff_url": "https://github.com/huggingface/datasets/pull/6024.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6024.patch", "merged_at": "2023-07-13T12:37...
true
1,801,272,420
6,023
Fix `ClassLabel` min max check for `None` values
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-07-12T15:46:12
2023-07-12T16:29:26
2023-07-12T16:18:04
Fix #6022
mariosasko
https://github.com/huggingface/datasets/pull/6023
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6023", "html_url": "https://github.com/huggingface/datasets/pull/6023", "diff_url": "https://github.com/huggingface/datasets/pull/6023.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6023.patch", "merged_at": "2023-07-12T16:18...
true
1,800,092,589
6,022
Batch map raises TypeError: '>=' not supported between instances of 'NoneType' and 'int'
closed
[ "Thanks for reporting! I've opened a PR with a fix." ]
2023-07-12T03:20:17
2023-07-12T16:18:06
2023-07-12T16:18:05
### Describe the bug When mapping some datasets with `batched=True`, datasets may raise an exeception: ```python Traceback (most recent call last): File "/Users/codingl2k1/Work/datasets/venv/lib/python3.11/site-packages/multiprocess/pool.py", line 125, in worker result = (True, func(*args, **kwds)) ...
codingl2k1
https://github.com/huggingface/datasets/issues/6022
null
false
1,799,785,904
6,021
[docs] Update return statement of index search
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-07-11T21:33:32
2023-07-12T17:13:02
2023-07-12T17:03:00
Clarifies in the return statement of the docstring that the retrieval score is `IndexFlatL2` by default (see [PR](https://github.com/huggingface/transformers/issues/24739) and internal Slack [convo](https://huggingface.slack.com/archives/C01229B19EX/p1689105179711689)), and fixes the formatting because multiple return ...
stevhliu
https://github.com/huggingface/datasets/pull/6021
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6021", "html_url": "https://github.com/huggingface/datasets/pull/6021", "diff_url": "https://github.com/huggingface/datasets/pull/6021.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6021.patch", "merged_at": "2023-07-12T17:03...
true
1,799,720,536
6,020
Inconsistent "The features can't be aligned" error when combining map, multiprocessing, and variable length outputs
open
[ "This scenario currently requires explicitly passing the target features (to avoid the error): \r\n```python\r\nimport datasets\r\n\r\n...\r\n\r\nfeatures = dataset.features\r\nfeatures[\"output\"] = = [{\"test\": datasets.Value(\"int64\")}]\r\ntest2 = dataset.map(lambda row, idx: test_func(row, idx), with_indices=...
2023-07-11T20:40:38
2024-10-27T06:30:13
null
### Describe the bug I'm using a dataset with map and multiprocessing to run a function that returned a variable length list of outputs. This output list may be empty. Normally this is handled fine, but there is an edge case that crops up when using multiprocessing. In some cases, an empty list result ends up in a dat...
kheyer
https://github.com/huggingface/datasets/issues/6020
null
false
1,799,532,822
6,019
Improve logging
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-07-11T18:30:23
2023-07-12T19:34:14
2023-07-12T17:19:28
Adds the StreamHandler (as `hfh` and `transformers` do) to the library's logger to log INFO messages and logs the messages about "loading a cached result" (and some other warnings) as INFO (Also removes the `leave=False` arg in the progress bars to be consistent with `hfh` and `transformers` - progress bars serve as...
mariosasko
https://github.com/huggingface/datasets/pull/6019
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6019", "html_url": "https://github.com/huggingface/datasets/pull/6019", "diff_url": "https://github.com/huggingface/datasets/pull/6019.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6019.patch", "merged_at": "2023-07-12T17:19...
true
1,799,411,999
6,018
test1
closed
[ "We no longer host datasets in this repo. You should use the HF Hub instead." ]
2023-07-11T17:25:49
2023-07-20T10:11:41
2023-07-20T10:11:41
null
ognjenovicj
https://github.com/huggingface/datasets/pull/6018
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6018", "html_url": "https://github.com/huggingface/datasets/pull/6018", "diff_url": "https://github.com/huggingface/datasets/pull/6018.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6018.patch", "merged_at": null }
true
1,799,309,132
6,017
Switch to huggingface_hub's HfFileSystem
closed
[]
2023-07-11T16:24:40
2023-07-17T17:01:01
2023-07-17T17:01:01
instead of the current datasets.filesystems.hffilesystem.HfFileSystem which can be slow in some cases related to https://github.com/huggingface/datasets/issues/5846 and https://github.com/huggingface/datasets/pull/5919
lhoestq
https://github.com/huggingface/datasets/issues/6017
null
false
1,798,968,033
6,016
Dataset string representation enhancement
open
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_6016). All of your documentation changes will be reflected on that endpoint.", "It we could have something similar to Polars, that would be great.\r\n\r\nThis is what Polars outputs: \r\n* `__repr__`/`__str__` :\r\n```\r\nshape...
2023-07-11T13:38:25
2023-07-16T10:26:18
null
my attempt at #6010 not sure if this is the right way to go about it, I will wait for your feedback
Ganryuu
https://github.com/huggingface/datasets/pull/6016
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6016", "html_url": "https://github.com/huggingface/datasets/pull/6016", "diff_url": "https://github.com/huggingface/datasets/pull/6016.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6016.patch", "merged_at": null }
true
1,798,807,893
6,015
Add metadata ui screenshot in docs
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-07-11T12:16:29
2023-07-11T16:07:28
2023-07-11T15:56:46
null
lhoestq
https://github.com/huggingface/datasets/pull/6015
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6015", "html_url": "https://github.com/huggingface/datasets/pull/6015", "diff_url": "https://github.com/huggingface/datasets/pull/6015.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6015.patch", "merged_at": "2023-07-11T15:56...
true
1,798,213,816
6,014
Request to Share/Update Dataset Viewer Code
closed
[ "Hi ! The huggingface/dataset-viewer code was not maintained anymore because we switched to a new dataset viewer that is deployed available for each dataset the Hugging Face website.\r\n\r\nWhat are you using this old repository for ?", "I think these parts are outdated:\r\n\r\n* https://github.com/huggingface/da...
2023-07-11T06:36:09
2024-07-20T07:29:08
2023-09-25T12:01:17
Overview: The repository (huggingface/datasets-viewer) was recently archived and when I tried to run the code, there was the error message "AttributeError: module 'datasets.load' has no attribute 'prepare_module'". I could not resolve the issue myself due to lack of documentation of that attribute. Request: I k...
lilyorlilypad
https://github.com/huggingface/datasets/issues/6014
null
false
1,796,083,437
6,013
[FR] `map` should reuse unchanged columns from the previous dataset to avoid disk usage
open
[ "You can use the `remove_columns` parameter in `map` to avoid duplicating the columns (and save disk space) and then concatenate the original dataset with the map result:\r\n```python\r\nfrom datasets import concatenate_datasets\r\n# dummy example\r\nds_new = ds.map(lambda x: {\"new_col\": x[\"col\"] + 2}, remove_c...
2023-07-10T06:42:20
2025-06-19T06:30:38
null
### Feature request Currently adding a new column with `map` will cause all the data in the dataset to be duplicated and stored/cached on the disk again. It should reuse unchanged columns. ### Motivation This allows having datasets with different columns but sharing some basic columns. Currently, these datasets wou...
NightMachinery
https://github.com/huggingface/datasets/issues/6013
null
false
1,795,575,432
6,012
[FR] Transform Chaining, Lazy Mapping
open
[ "You can use `with_transform` to get a new dataset object.\r\n\r\nSupport for lazy `map` has already been discussed [here](https://github.com/huggingface/datasets/issues/3385) a little bit. Personally, I'm not a fan, as this would make `map` even more complex. ", "> You can use `with_transform` to get a new datas...
2023-07-09T21:40:21
2025-01-20T14:06:28
null
### Feature request Currently using a `map` call processes and duplicates the whole dataset, which takes both time and disk space. The solution is to allow lazy mapping, which is essentially a saved chain of transforms that are applied on the fly whenever a slice of the dataset is requested. The API should look ...
NightMachinery
https://github.com/huggingface/datasets/issues/6012
null
false
1,795,296,568
6,011
Documentation: wiki_dpr Dataset has no metric_type for Faiss Index
closed
[ "Hi! You can do `ds.get_index(\"embeddings\").faiss_index.metric_type` to get the metric type and then match the result with the FAISS metric [enum](https://github.com/facebookresearch/faiss/blob/43d86e30736ede853c384b24667fc3ab897d6ba9/faiss/MetricType.h#L22-L36) (should be L2).", "Ah! Thank you for pointing thi...
2023-07-09T08:30:19
2023-07-11T03:02:36
2023-07-11T03:02:36
### Describe the bug After loading `wiki_dpr` using: ```py ds = load_dataset(path='wiki_dpr', name='psgs_w100.multiset.compressed', split='train') print(ds.get_index("embeddings").metric_type) # prints nothing because the value is None ``` the index does not have a defined `metric_type`. This is an issue because ...
YichiRockyZhang
https://github.com/huggingface/datasets/issues/6011
null
false
1,793,838,152
6,010
Improve `Dataset`'s string representation
open
[ "I want to take a shot at this if possible ", "Yes, feel free to work on this.\r\n\r\nYou can check the PyArrow Table `__repr__` and Polars DataFrame `__repr__`/`_repr_html_` implementations for some pointers/ideas.", "@mariosasko are there any other similar issues that I could work on? I see this has been alr...
2023-07-07T16:38:03
2023-09-01T03:45:07
null
Currently, `Dataset.__repr__` outputs a dataset's column names and the number of rows. We could improve it by printing its features and the first few rows. We should also implement `_repr_html_` to have a rich HTML representation in notebooks/Streamlit.
mariosasko
https://github.com/huggingface/datasets/issues/6010
null
false
1,792,059,808
6,009
Fix cast for dictionaries with no keys
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-07-06T18:48:14
2023-07-07T14:13:00
2023-07-07T14:01:13
Fix #5677
mariosasko
https://github.com/huggingface/datasets/pull/6009
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6009", "html_url": "https://github.com/huggingface/datasets/pull/6009", "diff_url": "https://github.com/huggingface/datasets/pull/6009.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6009.patch", "merged_at": "2023-07-07T14:01...
true
1,789,869,344
6,008
Dataset.from_generator consistently freezes at ~1000 rows
closed
[ "By default, we write data to disk (so it can be memory-mapped) every 1000 rows/samples. You can control this with the `writer_batch_size` parameter. Also, when working with fixed-size arrays, the `ArrayXD` feature types yield better performance (e.g., in your case, `features=datasets.Features({\"i\": datasets.Arra...
2023-07-05T16:06:48
2023-07-10T13:46:39
2023-07-10T13:46:39
### Describe the bug Whenever I try to create a dataset which contains images using `Dataset.from_generator`, it freezes around 996 rows. I suppose it has something to do with memory consumption, but there's more memory available. I Somehow it worked a few times but mostly this makes the datasets library much more ...
andreemic
https://github.com/huggingface/datasets/issues/6008
null
false
1,789,782,693
6,007
Get an error "OverflowError: Python int too large to convert to C long" when loading a large dataset
open
[ "This error means that one of the int32 (`Value(\"int32\")`) columns in the dataset has a value that is out of the valid (int32) range.\r\n\r\nI'll open a PR to print the name of a problematic column to make debugging such errors easier.", "I am afraid int32 is not the reason for this error.\r\n\r\nI have submitt...
2023-07-05T15:16:50
2024-02-07T22:22:35
null
### Describe the bug When load a large dataset with the following code ```python from datasets import load_dataset dataset = load_dataset("liwu/MNBVC", 'news_peoples_daily', split='train') ``` We encountered the error: "OverflowError: Python int too large to convert to C long" The error look something like...
silverriver
https://github.com/huggingface/datasets/issues/6007
null
false
1,788,855,582
6,006
NotADirectoryError when loading gigawords
closed
[ "issue due to corrupted download files. resolved after cleaning download cache. sorry for any inconvinence." ]
2023-07-05T06:23:41
2023-07-05T06:31:02
2023-07-05T06:31:01
### Describe the bug got `NotADirectoryError` whtn loading gigawords dataset ### Steps to reproduce the bug When running ``` import datasets datasets.load_dataset('gigaword') ``` Got the following exception: ```bash Traceback (most recent call last): ...
xipq
https://github.com/huggingface/datasets/issues/6006
null
false
1,788,103,576
6,005
Drop Python 3.7 support
closed
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
2023-07-04T15:02:37
2023-07-06T15:32:41
2023-07-06T15:22:43
`hfh` and `transformers` have dropped Python 3.7 support, so we should do the same :). (Based on the stats, it seems less than 10% of the users use `datasets` with Python 3.7)
mariosasko
https://github.com/huggingface/datasets/pull/6005
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6005", "html_url": "https://github.com/huggingface/datasets/pull/6005", "diff_url": "https://github.com/huggingface/datasets/pull/6005.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6005.patch", "merged_at": "2023-07-06T15:22...
true
1,786,636,368
6,004
Misc improvements
closed
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
2023-07-03T18:29:14
2023-07-06T17:04:11
2023-07-06T16:55:25
Contains the following improvements: * fixes a "share dataset" link in README and modifies the "hosting" part in the disclaimer section * updates `Makefile` to also run the style checks on `utils` and `setup.py` * deletes a test for GH-hosted datasets (no longer supported) * deletes `convert_dataset.sh` (outdated...
mariosasko
https://github.com/huggingface/datasets/pull/6004
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6004", "html_url": "https://github.com/huggingface/datasets/pull/6004", "diff_url": "https://github.com/huggingface/datasets/pull/6004.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6004.patch", "merged_at": "2023-07-06T16:55...
true
1,786,554,110
6,003
interleave_datasets & DataCollatorForLanguageModeling having a conflict ?
open
[]
2023-07-03T17:15:31
2023-07-03T17:15:31
null
### Describe the bug Hi everyone :) I have two local & custom datasets (1 "sentence" per line) which I split along the 95/5 lines for pre-training a Bert model. I use a modified version of `run_mlm.py` in order to be able to make use of `interleave_dataset`: - `tokenize()` runs fine - `group_text()` runs fine ...
PonteIneptique
https://github.com/huggingface/datasets/issues/6003
null
false
1,786,053,060
6,002
Add KLUE-MRC metrics
closed
[ "The metrics API in `datasets` is deprecated as of version 2.0, and `evaulate` is our new library for metrics. You can add a new metric to it by following [these steps](https://huggingface.co/docs/evaluate/creating_and_sharing)." ]
2023-07-03T12:11:10
2023-07-09T11:57:20
2023-07-09T11:57:20
## Metrics for KLUE-MRC (Korean Language Understanding Evaluation — Machine Reading Comprehension) Adding metrics for [KLUE-MRC](https://huggingface.co/datasets/klue). KLUE-MRC is very similar to SQuAD 2.0 but has a slightly different format which is why I added metrics for KLUE-MRC. Specifically, in the case of...
ingyuseong
https://github.com/huggingface/datasets/pull/6002
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6002", "html_url": "https://github.com/huggingface/datasets/pull/6002", "diff_url": "https://github.com/huggingface/datasets/pull/6002.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6002.patch", "merged_at": null }
true
1,782,516,627
6,001
Align `column_names` type check with type hint in `sort`
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-30T13:15:50
2023-06-30T14:18:32
2023-06-30T14:11:24
Fix #5998
mariosasko
https://github.com/huggingface/datasets/pull/6001
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6001", "html_url": "https://github.com/huggingface/datasets/pull/6001", "diff_url": "https://github.com/huggingface/datasets/pull/6001.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6001.patch", "merged_at": "2023-06-30T14:11...
true
1,782,456,878
6,000
Pin `joblib` to avoid `joblibspark` test failures
closed
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
2023-06-30T12:36:54
2023-06-30T13:17:05
2023-06-30T13:08:27
`joblibspark` doesn't support the latest `joblib` release. See https://github.com/huggingface/datasets/actions/runs/5401870932/jobs/9812337078 for the errors
mariosasko
https://github.com/huggingface/datasets/pull/6000
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/6000", "html_url": "https://github.com/huggingface/datasets/pull/6000", "diff_url": "https://github.com/huggingface/datasets/pull/6000.diff", "patch_url": "https://github.com/huggingface/datasets/pull/6000.patch", "merged_at": "2023-06-30T13:08...
true
1,781,851,513
5,999
Getting a 409 error while loading xglue dataset
closed
[ "Thanks for reporting, @Praful932.\r\n\r\nLet's continue the conversation on the Hub: https://huggingface.co/datasets/xglue/discussions/5" ]
2023-06-30T04:13:54
2023-06-30T05:57:23
2023-06-30T05:57:22
### Describe the bug Unable to load xglue dataset ### Steps to reproduce the bug ```python import datasets dataset = datasets.load_dataset("xglue", "ntg") ``` > ConnectionError: Couldn't reach https://xglue.blob.core.windows.net/xglue/xglue_full_dataset.tar.gz (error 409) ### Expected behavior Expected the...
Praful932
https://github.com/huggingface/datasets/issues/5999
null
false
1,781,805,018
5,998
The current implementation has a potential bug in the sort method
closed
[ "Thanks for reporting, @wangyuxinwhy. " ]
2023-06-30T03:16:57
2023-06-30T14:21:03
2023-06-30T14:11:25
### Describe the bug In the sort method,here's a piece of code ```python # column_names: Union[str, Sequence_[str]] # Check proper format of and for duplicates in column_names if not isinstance(column_names, list): column_names = [column_names] ``` I get an error when I pass in a tuple based on the ...
wangyuxinwhy
https://github.com/huggingface/datasets/issues/5998
null
false
1,781,582,818
5,997
extend the map function so it can wrap around long text that does not fit in the context window
open
[ "I just noticed the [docs](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L2881C11-L2881C200) say:\r\n\r\n>If batched is `True` and `batch_size` is `n > 1`, then the function takes a batch of `n` examples as input and can return a batch with `n` examples, or with an arbitrary number...
2023-06-29T22:15:21
2023-07-03T17:58:52
null
### Feature request I understand `dataset` provides a [`map`](https://github.com/huggingface/datasets/blob/main/src/datasets/arrow_dataset.py#L2849) function. This function in turn takes in a callable that is used to tokenize the text on which a model is trained. Frequently this text will not fit within a models's con...
siddhsql
https://github.com/huggingface/datasets/issues/5997
null
false
1,779,294,374
5,996
Deprecate `use_auth_token` in favor of `token`
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-28T16:26:38
2023-07-05T15:22:20
2023-07-03T16:03:33
... to be consistent with `transformers` and `huggingface_hub`.
mariosasko
https://github.com/huggingface/datasets/pull/5996
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5996", "html_url": "https://github.com/huggingface/datasets/pull/5996", "diff_url": "https://github.com/huggingface/datasets/pull/5996.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5996.patch", "merged_at": "2023-07-03T16:03...
true
1,777,088,925
5,995
Support returning dataframe in map transform
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-27T14:15:08
2023-06-28T13:56:02
2023-06-28T13:46:33
Allow returning Pandas DataFrames in `map` transforms. (Plus, raise an error in the non-batched mode if a returned PyArrow table/Pandas DataFrame has more than one row)
mariosasko
https://github.com/huggingface/datasets/pull/5995
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5995", "html_url": "https://github.com/huggingface/datasets/pull/5995", "diff_url": "https://github.com/huggingface/datasets/pull/5995.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5995.patch", "merged_at": "2023-06-28T13:46...
true
1,776,829,004
5,994
Fix select_columns columns order
closed
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
2023-06-27T12:32:46
2023-06-27T15:40:47
2023-06-27T15:32:43
Fix the order of the columns in dataset.features when the order changes with `dataset.select_columns()`. I also fixed the same issue for `dataset.flatten()` Close https://github.com/huggingface/datasets/issues/5993
lhoestq
https://github.com/huggingface/datasets/pull/5994
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5994", "html_url": "https://github.com/huggingface/datasets/pull/5994", "diff_url": "https://github.com/huggingface/datasets/pull/5994.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5994.patch", "merged_at": "2023-06-27T15:32...
true
1,776,643,555
5,993
ValueError: Table schema does not match schema used to create file
closed
[ "We'll do a new release of `datasets` soon to make the fix available :)\r\n\r\nIn the meantime you can use `datasets` from source (main)", "Thank you very much @lhoestq ! 🚀 " ]
2023-06-27T10:54:07
2023-06-27T15:36:42
2023-06-27T15:32:44
### Describe the bug Saving a dataset as parquet fails with a `ValueError: Table schema does not match schema used to create file` if the dataset was obtained out of a `.select_columns()` call with columns selected out of order. ### Steps to reproduce the bug ```python import datasets dataset = datasets.Dataset...
exs-avianello
https://github.com/huggingface/datasets/issues/5993
null
false
1,776,460,964
5,992
speedup
closed
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5992). All of your documentation changes will be reflected on that endpoint." ]
2023-06-27T09:17:58
2023-06-27T09:23:07
2023-06-27T09:18:04
null
qgallouedec
https://github.com/huggingface/datasets/pull/5992
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5992", "html_url": "https://github.com/huggingface/datasets/pull/5992", "diff_url": "https://github.com/huggingface/datasets/pull/5992.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5992.patch", "merged_at": null }
true
1,774,456,518
5,991
`map` with any joblib backend
open
[ "#self-assign\n\nHi @lhoestq 👋🏼\n\nI’d like to work on this!\n\nPlanning to support progress tracking with `map()` using any joblib backend (like \"loky\") by replacing the Queue-based approach in `iflatmap_unordered` with a file-based progress tracking mechanism (e.g. shared temp file with periodic updates).\n\n...
2023-06-26T10:33:42
2025-09-04T10:43:06
null
We recently enabled the (experimental) parallel backend switch for data download and extraction but not for `map` yet. Right now we're using our `iflatmap_unordered` implementation for multiprocessing that uses a shared Queue to gather progress updates from the subprocesses and show a progress bar in the main proces...
lhoestq
https://github.com/huggingface/datasets/issues/5991
null
false
1,774,134,091
5,989
Set a rule on the config and split names
open
[ "in this case we need to decide what to do with the existing datasets with white space characters (there shouldn't be a lot of them I think)", "I imagine that we should stop supporting them, and help the user fix them?", "See a report where the datasets server fails: https://huggingface.co/datasets/poloclub/dif...
2023-06-26T07:34:14
2023-07-19T14:22:54
null
> should we actually allow characters like spaces? maybe it's better to add validation for whitespace symbols and directly in datasets and raise https://github.com/huggingface/datasets-server/issues/853
severo
https://github.com/huggingface/datasets/issues/5989
null
false
1,773,257,828
5,988
ConnectionError: Couldn't reach dataset_infos.json
closed
[ "Unfortunately, I can't reproduce the error. What does the following code return for you?\r\n```python\r\nimport requests\r\nfrom huggingface_hub import hf_hub_url\r\nr = requests.get(hf_hub_url(\"codeparrot/codeparrot-clean-train\", \"dataset_infos.json\", repo_type=\"dataset\"))\r\n```\r\n\r\nAlso, can you provid...
2023-06-25T12:39:31
2023-07-07T13:20:57
2023-07-07T13:20:57
### Describe the bug I'm trying to load codeparrot/codeparrot-clean-train, but get the following error: ConnectionError: Couldn't reach https://huggingface.co/datasets/codeparrot/codeparrot-clean-train/resolve/main/dataset_infos.json (ConnectionError(ProtocolError('Connection aborted.', ConnectionResetError(104, 'C...
yulingao
https://github.com/huggingface/datasets/issues/5988
null
false
1,773,047,909
5,987
Why max_shard_size is not supported in load_dataset and passed to download_and_prepare
closed
[ "Can you explain your use case for `max_shard_size`? \r\n\r\nOn some systems, there is a limit to the size of a memory-mapped file, so we could consider exposing this parameter in `load_dataset`.", "In my use case, users may choose a proper size to balance the cost and benefit of using large shard size. (On azure...
2023-06-25T04:19:13
2023-06-29T16:06:08
2023-06-29T16:06:08
### Describe the bug https://github.com/huggingface/datasets/blob/a8a797cc92e860c8d0df71e0aa826f4d2690713e/src/datasets/load.py#L1809 What I can to is break the `load_dataset` and use `load_datset_builder` + `download_and_prepare` instead. ### Steps to reproduce the bug https://github.com/huggingface/datasets/blo...
npuichigo
https://github.com/huggingface/datasets/issues/5987
null
false
1,772,233,111
5,986
Make IterableDataset.from_spark more efficient
closed
[ "@lhoestq would you be able to review this please and also approve the workflow?", "Sounds good to me :) feel free to run `make style` to apply code formatting", "_The documentation is not available anymore as the PR was closed or merged._", "cool ! I think we can merge once all comments have been addressed",...
2023-06-23T22:18:20
2023-07-07T10:05:58
2023-07-07T09:56:09
Moved the code from using collect() to using toLocalIterator, which allows for prefetching partitions that will be selected next, thus allowing for better performance when iterating.
mathewjacob1002
https://github.com/huggingface/datasets/pull/5986
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5986", "html_url": "https://github.com/huggingface/datasets/pull/5986", "diff_url": "https://github.com/huggingface/datasets/pull/5986.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5986.patch", "merged_at": "2023-07-07T09:56...
true
1,771,588,158
5,985
Cannot reuse tokenizer object for dataset map
closed
[ "This is a known issue: https://github.com/huggingface/datasets/issues/3847.\r\n\r\nFixing this requires significant work - rewriting the `tokenizers` lib to make them immutable.\r\n\r\nThe current solution is to pass `cache_file_name` to `map` to use that file for caching or calling a tokenizer before `map` (with ...
2023-06-23T14:45:31
2023-07-21T14:09:14
2023-07-21T14:09:14
### Describe the bug Related to https://github.com/huggingface/transformers/issues/24441. Not sure if this is a tokenizer issue or caching issue, so filing in both. Passing the tokenizer to the dataset map function causes the tokenizer to be fingerprinted weirdly. After calling the tokenizer with arguments like pad...
vikigenius
https://github.com/huggingface/datasets/issues/5985
null
false
1,771,571,458
5,984
AutoSharding IterableDataset's when num_workers > 1
open
[ "For this to be possible, we would have to switch from the \"Streaming\" Arrow format to the \"Random Access\" (IPC/Feather) format, which allows reading arbitrary record batches (explained [here](https://arrow.apache.org/docs/python/ipc.html)). We could then use these batches to construct shards.\r\n\r\n@lhoestq @...
2023-06-23T14:34:20
2024-03-22T15:01:14
null
### Feature request Minimal Example ``` import torch from datasets import IterableDataset d = IterableDataset.from_file(<file_name>) dl = torch.utils.data.dataloader.DataLoader(d,num_workers=3) for sample in dl: print(sample) ``` Warning: Too many dataloader workers: 2 (max is dataset.n_shard...
mathephysicist
https://github.com/huggingface/datasets/issues/5984
null
false
1,770,578,804
5,983
replaced PathLike as a variable for save_to_disk for dataset_path wit…
closed
[]
2023-06-23T00:57:05
2023-09-11T04:17:17
2023-09-11T04:17:17
…h str like that of load_from_disk
benjaminbrown038
https://github.com/huggingface/datasets/pull/5983
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5983", "html_url": "https://github.com/huggingface/datasets/pull/5983", "diff_url": "https://github.com/huggingface/datasets/pull/5983.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5983.patch", "merged_at": null }
true
1,770,333,296
5,982
404 on Datasets Documentation Page
closed
[ "This wasn’t working for me a bit earlier, but it looks to be back up now", "We had a minor issue updating the docs after the latest release. It should work now :)." ]
2023-06-22T20:14:57
2023-06-26T15:45:03
2023-06-26T15:45:03
### Describe the bug Getting a 404 from the Hugging Face Datasets docs page: https://huggingface.co/docs/datasets/index ### Steps to reproduce the bug 1. Go to URL https://huggingface.co/docs/datasets/index 2. Notice 404 not found ### Expected behavior URL should either show docs or redirect to new location #...
kmulka-bloomberg
https://github.com/huggingface/datasets/issues/5982
null
false
1,770,310,087
5,981
Only two cores are getting used in sagemaker with pytorch 3.10 kernel
closed
[ "I think it's more likely that this issue is related to PyTorch than Datasets, as PyTorch (on import) registers functions to execute when forking a process. Maybe this is the culprit: https://github.com/pytorch/pytorch/issues/99625", "From reading that ticket, it may be down in mkl? Is it worth hotfixing in the ...
2023-06-22T19:57:31
2023-10-30T06:17:40
2023-07-24T11:54:52
### Describe the bug When using the newer pytorch 3.10 kernel, only 2 cores are being used by huggingface filter and map functions. The Pytorch 3.9 kernel would use as many cores as specified in the num_proc field. We have solved this in our own code by placing the following snippet in the code that is called insi...
mmr-crexi
https://github.com/huggingface/datasets/issues/5981
null
false
1,770,255,973
5,980
Viewing dataset card returns “502 Bad Gateway”
closed
[ "Can you try again? Maybe there was a minor outage.", "Yes, it seems to be working now. In case it's helpful, the outage lasted several days. It was failing as late as yesterday morning. ", "we fixed something on the server side, glad it's fixed now" ]
2023-06-22T19:14:48
2023-06-27T08:38:19
2023-06-26T14:42:45
The url is: https://huggingface.co/datasets/Confirm-Labs/pile_ngrams_trigrams I am able to successfully view the “Files and versions” tab: [Confirm-Labs/pile_ngrams_trigrams at main](https://huggingface.co/datasets/Confirm-Labs/pile_ngrams_trigrams/tree/main) Any help would be appreciated! Thanks! I hope this is ...
tbenthompson
https://github.com/huggingface/datasets/issues/5980
null
false
1,770,198,250
5,979
set dev version
closed
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5979). All of your documentation changes will be reflected on that endpoint.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma...
2023-06-22T18:32:14
2023-06-22T18:42:22
2023-06-22T18:32:22
null
lhoestq
https://github.com/huggingface/datasets/pull/5979
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5979", "html_url": "https://github.com/huggingface/datasets/pull/5979", "diff_url": "https://github.com/huggingface/datasets/pull/5979.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5979.patch", "merged_at": "2023-06-22T18:32...
true
1,770,187,053
5,978
Release: 2.13.1
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-22T18:23:11
2023-06-22T18:40:24
2023-06-22T18:30:16
null
lhoestq
https://github.com/huggingface/datasets/pull/5978
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5978", "html_url": "https://github.com/huggingface/datasets/pull/5978", "diff_url": "https://github.com/huggingface/datasets/pull/5978.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5978.patch", "merged_at": "2023-06-22T18:30...
true
1,768,503,913
5,976
Avoid stuck map operation when subprocesses crashes
closed
[ "Hi ! Do you think this can be fixed at the Pool level ? Ideally it should be the Pool responsibility to handle this, not the `map` code. We could even subclass Pool if needed (at least the one from `multiprocess`)", "@lhoestq it makes sense to me. Just pushed a refactoring creating a `class ProcessPool(multiproc...
2023-06-21T21:18:31
2023-07-10T09:58:39
2023-07-10T09:50:07
I've been using Dataset.map() with `num_proc=os.cpu_count()` to leverage multicore processing for my datasets, but from time to time I get stuck processes waiting forever. Apparently, when one of the subprocesses is abruptly killed (OOM killer, segfault, SIGKILL, etc), the main process keeps waiting for the async task ...
pappacena
https://github.com/huggingface/datasets/pull/5976
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5976", "html_url": "https://github.com/huggingface/datasets/pull/5976", "diff_url": "https://github.com/huggingface/datasets/pull/5976.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5976.patch", "merged_at": "2023-07-10T09:50...
true
1,768,271,343
5,975
Streaming Dataset behind Proxy - FileNotFoundError
closed
[ "Duplicate of #", "Hi ! can you try to set the upper case environment variables `HTTP_PROXY` and `HTTPS_PROXY` ?\r\n\r\nWe use `aiohttp` for streaming and it uses case sensitive environment variables", "Hi, thanks for the quick reply.\r\n\r\nI set the uppercase env variables with\r\n\r\n`\r\nos.environ['HTTP_PR...
2023-06-21T19:10:02
2023-06-30T05:55:39
2023-06-30T05:55:38
### Describe the bug When trying to stream a dataset i get the following error after a few minutes of waiting. ``` FileNotFoundError: https://huggingface.co/datasets/facebook/voxpopuli/resolve/main/data/n_files.json If the repo is private or gated, make sure to log in with `huggingface-cli login`. ``` I hav...
Veluchs
https://github.com/huggingface/datasets/issues/5975
null
false
1,767,981,231
5,974
Deprecate `errors` param in favor of `encoding_errors` in text builder
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-21T16:31:38
2023-06-26T10:34:43
2023-06-26T10:27:40
For consistency with the JSON builder and Pandas
mariosasko
https://github.com/huggingface/datasets/pull/5974
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5974", "html_url": "https://github.com/huggingface/datasets/pull/5974", "diff_url": "https://github.com/huggingface/datasets/pull/5974.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5974.patch", "merged_at": "2023-06-26T10:27...
true
1,767,897,485
5,972
Filter unsupported extensions
closed
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
2023-06-21T15:43:01
2023-06-22T14:23:29
2023-06-22T14:16:26
I used a regex to filter the data files based on their extension for packaged builders. I tried and a regex is 10x faster that using `in` to check if the extension is in the list of supported extensions. Supersedes https://github.com/huggingface/datasets/pull/5850 Close https://github.com/huggingface/datasets/...
lhoestq
https://github.com/huggingface/datasets/pull/5972
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5972", "html_url": "https://github.com/huggingface/datasets/pull/5972", "diff_url": "https://github.com/huggingface/datasets/pull/5972.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5972.patch", "merged_at": "2023-06-22T14:16...
true
1,767,053,635
5,971
Docs: make "repository structure" easier to find
open
[ "Loading a local dataset also works the same way when `data_files` are not specified, so I agree we should make this info easier to discover \r\n\r\ncc @stevhliu ", "Is this issue open? If so, I will self assign. ", "@benjaminbrown038 Yes, it is. Maybe @stevhliu can give some pointers on improving this doc pag...
2023-06-21T08:26:44
2023-07-05T06:51:38
null
The page https://huggingface.co/docs/datasets/repository_structure explains how to create a simple repository structure without a dataset script. It's the simplest way to create a dataset and should be easier to find, particularly on the docs' first pages.
severo
https://github.com/huggingface/datasets/issues/5971
null
false
1,766,010,356
5,970
description disappearing from Info when Uploading a Dataset Created with `from_dict`
open
[ "Here's a minimal way to reproduce the bug, for the sake of convenience.\r\n````\r\nfrom datasets import Dataset, DatasetInfo, load_dataset\r\n\r\n\r\nepisodes_dict = {\"test\":[1,2,3],\"test2\": [1,2,4]}\r\n\r\nhugging_face_dataset = Dataset.from_dict(\r\n episodes_dict, info=DatasetInfo(description=\"test_str\...
2023-06-20T19:18:26
2023-06-22T14:23:56
null
### Describe the bug When uploading a dataset created locally using `from_dict` with a specified `description` field. It appears before upload, but is missing after upload and re-download. ### Steps to reproduce the bug I think the most relevant pattern in the code might be the following lines: ``` descr...
balisujohn
https://github.com/huggingface/datasets/issues/5970
null
false
1,765,529,905
5,969
Add `encoding` and `errors` params to JSON loader
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-20T14:28:35
2023-06-21T13:39:50
2023-06-21T13:32:22
"Requested" in https://discuss.huggingface.co/t/utf-16-for-datasets/43828/3. `pd.read_json` also has these parameters, so it makes sense to be consistent.
mariosasko
https://github.com/huggingface/datasets/pull/5969
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5969", "html_url": "https://github.com/huggingface/datasets/pull/5969", "diff_url": "https://github.com/huggingface/datasets/pull/5969.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5969.patch", "merged_at": "2023-06-21T13:32...
true
1,765,252,561
5,968
Common Voice datasets still need `use_auth_token=True`
closed
[ "cc @pcuenca as well. \r\n\r\nNot super urgent btw", "The issue commes from the dataset itself and is not related to the `datasets` lib\r\n\r\nsee https://huggingface.co/datasets/mozilla-foundation/common_voice_6_1/blob/2c475b3b88e0f2e5828f830a4b91618a25ff20b7/common_voice_6_1.py#L148-L152", "Let's remove these...
2023-06-20T11:58:37
2023-07-29T16:08:59
2023-07-29T16:08:58
### Describe the bug We don't need to pass `use_auth_token=True` anymore to download gated datasets or models, so the following should work if correctly logged in. ```py from datasets import load_dataset load_dataset("mozilla-foundation/common_voice_6_1", "tr", split="train+validation") ``` However it throw...
patrickvonplaten
https://github.com/huggingface/datasets/issues/5968
null
false
1,763,926,520
5,967
Config name / split name lost after map with multiproc
open
[ "This must be due to DatasetInfo.from_merge which drops them and is used in `concatenate_datasets`.\r\n\r\nAnd you're experiencing this issue because multiprocessing does concatenate the resulting datasets from each process.\r\n\r\nMaybe they should be kept if all the subdatasets share the same values for config_na...
2023-06-19T17:27:36
2023-06-28T08:55:25
null
### Describe the bug Performing a `.map` method on a dataset loses it's config name / split name only if run with multiproc ### Steps to reproduce the bug ```python from datasets import Audio, load_dataset from transformers import AutoFeatureExtractor import numpy as np # load dummy dataset libri = load_datas...
sanchit-gandhi
https://github.com/huggingface/datasets/issues/5967
null
false
1,763,885,914
5,966
Fix JSON generation in benchmarks CI
closed
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
2023-06-19T16:56:06
2023-06-19T17:29:11
2023-06-19T17:22:10
Related to changes made in https://github.com/iterative/dvc/pull/9475
mariosasko
https://github.com/huggingface/datasets/pull/5966
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5966", "html_url": "https://github.com/huggingface/datasets/pull/5966", "diff_url": "https://github.com/huggingface/datasets/pull/5966.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5966.patch", "merged_at": "2023-06-19T17:22...
true
1,763,648,540
5,965
"Couldn't cast array of type" in complex datasets
closed
[ "Thanks for reporting! \r\n\r\nSpecifying the target features explicitly should avoid this error:\r\n```python\r\ndataset = dataset.map(\r\n batch_process,\r\n batched=True,\r\n batch_size=1,\r\n num_proc=1,\r\n remove_columns=dataset.column_names,\r\n features=datasets.Features({\"texts\": datase...
2023-06-19T14:16:14
2023-07-26T15:13:53
2023-07-26T15:13:53
### Describe the bug When doing a map of a dataset with complex types, sometimes `datasets` is unable to interpret the valid schema of a returned datasets.map() function. This often comes from conflicting types, like when both empty lists and filled lists are competing for the same field value. This is prone to hap...
piercefreeman
https://github.com/huggingface/datasets/issues/5965
null
false
1,763,513,574
5,964
Always return list in `list_datasets`
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-19T13:07:08
2023-06-19T17:29:37
2023-06-19T17:22:41
Fix #5925 Plus, deprecate `list_datasets`/`inspect_dataset` in favor of `huggingface_hub.list_datasets`/"git clone workflow" (downloads data files)
mariosasko
https://github.com/huggingface/datasets/pull/5964
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5964", "html_url": "https://github.com/huggingface/datasets/pull/5964", "diff_url": "https://github.com/huggingface/datasets/pull/5964.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5964.patch", "merged_at": "2023-06-19T17:22...
true
1,762,774,457
5,963
Got an error _pickle.PicklingError use Dataset.from_spark.
closed
[ "i got error using method from_spark when using multi-node Spark cluster. seems could only use \"from_spark\" in local?", "@lhoestq ", "cc @maddiedawson it looks like there an issue with `_validate_cache_dir` ?\r\n\r\nIt looks like the function passed to mapPartitions has a reference to the Spark dataset build...
2023-06-19T05:30:35
2023-07-24T11:55:46
2023-07-24T11:55:46
python 3.9.2 Got an error _pickle.PicklingError use Dataset.from_spark. Did the dataset import load data from spark dataframe using multi-node Spark cluster df = spark.read.parquet(args.input_data).repartition(50) ds = Dataset.from_spark(df, keep_in_memory=True, cache_dir="...
yanzia12138
https://github.com/huggingface/datasets/issues/5963
null
false
1,761,589,882
5,962
Issue with train_test_split maintaining the same underlying PyArrow Table
open
[]
2023-06-17T02:19:58
2023-06-17T02:19:58
null
### Describe the bug I've been using the train_test_split method in the datasets module to split my HuggingFace Dataset into separate training, validation, and testing subsets. However, I've noticed an issue where the split datasets appear to maintain the same underlying PyArrow Table. ### Steps to reproduce the bug ...
Oziel14
https://github.com/huggingface/datasets/issues/5962
null
false
1,758,525,111
5,961
IterableDataset: split by node and map may preprocess samples that will be skipped anyway
open
[ "Does \"number of shards\" refer to the total number of data?\r\n\r\nmy config:\r\nnproc_per_node=2\r\nds=ds['train'] = load_dataset(streaming=True).take(50000)\r\n\r\nI'm test again: in prepare_data(), data have the same for each GPU\r\n", "The number of shards is `ds.n_shards`. It corresponds generally to the ...
2023-06-15T10:29:10
2023-09-01T10:35:11
null
There are two ways an iterable dataset can be split by node: 1. if the number of shards is a factor of number of GPUs: in that case the shards are evenly distributed per GPU 2. otherwise, each GPU iterate on the data and at the end keeps 1 sample out of n(GPUs) - skipping the others. In case 2. it's ...
johnchienbronci
https://github.com/huggingface/datasets/issues/5961
null
false
1,757,397,507
5,959
read metric glue.py from local file
closed
[ "Sorry, I solve this by call `evaluate.load('glue_metric.py','sst-2')`\r\n" ]
2023-06-14T17:59:35
2023-06-14T18:04:16
2023-06-14T18:04:16
### Describe the bug Currently, The server is off-line. I am using the glue metric from the local file downloaded from the hub. I download / cached datasets using `load_dataset('glue','sst2', cache_dir='/xxx')` to cache them and then in the off-line mode, I use `load_dataset('xxx/glue.py','sst2', cache_dir='/xxx'...
JiazhaoLi
https://github.com/huggingface/datasets/issues/5959
null
false
1,757,265,971
5,958
set dev version
closed
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_5958). All of your documentation changes will be reflected on that endpoint.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchma...
2023-06-14T16:26:34
2023-06-14T16:34:55
2023-06-14T16:26:51
null
lhoestq
https://github.com/huggingface/datasets/pull/5958
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5958", "html_url": "https://github.com/huggingface/datasets/pull/5958", "diff_url": "https://github.com/huggingface/datasets/pull/5958.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5958.patch", "merged_at": "2023-06-14T16:26...
true
1,757,252,466
5,957
Release: 2.13.0
closed
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
2023-06-14T16:17:26
2023-06-14T16:33:39
2023-06-14T16:24:39
null
lhoestq
https://github.com/huggingface/datasets/pull/5957
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5957", "html_url": "https://github.com/huggingface/datasets/pull/5957", "diff_url": "https://github.com/huggingface/datasets/pull/5957.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5957.patch", "merged_at": "2023-06-14T16:24...
true
1,756,959,367
5,956
Fix ArrowExamplesIterable.shard_data_sources
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-14T13:50:38
2023-06-14T14:43:12
2023-06-14T14:33:45
ArrowExamplesIterable.shard_data_sources was outdated I also fixed a warning message by not using format_type= in with_format()
lhoestq
https://github.com/huggingface/datasets/pull/5956
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5956", "html_url": "https://github.com/huggingface/datasets/pull/5956", "diff_url": "https://github.com/huggingface/datasets/pull/5956.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5956.patch", "merged_at": "2023-06-14T14:33...
true
1,756,827,133
5,955
Strange bug in loading local JSON files, using load_dataset
closed
[ "This is the actual error:\r\n```\r\nFailed to read file '/home/lakala/hjc/code/pycode/glm/temp.json' with error <class 'pyarrow.lib.ArrowInvalid'>: cannot mix list and non-list, non-null values\r\n```\r\nWhich means some samples are incorrectly formatted.\r\n\r\nPyArrow, a storage backend that we use under the hoo...
2023-06-14T12:46:00
2023-06-21T14:42:15
2023-06-21T14:42:15
### Describe the bug I am using 'load_dataset 'loads a JSON file, but I found a strange bug: an error will be reported when the length of the JSON file exceeds 160000 (uncertain exact number). I have checked the data through the following code and there are no issues. So I cannot determine the true reason for this err...
Night-Quiet
https://github.com/huggingface/datasets/issues/5955
null
false
1,756,572,994
5,954
Better filenotfound for gated
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-14T10:33:10
2023-06-14T12:33:27
2023-06-14T12:26:31
close https://github.com/huggingface/datasets/issues/5953 <img width="1292" alt="image" src="https://github.com/huggingface/datasets/assets/42851186/270fe5bc-1739-4878-b7bc-ab6d35336d4d">
lhoestq
https://github.com/huggingface/datasets/pull/5954
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5954", "html_url": "https://github.com/huggingface/datasets/pull/5954", "diff_url": "https://github.com/huggingface/datasets/pull/5954.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5954.patch", "merged_at": "2023-06-14T12:26...
true
1,756,520,523
5,953
Bad error message when trying to download gated dataset
closed
[ "cc @sanchit-gandhi @Vaibhavs10 @lhoestq - this is mainly for demos that use Common Voice datasets as done here: https://github.com/facebookresearch/fairseq/tree/main/examples/mms#-transformers\r\n", "Hi ! the error for me is\r\n\r\n```\r\nFileNotFoundError: Couldn't find a dataset script at /content/mozilla-foun...
2023-06-14T10:03:39
2023-06-14T16:36:51
2023-06-14T12:26:32
### Describe the bug When I attempt to download a model from the Hub that is gated without being logged in, I get a nice error message. E.g.: E.g. ```sh Repository Not Found for url: https://huggingface.co/api/models/DeepFloyd/IF-I-XL-v1.0. Please make sure you specified the correct `repo_id` and `repo_type`. I...
patrickvonplaten
https://github.com/huggingface/datasets/issues/5953
null
false
1,756,481,591
5,952
Add Arrow builder docs
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-14T09:42:46
2023-06-14T14:42:31
2023-06-14T14:34:39
following https://github.com/huggingface/datasets/pull/5944
lhoestq
https://github.com/huggingface/datasets/pull/5952
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5952", "html_url": "https://github.com/huggingface/datasets/pull/5952", "diff_url": "https://github.com/huggingface/datasets/pull/5952.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5952.patch", "merged_at": "2023-06-14T14:34...
true
1,756,363,546
5,951
What is the Right way to use discofuse dataset??
closed
[ "Thanks for opening https://huggingface.co/datasets/discofuse/discussions/3, let's continue the discussion over there if you don't mind", "I have posted there also sir, please check\r\n@lhoestq" ]
2023-06-14T08:38:39
2023-06-14T13:25:06
2023-06-14T12:10:16
[Click here for Dataset link](https://huggingface.co/datasets/discofuse/viewer/discofuse-wikipedia/train?row=6) **Below is the following way, as per my understanding , Is it correct :question: :question:** The **columns/features from `DiscoFuse dataset`** that will be the **input to the `encoder` and `decoder`** ar...
akesh1235
https://github.com/huggingface/datasets/issues/5951
null
false
1,755,197,946
5,950
Support for data with instance-wise dictionary as features
open
[ "Hi ! We use the Arrow columnar format under the hood, which doesn't support such dictionaries: each field must have a fixed type and exist in each sample.\r\n\r\nInstead you can restructure your data like\r\n```\r\n{\r\n \"index\": 0,\r\n \"keys\": [\"2 * x + y >= 3\"],\r\n \"values\": [[\"2 * x + y >= 3\...
2023-06-13T15:49:00
2025-04-07T13:20:37
null
### Feature request I notice that when loading data instances with feature type of python dictionary, the dictionary keys would be broadcast so that every instance has the same set of keys. Please see an example in the Motivation section. It is possible to avoid this behavior, i.e., load dictionary features as it i...
richardwth
https://github.com/huggingface/datasets/issues/5950
null
false
1,754,843,717
5,949
Replace metadata utils with `huggingface_hub`'s RepoCard API
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-13T13:03:19
2023-06-27T16:47:51
2023-06-27T16:38:32
Use `huggingface_hub`'s RepoCard API instead of `DatasetMetadata` for modifying the card's YAML, and deprecate `datasets.utils.metadata` and `datasets.utils.readme`. After removing these modules, we can also delete `datasets.utils.resources` since the moon landing repo now stores its own version of these resources f...
mariosasko
https://github.com/huggingface/datasets/pull/5949
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5949", "html_url": "https://github.com/huggingface/datasets/pull/5949", "diff_url": "https://github.com/huggingface/datasets/pull/5949.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5949.patch", "merged_at": "2023-06-27T16:38...
true
1,754,794,611
5,948
Fix sequence of array support for most dtype
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-13T12:38:59
2023-06-14T15:11:55
2023-06-14T15:03:33
Fixes #5936 Also, a related fix to #5927
qgallouedec
https://github.com/huggingface/datasets/pull/5948
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5948", "html_url": "https://github.com/huggingface/datasets/pull/5948", "diff_url": "https://github.com/huggingface/datasets/pull/5948.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5948.patch", "merged_at": "2023-06-14T15:03...
true
1,754,359,316
5,947
Return the audio filename when decoding fails due to corrupt files
open
[ "Hi ! The audio data don't always exist as files on disk - the blobs are often stored in the Arrow files. For now I'd suggest disabling decoding with `.cast_column(\"audio\", Audio(decode=False))` and apply your own decoding that handles corrupted files (maybe to filter them out ?)\r\n\r\ncc @sanchit-gandhi since i...
2023-06-13T08:44:09
2023-06-14T12:45:01
null
### Feature request Return the audio filename when the audio decoding fails. Although currently there are some checks for mp3 and opus formats with the library version there are still cases when the audio decoding could fail, eg. Corrupt file. ### Motivation When you try to load an object file dataset and the...
wetdog
https://github.com/huggingface/datasets/issues/5947
null
false
1,754,234,469
5,946
IndexError Not Solving -> IndexError: Invalid key: ?? is out of bounds for size 0 or ??
open
[ "https://colab.research.google.com/#scrollTo=AQ_HCYruWIHU&fileId=https%3A//huggingface.co/dfurman/falcon-40b-chat-oasst1/blob/main/finetune_falcon40b_oasst1_with_bnb_peft.ipynb\r\n\r\nI ran the same administration exactly the same but got the same error", "Looks related to https://discuss.huggingface.co/t/indexer...
2023-06-13T07:34:15
2023-07-14T12:04:48
null
### Describe the bug in <cell line: 1>:1 │ │ │ │ /usr/local/lib/python3.10/dist-packages/transformers/trainer.py:1537 in train ...
syngokhan
https://github.com/huggingface/datasets/issues/5946
null
false
1,754,084,577
5,945
Failing to upload dataset to the hub
closed
[ "Hi ! Feel free to re-run your code later, it will resume automatically where you left", "Tried many times in the last 2 weeks, problem remains.", "Alternatively you can save your dataset in parquet files locally and upload them to the hub manually\r\n\r\n```python\r\nfrom tqdm import tqdm\r\nnum_shards = 60\r\...
2023-06-13T05:46:46
2023-07-24T11:56:40
2023-07-24T11:56:40
### Describe the bug Trying to upload a dataset of hundreds of thousands of audio samples (the total volume is not very large, 60 gb) to the hub with push_to_hub, it doesn't work. From time to time one piece of the data (parquet) gets pushed and then I get RemoteDisconnected even though my internet is stable. Please...
Ar770
https://github.com/huggingface/datasets/issues/5945
null
false
1,752,882,200
5,944
Arrow dataset builder to be able to load and stream Arrow datasets
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "@lhoestq tips applied. Thanks for a review. :smile: It's a lot of fun to improve this project. ", "Let's add some documentation in a subsequent PR :)\r\n\r\nIn particular @mariosasko and I think it's important to note to users tha...
2023-06-12T14:21:49
2023-06-13T17:36:02
2023-06-13T17:29:01
This adds a Arrow dataset builder to be able to load and stream from already preprocessed Arrow files. It's related to https://github.com/huggingface/datasets/issues/3035
mariusz-jachimowicz-83
https://github.com/huggingface/datasets/pull/5944
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5944", "html_url": "https://github.com/huggingface/datasets/pull/5944", "diff_url": "https://github.com/huggingface/datasets/pull/5944.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5944.patch", "merged_at": "2023-06-13T17:29...
true
1,752,021,681
5,942
Pass datasets-cli additional args as kwargs to DatasetBuilder in `run_beam.py`
open
[]
2023-06-12T06:50:50
2023-06-30T09:15:00
null
Hi, Following this <https://discuss.huggingface.co/t/how-to-preprocess-a-wikipedia-dataset-using-dataflowrunner/41991/3>, here is a simple PR to pass any additional args to datasets-cli as kwargs in the DatasetBuilder in `run_beam.py`. I also took the liberty to add missing setup steps to the `beam.mdx` docs in o...
graelo
https://github.com/huggingface/datasets/pull/5942
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5942", "html_url": "https://github.com/huggingface/datasets/pull/5942", "diff_url": "https://github.com/huggingface/datasets/pull/5942.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5942.patch", "merged_at": null }
true
1,751,838,897
5,941
Load Data Sets Too Slow In Train Seq2seq Model
closed
[ "Hi ! you can speed it up using multiprocessing by passing `num_proc=` to `load_dataset()`", "already did,but not useful for step Generating train split,it works in step \"Resolving data files\" & \"Downloading data files\" ", "@mariosasko some advice , thanks!", "I met the same problem, terrible experience...
2023-06-12T03:58:43
2023-08-15T02:52:22
2023-08-15T02:52:22
### Describe the bug step 'Generating train split' in load_dataset is too slow: ![image](https://github.com/huggingface/datasets/assets/19569322/d9b08eee-95fe-4741-a346-b70416c948f8) ### Steps to reproduce the bug Data: own data,16K16B Mono wav Oficial Script:[ run_speech_recognition_seq2seq.py](https://github...
xyx361100238
https://github.com/huggingface/datasets/issues/5941
null
false
1,774,389,854
5,990
Pushing a large dataset on the hub consistently hangs
open
[ "Hi @AntreasAntoniou , sorry to know you are facing this issue. To help debugging it, could you tell me:\r\n- What is the total dataset size?\r\n- Is it always failing on the same shard or is the hanging problem happening randomly?\r\n- Were you able to save the dataset as parquet locally? This would help us determ...
2023-06-10T14:46:47
2025-02-15T09:29:10
null
### Describe the bug Once I have locally built a large dataset that I want to push to hub, I use the recommended approach of .push_to_hub to get the dataset on the hub, and after pushing a few shards, it consistently hangs. This has happened over 40 times over the past week, and despite my best efforts to try and catc...
AntreasAntoniou
https://github.com/huggingface/datasets/issues/5990
null
false
1,749,955,883
5,939
.
closed
[]
2023-06-09T14:01:34
2023-06-12T12:19:34
2023-06-12T12:19:19
null
flckv
https://github.com/huggingface/datasets/issues/5939
null
false
1,749,462,851
5,938
Make get_from_cache use custom temp filename that is locked
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-09T09:01:13
2023-06-14T13:35:38
2023-06-14T13:27:24
This PR ensures that the temporary filename created is the same as the one that is locked, while writing to the cache. This PR stops using `tempfile` to generate the temporary filename. Additionally, the behavior now is aligned for both `resume_download` `True` and `False`. Refactor temp_file_manager so that i...
albertvillanova
https://github.com/huggingface/datasets/pull/5938
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5938", "html_url": "https://github.com/huggingface/datasets/pull/5938", "diff_url": "https://github.com/huggingface/datasets/pull/5938.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5938.patch", "merged_at": "2023-06-14T13:27...
true
1,749,388,597
5,937
Avoid parallel redownload in cache
closed
[ "_The documentation is not available anymore as the PR was closed or merged._", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | rea...
2023-06-09T08:18:36
2023-06-14T12:30:59
2023-06-14T12:23:57
Avoid parallel redownload in cache by retrying inside the lock if path exists.
albertvillanova
https://github.com/huggingface/datasets/pull/5937
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5937", "html_url": "https://github.com/huggingface/datasets/pull/5937", "diff_url": "https://github.com/huggingface/datasets/pull/5937.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5937.patch", "merged_at": "2023-06-14T12:23...
true
1,748,424,388
5,936
Sequence of array not supported for most dtype
closed
[ "Related, `float16` is the only dtype not supported by `Array2D` (probably by every `ArrayND`):\r\n\r\n```python\r\nfrom datasets import Array2D, Features, Dataset\r\n\r\nimport numpy as np\r\n\r\nfor dtype in [\r\n \"bool\", # ok\r\n \"int8\", # ok\r\n \"int16\", # ok\r\n \"int32\", # ok\r\n \"i...
2023-06-08T18:18:07
2023-06-14T15:03:34
2023-06-14T15:03:34
### Describe the bug Create a dataset composed of sequence of array fails for most dtypes (see code below). ### Steps to reproduce the bug ```python from datasets import Sequence, Array2D, Features, Dataset import numpy as np for dtype in [ "bool", # ok "int8", # failed "int16", # failed ...
qgallouedec
https://github.com/huggingface/datasets/issues/5936
null
false
1,748,090,220
5,935
Better row group size in push_to_hub
closed
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
2023-06-08T15:01:15
2023-06-09T17:47:37
2023-06-09T17:40:09
This is a very simple change that improves `to_parquet` to use a more reasonable row group size for image and audio datasets. This is especially useful for `push_to_hub` and will provide a better experience with the dataset viewer on HF
lhoestq
https://github.com/huggingface/datasets/pull/5935
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5935", "html_url": "https://github.com/huggingface/datasets/pull/5935", "diff_url": "https://github.com/huggingface/datasets/pull/5935.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5935.patch", "merged_at": "2023-06-09T17:40...
true
1,747,904,840
5,934
Modify levels of some logging messages
closed
[ "I've addressed this as part of #6019, so feel free to close this PR. ", "Thanks !" ]
2023-06-08T13:31:44
2023-07-12T18:21:03
2023-07-12T18:21:02
Some warning messages didn't quite sound like warnings so I modified their logging levels to info.
Laurent2916
https://github.com/huggingface/datasets/pull/5934
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5934", "html_url": "https://github.com/huggingface/datasets/pull/5934", "diff_url": "https://github.com/huggingface/datasets/pull/5934.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5934.patch", "merged_at": null }
true
1,747,382,500
5,933
Fix `to_numpy` when None values in the sequence
closed
[ "I just added the same test with dynamic shape", "_The documentation is not available anymore as the PR was closed or merged._", "Awesome ! I'm merging now if you don't mind :)\r\nWe should probably give you permissions to merge your own PRs when you have an approval", "<details>\n<summary>Show benchmarks</su...
2023-06-08T08:38:56
2023-06-09T13:49:41
2023-06-09T13:23:48
Closes #5927 I've realized that the error was overlooked during testing due to the presence of only one None value in the sequence. Unfortunately, it was the only case where the function works as expected. When the sequence contained more than one None value, the function failed. Consequently, I've updated the tests...
qgallouedec
https://github.com/huggingface/datasets/pull/5933
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5933", "html_url": "https://github.com/huggingface/datasets/pull/5933", "diff_url": "https://github.com/huggingface/datasets/pull/5933.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5933.patch", "merged_at": "2023-06-09T13:23...
true
1,746,249,161
5,932
[doc build] Use secrets
closed
[ "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>Show updated benchmarks!</summary>\n\n### Benchmark: benchmark_array_xd.json\n\n| metric | read_batch_formatted_as_numpy after write_array2d | read_batch_formatted_as_numpy after write_flattened_sequence | read_batch_formatted_a...
2023-06-07T16:09:39
2023-06-09T10:16:58
2023-06-09T09:53:16
Companion pr to https://github.com/huggingface/doc-builder/pull/379
mishig25
https://github.com/huggingface/datasets/pull/5932
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/5932", "html_url": "https://github.com/huggingface/datasets/pull/5932", "diff_url": "https://github.com/huggingface/datasets/pull/5932.diff", "patch_url": "https://github.com/huggingface/datasets/pull/5932.patch", "merged_at": "2023-06-09T09:53...
true
1,745,408,784
5,931
`datasets.map` not reusing cached copy by default
closed
[ "This can happen when a map transform cannot be hashed deterministically (e.g., an object referenced by the transform changes its state after the first call - an issue with fast tokenizers). The solution is to provide `cache_file_name` in the `map` call to check this file for the cached result instead of relying on...
2023-06-07T09:03:33
2023-06-21T16:15:40
2023-06-21T16:15:40
### Describe the bug When I load the dataset from local directory, it's cached copy is picked up after first time. However, for `map` operation, the operation is applied again and cached copy is not picked up. Is there any way to pick cached copy instead of processing it again? The only solution I could think of was...
bhavitvyamalik
https://github.com/huggingface/datasets/issues/5931
null
false